<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">A Knowledge-Based Service Architecture for Legal Document Building</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Silvana</forename><surname>Castano</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Department of Computer Science</orgName>
								<orgName type="institution">Università degli Studi di Milano</orgName>
								<address>
									<addrLine>Via Celoria</addrLine>
									<postCode>18 -20133</postCode>
									<settlement>Milano</settlement>
									<country key="IT">Italy</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Alfio</forename><surname>Ferrara</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Department of Computer Science</orgName>
								<orgName type="institution">Università degli Studi di Milano</orgName>
								<address>
									<addrLine>Via Celoria</addrLine>
									<postCode>18 -20133</postCode>
									<settlement>Milano</settlement>
									<country key="IT">Italy</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Stefano</forename><surname>Montanelli</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Department of Computer Science</orgName>
								<orgName type="institution">Università degli Studi di Milano</orgName>
								<address>
									<addrLine>Via Celoria</addrLine>
									<postCode>18 -20133</postCode>
									<settlement>Milano</settlement>
									<country key="IT">Italy</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Sergio</forename><surname>Picascia</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Department of Computer Science</orgName>
								<orgName type="institution">Università degli Studi di Milano</orgName>
								<address>
									<addrLine>Via Celoria</addrLine>
									<postCode>18 -20133</postCode>
									<settlement>Milano</settlement>
									<country key="IT">Italy</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Davide</forename><surname>Riva</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Department of Computer Science</orgName>
								<orgName type="institution">Università degli Studi di Milano</orgName>
								<address>
									<addrLine>Via Celoria</addrLine>
									<postCode>18 -20133</postCode>
									<settlement>Milano</settlement>
									<country key="IT">Italy</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">A Knowledge-Based Service Architecture for Legal Document Building</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">E97D8359B6B33C046311CE3EEB95FEF7</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T19:26+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Digital Justice</term>
					<term>Knowledge Extraction</term>
					<term>Legal Concept Graph</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>In this paper, we propose a knowledge-based service architecture for legal document building based on Natural Language Processing and learning techniques, to semantically analyze a database of ingested legal documents and propose the most prominent and pertinent textual suggestions for new document composition. After describing the proposed NLP services for knowledge extraction and textual suggestion selection and proposition, we describe the application of proposed document builder architecture by considering a case study of Italian civil judgements.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>Legal documents constantly produced by Parliaments, Courts, and other institutional bodies constitute a prominent source of information and knowledge not only for legal actors, like judges or lawyers, but also for general subjects, like citizens or private and public organizations. To improve both efficiency and effectiveness of courthouses and legal record offices, and to foster digital justice, a significant effort is being devoted in almost all countries to digital transformation projects, by developing legal information systems and modular architectures providing a variety of services for acquisition, management, classification, exploration, and retrieval of legal documents <ref type="bibr" target="#b0">[1]</ref>. A main issue fostering digital transformation and courthouses efficiency, is related to the availability of tools to support legal actors in the complex process of producing a new document for the case at hand, by making available databases of previous documentation as well as workflow management services for the different stages of the legal proceedings <ref type="bibr" target="#b1">[2]</ref>. In particular, advanced legal document building environments are required, to assist legal actors in the production of a new document, like judgements or lawyer acts, by relying on predefined document templates and knowledge-based services to properly extract useful information available in previously produced legal documents to propose in form of 2nd Workshop on Knowledge Management and Process Mining for Law, 9th Joint Ontology Workshops (JOWO 2023), co-located with FOIS 2023, 19-20 July, 2023, Sherbrooke, Québec, Canada Envelope silvana.castano@unimi.it (S. Castano); alfio.ferrara@unimi.it (A. Ferrara); stefano.montanelli@unimi.it (S. Montanelli); sergio.picascia@unimi.it (S. Picascia); davide.riva1@unimi.it (D. Riva) Orcid 0000-0002-3826-2407 (S. Castano); 0000-0002-4991-4984 (A. Ferrara); 0000-0002-6594-6644 (S. Montanelli); 0000-0001-6863-0082 (S. Picascia); 0009-0003-9681-9423 (D. Riva) suggestions. The availability of predefined document templates <ref type="bibr" target="#b2">[3]</ref> and rules <ref type="bibr" target="#b3">[4]</ref> promotes a disciplined structuring of legal documents, which will be eventually ingested in the document repository for subsequent processing. The availability of a legal document repository where ingested documents have a very homogeneous structure can, in turn, facilitate the application of AI techniques, for semantic processing and classification of document contents.</p><p>In this paper, we propose a knowledge-based service architecture for legal document building based on Natural Language Processing and Zero-Shot Learning techniques, to semantically analyze a database of ingested legal documents and propose the most prominent and pertinent textual suggestions for the composition of a new document. In particular, the document building process relies on a predefined document template organized in sections, according to a segmentation schema, and on a knowledge-extraction service and a suggestion-extraction service to enforce the document building process. The knowledge-extraction service is responsible for i) mining a set of featuring concepts that provide a topic-oriented description of textual contents of ingested documents and ii) implementing a fine-grained semantic classification of textual contents, where each concept is connected to the document portions from which the concept emerged. The suggestion-extraction service is responsible for i) Search-by-Content, to retrieve text portions of ingested documents that are most pertinent to the current text of the document section under preparation; ii) Search-by-Concept, to retrieve text portions of ingested documents that are most pertinent to one or more concepts, and possibly document section(s) of interest, specified by the legal actor, that is, text portions that contain terminological occurrences of the considered concept(s), coming from specified section(s), if any. The proposed service architecture supports the legal document building process according to a "human-in-the-loop approach", where the document author, namely the legal actor, drafts the document under preparation following the document template, by actively working on the text portions provided by the suggestion service.</p><p>The paper is organized as follows. In Section 2 we introduce the proposed service architecture for a legal document builder. Section 3 describes the knowledge-extraction service, while Section 4 illustrates the suggestion-extraction service. In Section 5 we present a case study applying the architecture on a corpus of Italian case law decisions in the framework of the NGUPP project. Section 6 is devoted to related work. Finally, Section 7 provides the final remarks.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Legal Document Builder: the Proposed Architecture</head><p>Document building, sometimes referred to as document assembly, is a process that aims at producing a textual document following a predefined schema with the support of digital, automated tools. The task, even when entirely performed by humans, can be thought of as a sequence of three phases: (1) definition of the document format and structure; (2) content draft of each part or section; (3) editing of the document. Each of these phases is susceptible to automated support to a variable extent, ranging from interactive, human-in-the-loop approaches to fully automated services. For instance, document structure can be totally or partially based on predefined templates, possibly tailored to user needs. At the same time, editing may be supported by solutions ranging from error detection to automated rephrasing and text generation.</p><p>Our proposed architecture is designed to support experts in producing documents specialized domains that require (or would benefit from) a well-defined document structure, such as the legal domain, which is the context we consider in this paper. A main requirement in designing a document builder architecture for the legal domain is to preserve the importance of the legal actor role and work with capability to provide automated support while preserving the uniqueness/autonomy of the assessment and judgment capability of legal actors in document writing. As a consequence, a fully automated approach is not viable nor desirable, and interactive tool environments should be provided, to keep the legal actor "in-the-loop" and to reduce also the risk of excessive standardization of legal documents, which is a requirement too, given the value and purposes of legal documents.</p><p>The architecture provides a set of services to support the document assembly process by relying on predefined document template(s). This way, document building promotes a disciplined approach to document generation, by enhancing the readability, homogeneity, and quality of produced documentation, very important also for subsequent analysis and knowledge extraction purposes.</p><p>The proposed service architecture for legal document building is shown in Figure <ref type="figure" target="#fig_0">1</ref>. The architecture aims to support legal actors like judges and judicial officers in the preparation of official legal documents. In particular, the legal actor, namely the document author, starts working by selecting a template among those available and she/he is supported in building the document by receiving suggestions in the form of pertinent chunks of text that can be reused "as-is" or adjusted/edited for insertion in the document under preparation. Suggestions are extracted from a corpus of legal documents (e.g., judgements, case-law decisions) that have been previously ingested and processed. To this end, a storage layer is defined to maintain i) the document database for the raw ingested documents and corresponding texts; ii) the document chunks and related embeddings extracted from documents for classification; iii) the concept graph to store entries for the concepts extracted from documents.</p><p>Two main service pipelines are defined: the knowledge-extraction service and the suggestionextraction service. In the knowledge-extraction service, appropriate components are defined to exploit the ingested documents for mining a set of featuring concepts that provide a topicoriented description of underlying textual contents. The concepts extracted from the documents are organized in a graph, where a pair of similar concepts is linked by an edge. Each concept is also connected to the document portions from which the concept emerged, meaning that we can retrieve the pertinent document segments where a certain concept occurs. For knowledge extraction, the involved components are document segmentation and data preparation, and concept mining.</p><p>The suggestion-extraction service enforces two modalities to detect pertinent chunks to propose to the user for document drafting: i) Search-by-Content, which retrieves similar chunks as suggestions by considering the current text of the document section under preparation; ii) Search-by-Concept, which is based on a concept specified by the legal actor to retrieve and provide as suggestions those phrases that contain terminological occurrences of the considered concept.</p><p>In the following sections, we describe each service in more detail.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">The Knowledge-Extraction Service</head><p>Ingested legal documents are submitted to a knowledge-extraction pipeline based on document annotation/segmentation, and fine-grained indexing and classification techniques.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1.">Document Segmentation</head><p>Document segmentation is concerned with the identification of relevant sections of a document which are exploited to construct new documents and to filter suggestions for the document drafting. The idea is that the legal actor receives suggestions extracted from the same, or strictly related, document section that is currently under construction. Document segmentation is a type of document annotation that consists in dividing a document into parts (known as segments) that "display local coherence" <ref type="bibr" target="#b4">[5]</ref>, eventually assigning a label to each of them. Since in our case we focus only on textual data, such parts may be contiguous sequences of words, clauses or sentences. Local coherence is here interpreted from a functional point of view, i.e. a text segment plays a single and indivisible function in the structure of the document, e.g. introduction, argumentation, conclusion, and others.</p><p>The first problem to solve is therefore the definition of a segmentation schema comprising a set of functional segments and the descriptions of the respective text characteristics, which serve as guidelines for annotation. Binding rules over the annotation output can be specified by an additional set of axioms, including for instance a lower and/or upper bound on the number of segments of a certain type. In most cases, definition of segment functions and characteristics and axioms over them requires expert knowledge of the domain at hand, and shall be general enough to encompass the variety of documents in the corpus of interest.</p><p>The second problem is the actual annotation of documents, complying with the schema. Given a segmentation schema and a corpus to annotate, the annotation can be performed either manually (by human annotators), automatically (by rule-based or machine learning techniques), or by a hybrid approach. Manual Segmentation. In the manual case, the annotation is performed by a group of human annotators, generally, experts of the domain, who split each document into segments and assign a label from the segmentation schema to each segment using a digital text annotation tool. Leveraging expert knowledge can benefit annotation accuracy, especially in case of complex documents or domain corpora like in the legal domain, but it hinders scalability, since manual annotation is notoriously a labour-intensive, time-consuming activity. Moreover, despite increasing accuracy, expert knowledge doesn't ensure agreement among annotators, which is often necessary to define a unique and homogeneous ground truth. Automated Segmentation. Automated systems for text segmentation typically rely on either rule-based or machine learning models. Such models ensure two main advantages with respect to manual annotation: they can achieve higher scalability, making it possible to process large corpora, and, if machine learning models are employed, text characteristics can be learned automatically instead of being a-priori defined. The main disadvantage is lower accuracy, especially on complex documents. The problem may intensify in case of totally automated systems, which rely on pre-trained models only, in the absence of any manually annotated data. Hybrid Segmentation. By hybrid segmentation we indicate a text segmentation system that receives as input a limited corpus of manually annotated documents and, training and validating an automated system on such data, extends the annotation to other documents.</p><p>In Section 5, we discuss legal document segmentation activity we performed for prototyping a first release of a legal document builder in the framework of an ongoing research project on digital justice in Italy. In this context, legal document segmentation has been manually performed by legal experts of the project on a dataset case study of judgements. Since rule-based models usually fall short in generalization, we deem machine learning models preferable to process large and possibly heterogeneous legal corpora. We plan experimentation of different classifiers as discussed in concluding remarks.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.">Data Preparation</head><p>Data preparation is concerned with the tokenization of ingested documents with the goal to split them into document chunks. A document chunk represents the text unit to consider for classification that can be associated with a concept. Document chunks associated with concepts represent the suggestions that can be provided to the author while building a document. We stress that the size of the document chunk should be large enough, so that the context can be captured, but not too much extended to avoid segments that are long to read and potentially noisy due to the presence of multiple concepts. In this paper, we choose to tokenize documents by defining a chunk for few sentence/phrase detected in a document, up to a maximum size of 128 tokens<ref type="foot" target="#foot_0">1</ref> . This is particularly appropriate for legal actors that are typically interested in retrieving precise suggestions in which a given concept of interest appears and can be rapidly read/assimilated.</p><p>As a further preparation step, the terms appearing in document chunks are lemmatized and a vector-based representation of each document chunk is finally built. The use of embedding techniques to represent chunks allows to map the document contents on a semantic vector space where the similarity of two chunks can be measured by comparing the corresponding vector representations through a similarity metric (e.g., cosine similarity). For embedding construction, Sentence-BERT <ref type="bibr" target="#b5">[6]</ref>, a modification of the original BERT model based on siamese and triplets networks, is employed to derive a semantically meaningful embedding for a given sentence/phrase. As such, a document chunk 𝑘 is associated with a set of terms 𝑊 𝑘 therein contained. Any term is described as 𝑤 = (𝑤 𝑙 , 𝑤 𝑑 , w ), where 𝑤 𝑙 is the label of the term (i.e., the lemma), 𝑤 𝑑 is a description of the term meaning taken from a reference dictionary/vocabulary (e.g., WordNet), and w is the corresponding vector-based representation according to Sentence-BERT, respectively. A document chunk 𝑘 has the form 𝑘 = (𝑠 𝑑 , 𝑘 𝑑 , k ), where 𝑠 𝑑 is the section of the document where the chunk occurs, 𝑘 𝑑 is the original textual content of the chunk, and k is the corresponding vector-based representation calculated as the mean of term vectors w with 𝑤 ∈ 𝑊 𝑘 . Embedding models have the capability to represent and compare the meaning of entire text blocks like document chunks. On such a target, context-aware embedding models fine-tuned on document similarity tasks, like Sentence-BERT, are appropriate. In the legal field, the phrase structure can be highly articulated, and some common terms can have a precise technical meaning when used in a court (e.g., citation, clemency, designation). Sentence-BERT can handle such a kind of situations, which may strongly deviate with respect to everyday conversations.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.3.">Concept Mining</head><p>Our solution to concept mining is called ASKE and it is based on zero-shot learning techniques and context-aware embedding models to enforce concept extraction <ref type="bibr" target="#b6">[7]</ref>.</p><p>Zero-shot learning is an unsupervised classification technique, characterized by the capability to enforce classification without requiring any pre-existing annotation of the considered documents. Initially, a seed knowledge is defined as a set of textual descriptions, each one featuring a concept of interest, namely a seed concept, to consider for classification. Typically, for a seed concept, a basic, gross-grained description is provided as a short text (e.g., one or two phrases) or a list of keywords. As an example, for a seed concept about banking contract, a corresponding textual description used for embedding is bank deposit, safe deposit box, bank credit opening, bank advance, bank account, bank discount. Further concepts are derived from seed ones during the extraction process, and they usually provide a more fine-grained description of the concept instances occurring in the document chunks. A concept 𝑐, either seed or derived, is defined as a pair 𝑐 = (𝑐 𝑙 , c</p><p>), where 𝑐 𝑙 is a label featuring the meaning of the concept expressed in a synthetic and human-understandable way, and c is a vector-based concept representation. Each concept 𝑐 is initially associated with the set of terms 𝑊 𝑐 extracted from the textual description of 𝑐. The vector concept c is built as the mean of the vectors of all the terms in 𝑊 𝑐 . Finally, the label 𝑐 𝑙 corresponds to the label 𝑤 𝑙 of the term 𝑤 ∈ 𝑊 𝑐 , whose vector representation w is closest to the concept vector c. Concept extraction is defined as a progressive, iterative process articulated in the following three steps:</p><p>Zero-Shot Classification. Given a set of concepts (i.e., the seed concepts at the beginning of the process), the document chunks are classified through zero-shot learning. A similarity measure 𝜎, e.g. cosine similarity, is calculated over any pair of embeddings between chunks and concepts. A document chunk 𝑘 is classified with the concept 𝑐 when the similarity value satisfies 𝜎 (𝑘, 𝑐) ≥ 𝛼, with 𝛼 defined as a similarity threshold configured in the system. The value of 𝛼 is empirically determined according to experimental results. In this paper, the value 𝛼 = 0.3 is employed in the proposed case-studies and experiments.</p><p>Terminology Enrichment. Given a document chunk 𝑘 classified with the concept 𝑐, the terms in 𝑊 𝑘 are exploited for enriching the term set 𝑊 𝑐 . The idea is that the initial description of the concept 𝑐 can become more detailed if we add terminology taken from chunks that are pertinent (i.e., classified) with 𝑐. This is done by summing, for each 𝑤 ∈ 𝑊 𝑘 , similarities 𝜎 (𝑤, 𝑐) and 𝜎(𝑤, 𝐾 𝑐 ), where 𝐾 𝑐 denotes the average embedding of document chunks classified with 𝑐. Terms 𝑤 ∈ 𝑊 𝑘 satisfying a system-defined 𝛽 similarity threshold are inserted in 𝑊 𝑐 .</p><p>Concept Derivation. By enriching the term set 𝑊 𝑐 , it is possible that more fine-grained concepts emerge from 𝑐, and they can be generated as new concepts. The discovery of possible new concepts emerging from 𝑐 is enforced by clustering the embedding vectors w of terms in 𝑊 𝑐 . The Affinity Propagation (AP) algorithm is adopted to this end, since it allows to detect the emergence of sub-groups of similar terms within 𝑊 𝑐 , without requiring to "a-priori define" the number of clusters to generate. A new concept 𝑐 ′ is created for each cluster returned by AP on the terms 𝑊 𝑐 of a concept 𝑐. A link is defined between a concept 𝑐 ′ and 𝑐 to denote that 𝑐 ′ is derived from 𝑐 and they are somehow similar/related in content. The concept 𝑐 is then updated since the terms in 𝑊 𝑐 can be changed due to enrichment. As a consequence, 𝑐 𝑙 and c are re-calculated.</p><p>The set of concepts obtained after derivation can trigger the execution of a new cycle based on the above three steps. New derived concepts can contribute to improve the classification of chunks with more fine-grained concepts. Further new concepts can be also discovered through a new execution of enrichment and derivation on the basis of a refined classification result. As such, concept extraction is characterized by a predefined endpoint condition based on a termination threshold. When the number of new concepts created in the derivation step is lower than the threshold, the concept extraction process is concluded. A concept graph providing a topic-based description of the underlying document corpus is finally stored.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Suggestion-Extraction for Document Drafting</head><p>This service supports the legal document building process, by implementing a human-in-theloop approach, where the document author, namely the legal actor, has to draft a document under preparation and she/he receives suggestions extracted by the ASKE engine according to the results of the knowledge-extraction service.</p><p>The Document Builder Environment is characterized by an editing area composed of a list of sections to be drafted, and a suggestion area where the author can retrieve pertinent chunks to be re-used and optionally changed in the currently-edited section (see Figure <ref type="figure" target="#fig_1">2</ref>). The author is mainly focused on the editing area with the aim to write and create the document. The structure of the document follows a template that the author has to choose within a predefined set. A document template is articulated in a list of sections 𝑠 1 , … , 𝑠 𝑛 . In the following, we call 𝑑 * the document under editing (i.e., the active document), and 𝑠 * the section on which the author is currently focused for drafting (i.e., the active section).</p><p>Two modalities are defined to provide suggestions to the author during document drafting:</p><p>• Search-by-Content, which enforces similarity-based retrieval of text suggestions for the section 𝑠 * by exploiting the chunks of ingested documents; • Search-by-Concept, which enforces concept-based retrieval of text suggestions for the section 𝑠 * by exploiting the concept graph. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1.">Search-by-Content</head><p>In the Search-by-Content modality, suggestions are represented by document chunks belonging to ingested documents that are considered most similar to the active one; the degree of similarity is computed taking into consideration document chunks belonging to a section preceding 𝑠 * .</p><p>The following steps are executed to determine the most pertinent suggestions for an active section 𝑠 * with text 𝑡 * . Detection of similar documents on previous sections. Call 𝑠 1 , … , 𝑠 𝑘 the sections of the active document 𝑑 * that precede the active section 𝑠 * . The ASKE engine is invoked to determine the ingested documents that contain similar chunks with respect to the texts 𝑡 1 , … , 𝑡 𝑘 of sections 𝑠 1 , … , 𝑠 𝑘 , respectively. For each section 𝑠 𝑖 of the active document 𝑑 * with 𝑖 ≤ 𝑘, the corresponding section content 𝑡 𝑖 is passed to the ASKE engine. The ASKE embedding model, namely the model used for knowledge extraction, is queried with the aim of extracting a vector representation ⃗ 𝑡 𝑖 of 𝑡 𝑖 2 . ASKE employs the cosine similarity function to compare ⃗ 𝑡 𝑖 against the document chunks of the 𝑠 𝑖 section belonging to any ingested document 𝑑. For each document 𝑑, a similarity degree 𝜎 𝑑 is defined by summing all the similarity values provided by the matching chunks 𝑘 𝑑 in 𝑠 𝑖 . For a section 𝑠 𝑖 , a set of similar documents 𝐷 𝑖 is returned as a result, where the similarity degree 𝜎 𝑑 of a document 𝑑 ∈ 𝐷 𝑖 is over a prefixed system-designed threshold. An overall set of similar documents 𝐷 = ⋂ 𝑖=𝑘 𝑖=1 𝐷 𝑖 is finally returned to use as input for the next step. Detection of pertinent chunks on the active section. Given the active section 𝑠 * , we retrieve a set of pertinent suggestions among the chunks of the similar documents in 𝐷. As a default behavior, the chunks of section 𝑠 * in the documents of 𝐷 are retrieved as suggestions to support the author in drafting the active section. Moreover, to obtain a refined set of suggestions, the author can start drafting the active section 𝑠 * by inserting an initial section content 𝑡 * . In this case, ASKE is invoked to derive the vector representation ⃗ 𝑡 * of 𝑡 * . The embeddings of chunks in documents of 𝐷 and section 𝑠 * are compared against ⃗ 𝑡 * . A ranked list of similar chunks are returned as suggestions to show in the right-hand panel of Figure <ref type="figure" target="#fig_1">2</ref> (descending order).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2.">Search-by-Concept</head><p>In the Search-by-Concept modality, suggestions represent the document chunks where an occurrence of a given concept of interest appears according to the classification results of ASKE.</p><p>In the suggestion area, the author can select a concept of interest 𝑐 * among a set of available ones, namely the set of concepts derived by ASKE during knowledge extraction from the ingested documents. The concept graph of ASKE is queried to retrieve the set of document chunks 𝐾 𝑐 * classified as 𝑐 * that are then returned as a result.</p><p>A further filtering option is provided in the suggestion area to enable the author to choose a target section of interest 𝑠 * . When a target section is selected, only the chunks of 𝐾 𝑐 * belonging to the section 𝑠 * in the ingested documents are finally shown as suggestions in the right-hand panel of Figure <ref type="figure" target="#fig_1">2</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Application to the Italian Legal Domain</head><p>The proposed service architecture has been applied to a legal case study in framework of the Next Generation UPP Italian project, aimed at providing artificial intelligence and advanced information management techniques for digital transformation of Italian legal processes and digital law in general.</p><p>We firstly introduce the dataset employed in the case study, explaining document preparation and segmentation. Then, we describe the scenario in which a legal actor, i.e. the judge, makes use of the document builder functionalities in order to generate a new judgement, with examples of Search-by-Content and Search-by-Concept text suggestions. The case study has been conducted on documents written in Italian; examples reported in the paper have been translated into English for the purpose of understanding. Also, for the sake of brevity, we report an excerpt of the text suggestions.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.1.">Dataset</head><p>The dataset consists in a corpus of 50 Italian case law decisions, retrieved from 12 different Courts located in Northern Italy. All the documents concern first degree civil law judgments regarding the matter of unfair competition. Such kind of documents comes in PDF files of different formats, depending on the court emitting it; plain text has been extracted from these files in order to be used in the document segmentation and the data preparation processes. For document segmentation, our legal partners of the project provided a document schema defined on the basis of the Rules of the Court (March 2022). This set of rules has the function of giving a rigid and predetermined structure to the introductory acts of the parties and to the decision of the Court. In particular, for example, Rule 74 states that "any judgment of the European Court of Human Rights must contain some basic information (such as the names of the parties, agents, lawyers or advisers of the parties and a report of the procedure followed) followed by some well-defined steps of the decision: the facts of the case; a summary of the arguments of the parties; the legal reasons; operational provisions". Similar criteria are also established for the drafting of the appeals of lawyers. Based on Rule 74 recommendations and related literature, our law project partners developed a document segmentation schema for Italian juridical judgements/case law decisions <ref type="bibr" target="#b2">[3]</ref>, which has been enforced in our proposed architecture, for document segmentation and for document building. The segmentation schema comprises five different structural sections with the following corresponding labels (identified directly by the domain experts): court and parties, providing information of the court, the panel of judges and the parties in the trial; background information, relating to the proceedings of the trial and to the reconstruction of the facts; claims and arguments, claims made by the plaintiffs and counterclaims made by the defendants; reasoning, for each decision about an individual claim; final decisions, for each individual claim. For the purposes of the case study, document segmentation has been performed manually by legal domain experts involved in the project.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.2.">Search-by-Content</head><p>In the Search-by-Content scenario, the legal actor is required to provide to the system basic knowledge regarding the case at hand, filling at least a section 𝑠 of the chosen document template, in order to receive suggestions on an active section 𝑠 * . These suggestions are retrieved from previous relevant case law decisions stored in the database of ingested documents and sorted based on the computed similarity scores. In particular, the legal actor could write down the sections 𝑠 1 and 𝑠 2 regarding background information and claims and arguments in order to receive suggestions for the reasoning section 𝑠 * . We exclude from the analysis the first section in the aforementioned annotation schema, court and parties, since it is auto-compiled by the software tools currently in use by legal actors and it includes general information that end up being not relevant for the search process.</p><p>As an example of document building, suppose to work on the generation of a new judgement regarding the counterfeiting of machinery, for which the following two sections have already been drafted:</p><p>Background Information: The Supreme Court confirmed what it had already ruled in previous decisions, ruling definitively on the validity of the XXXX patent and the intervening counterfeiting by YYYY; the plaintiff then filed a petition for the continuation of the suspended trial.</p><p>Claims and Arguments: XXXX seeks relief for the contraction of revenues afferent to the machinery, resulting from the presence of the machinery substitute of YYYY.</p><p>and the reasoning section is the active section to be generated. By invoking the Searchby-Content modality for drafting the reasoning section, the document builder retrieves as suggestions the most relevant document chunks of the reasoning sections of the judgements similar to the active counterfeiting of machinery document at hand. The top-2 retrieved document chunks ranked by similarity score are shown below.</p><p>(...) objected that such conduct challenged to XXXX would not be configurable since the corporate purpose of the two companies is different in the sense that XXXX is limited to carrying out only service activities for the repair of cleaning machinery, while YYYY carries out as its prominent activity the production and sale of machinery, which XXXX does not deal with, consequently, there could not be talk of unfair competition since the activity carried out by the defendant company is merely marketing and not production pertaining exclusively to the plaintiff company.</p><p>The plaintiff initially attached only pecuniary damage by referring to the royalties that the defendants would have to pay for the use of the trademark online, then to the cost of advertising investments for the launch of the products then not sold due to unfair competition and alternatively the retroversion of profits. Then referring to the fact that through unfair competition XXXX's turnover would have unlawfully grown by increasing its holding YYYY at the expense of ZZZZ with a reduction for the latter in the number of shares.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.3.">Search-by-Concept</head><p>In the Search-by-Concept scenario, it is possible to ask for text suggestions pertinent to one or more concepts 𝑐 * among the ones extracted by ASKE, which are representative of the actual content of the ingested documents. Besides concept(s) of pertinence, the legal actor can also choose the section(s) 𝑠 * to which retrieved text suggestions should belong to.</p><p>Below, we report an example of retrieved text suggestions pertinent to the concept public service located in background information section of ingested documents.</p><p>XXXX sued YYYY and ZZZZ, claiming: that it carries out funeral activities; that, by virtue of Article 5, paragraph 2, R.L. no. 19 of 2004, "in the event that the manager of public cemetery and necropsy services also carries out the funeral activity referred to in Article 13 of this law, corporate separation is mandatory... "; that, in implementation of this legislation, on <ref type="bibr">March 27, 2008</ref>  The company XXXX reported: that it was dedicated to the provision of logistics and material transport services for the "MotoGP" world championship races and other motorcycle sports; that it had entered into a contract with the organizing company YYYY, thus securing the status of "Express Supplier" and the ability to carry out all transports not only for YYYY, but also for the various teams and suppliers participating in the MotoGP championship; that in the structure of ZZZZ Mr. A. B. has always played a key role, following the conclusion of a project collaboration contract, renewed until 12/31/2008, by which he had been given the position of project manager organizing logistics for the MotoGP and other motorsport world championships, including the contract acquisition stages, with qualifications general director adviser and general director marketing and full decision-making and signing powers in the name and on behalf of ZZZZ; (...)</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.">Related Work</head><p>Work related to the legal document building issues regards legal document assembly, legal information retrieval, and document segmentation. The increasing interest towards AI applications in the legal field pushed a significant interest towards legal document assembly <ref type="bibr" target="#b1">[2]</ref> by relying on a predefined document structure. In <ref type="bibr" target="#b7">[8]</ref> the authors define a clear distinction between a document-oriented approach and an issue-oriented approach to document assembly. The former consists in the automatic selection of document components and the instantiation of user provided values for certain variables. The latter makes use of an explicit representation of legal rules, whose truths of the predicates are provided by the judge in the justification. These rules have been explicitly defined over the years in different forms, such as decision trees <ref type="bibr" target="#b8">[9]</ref> and interchange languages <ref type="bibr" target="#b3">[4]</ref>. Different collection of legal documents, such as Serbian case laws <ref type="bibr" target="#b9">[10]</ref> and the GDPR <ref type="bibr" target="#b10">[11]</ref>, have been modeled using these approaches.</p><p>Legal information retrieval (LIR) is the discipline that aims at extracting information from a corpus of legal documents, including case law decisions and legal codes. The digitization of these documents produced a significant boost to the field of LIR, with many methodologies being proposed for the problem: from boolean and rule-based approaches <ref type="bibr" target="#b11">[12]</ref>, from the exploitation of thesauri <ref type="bibr" target="#b12">[13]</ref> to ontologies <ref type="bibr" target="#b13">[14]</ref>. Recent studies <ref type="bibr" target="#b14">[15]</ref>  <ref type="bibr" target="#b15">[16]</ref> have been focused on the use of large language models (LLM) <ref type="bibr" target="#b16">[17]</ref>, that are capable of capturing the contextual representation of the text rather than simply focusing on the occurrences of certain terms. There have also been studies exploiting NLP techniques directly on Italian legal documents: ontology learning systems <ref type="bibr" target="#b17">[18]</ref>, article prediction <ref type="bibr" target="#b18">[19]</ref>, and fine-tuned LLMs <ref type="bibr" target="#b19">[20]</ref>.</p><p>The segmentation of legal documents is a task that can be performed either manually or automatically, with the respective advantages and disadvantages. Manual segmentation <ref type="bibr" target="#b20">[21]</ref>  <ref type="bibr" target="#b21">[22]</ref> relies on annotations performed by field experts on a corpus that do not contain many documents due to the complexity of the task, although the results are more precise than any automated system. Automatic approaches to text segmentation have been developed for general corpus in order to detect a shift in the topic discussed by subsequent sentences. In <ref type="bibr" target="#b22">[23]</ref> semantic related graphs, in which sentences are nodes and edges represents their relatedness, are exploited in order to find maximal cliques. More recent studies approach the task employing global <ref type="bibr" target="#b23">[24]</ref> and contextual <ref type="bibr" target="#b24">[25]</ref> vector representations of sentences. Automatic approaches have also been developed for the segmentation of legal documents, such as US Court Decisions <ref type="bibr" target="#b25">[26]</ref>, and Terms-of-Service documents <ref type="bibr" target="#b26">[27]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="7.">Concluding Remarks</head><p>In this paper, we presented a reference service architecture for knowledge-based legal document building. After discussing components and techniques composing the knowledge-extraction service and the suggestion-extraction service of the proposed architecture, some examples of the experimentation of a first legal document builder prototype on a corpus case study in the Italian legal domain are discussed. Ongoing work is devoted to the consolidation of the functionalities of the suggestion-extraction service of the document builder, with development of a graphical interface for the knowledge-based document building functionalities. Future work will be devoted to studying automated document segmentation techniques. In particular, we envision a clause-level classification model, which first maps each clause into a vector space exploiting a pre-trained contextual embedding model (e.g. Sentence-BERT <ref type="bibr" target="#b5">[6]</ref>) and then applies a classifier on the clause embedding vectors to determine the segment label of each clause. We plan experimentation of different classifiers, operating either on single clauses, clause sequences or clause networks.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: The proposed service architecture.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figure 2 :</head><label>2</label><figDesc>Figure 2: The Document Builder Environment.Editing area</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head></head><label></label><figDesc>the company YYYY had been established, a company that performs on behalf of the Municipality of Rimini the cemetery and necropsy activity as provided for by Art. 1 point 3 letter c of R.L. n. 19 of 2004 i.e. funeral transport for indigent people, funeral collection and transport on call of the Judicial Authority or for hygienic-sanitary needs, observation depot, morgue, mortuary sanitary service, necroscopic medicine activities; by notarial deed dated 30/09/2009, YYYY conferred the commercial activity concerning funeral services to ZZZZ. (...)</figDesc></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_0">The size of the document chunk is determined by the maximum number of tokens that can be processed by the considered embedding model.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_1">When the section content 𝑡 𝑖 exceeds the size of a document chunk, a tokenization mechanism is employed to spit 𝑡 𝑖 . For the sake of clarity, in the paper, we consider 𝑡 𝑖 as a single textual element with a corresponding vector-based representation ⃗ 𝑡 𝑖 .</note>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgments</head><p>This work is partially supported by i) the Next Generation UPP project within the PON programme of the Italian Ministry of Justice, and ii) the project SERICS (PE00000014) under the MUR NRRP funded by the EU -NextGenerationEU.</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Algorithm-assisted decision-making in the public sector: framing the issues using administrative law rules governing discretionary power</title>
		<author>
			<persName><forename type="first">M</forename><surname>Oswald</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Philosophical Transactions of the Royal Society A: Mathematical, Physical and Engineering Sciences</title>
		<imprint>
			<biblScope unit="volume">376</biblScope>
			<biblScope unit="page">20170359</biblScope>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">A theory construction approach to legal document assembly</title>
		<author>
			<persName><forename type="first">T</forename><surname>Gordon</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Pre-Proceedings of the Third International Conference on Logic, Informatics, and Law</title>
				<imprint>
			<publisher>Citeseer</publisher>
			<date type="published" when="1989">1989</date>
			<biblScope unit="page" from="485" to="498" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">A rule 74 for italian judges and lawyers</title>
		<author>
			<persName><forename type="first">G</forename><surname>Pinotti</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Santosuosso</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Fazio</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Advances in Conceptual Modeling: ER 2022 Workshops</title>
				<meeting><address><addrLine>CMLS, EmpER, and JUSMOD, Hyderabad, India</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2022">October 17-20, 2022. 2023</date>
			<biblScope unit="page" from="112" to="121" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Legalruleml: Xml-based rules and norms</title>
		<author>
			<persName><forename type="first">M</forename><surname>Palmirani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Governatori</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Rotolo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Tabet</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Boley</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Paschke</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">RuleML America</title>
		<imprint>
			<biblScope unit="volume">7018</biblScope>
			<biblScope unit="page" from="298" to="312" />
			<date type="published" when="2011">2011</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<monogr>
		<title level="m" type="main">Text segmentation based on similarity between words</title>
		<author>
			<persName><forename type="first">H</forename><surname>Kozima</surname></persName>
		</author>
		<idno>CoRR cmp-lg/9601005</idno>
		<ptr target="http://arxiv.org/abs/cmp-lg/9601005" />
		<imprint>
			<date type="published" when="1996">1996</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<monogr>
		<title level="m" type="main">Sentence-bert: Sentence embeddings using siamese bert-networks</title>
		<author>
			<persName><forename type="first">N</forename><surname>Reimers</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Gurevych</surname></persName>
		</author>
		<idno>CoRR abs/1908.10084</idno>
		<ptr target="http://arxiv.org/abs/1908.10084.arXiv:1908.10084" />
		<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Knowledge-Based Legal Document Retrieval: A Case Study on Italian Civil Court Decisions</title>
		<author>
			<persName><forename type="first">V</forename><surname>Bellandi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Castano</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Ceravolo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Damiani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Ferrara</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Montanelli</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Picascia</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Polimeno</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Riva</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. of the 1st Int. Knowledge Management for Law Workshop (KM4LAW)</title>
				<meeting>of the 1st Int. Knowledge Management for Law Workshop (KM4LAW)<address><addrLine>Bozen-Bolzano, Italy</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">An issue-oriented approach to judicial document assembly</title>
		<author>
			<persName><forename type="first">L</forename><forename type="middle">K</forename><surname>Branting</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 4th international conference on Artificial intelligence and law</title>
				<meeting>the 4th international conference on Artificial intelligence and law</meeting>
		<imprint>
			<date type="published" when="1993">1993</date>
			<biblScope unit="page" from="228" to="235" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Artificial intelligence and document assembly</title>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">B</forename><surname>Evans</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Law Prac. Mgmt</title>
		<imprint>
			<biblScope unit="volume">16</biblScope>
			<biblScope unit="page">18</biblScope>
			<date type="published" when="1990">1990</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<monogr>
		<author>
			<persName><forename type="first">M</forename><surname>Marković</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Gostojić</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2009.06611</idno>
		<title level="m">Knowledge-based legal document assembly</title>
				<imprint>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">The dapreco knowledge base: representing the gdpr in legalruleml</title>
		<author>
			<persName><forename type="first">L</forename><surname>Robaldo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Bartolini</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Lenzini</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of The 12th Language Resources and Evaluation Conference</title>
				<meeting>The 12th Language Resources and Evaluation Conference</meeting>
		<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="5688" to="5697" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Legal machine-learning analysis: First steps towards a.i. assisted legal research</title>
		<author>
			<persName><forename type="first">W</forename><forename type="middle">Y</forename><surname>Mok</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">R</forename><surname>Mok</surname></persName>
		</author>
		<idno type="DOI">10.1145/3322640.3326737</idno>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Seventeenth International Conference on Artificial Intelligence and Law, ICAIL &apos;19</title>
				<meeting>the Seventeenth International Conference on Artificial Intelligence and Law, ICAIL &apos;19<address><addrLine>New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>Association for Computing Machinery</publisher>
			<date type="published" when="2019">2019</date>
			<biblScope unit="page" from="266" to="267" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Thesaurus-based retrieval of case law</title>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">C</forename><surname>Klein</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Van Steenbergen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><forename type="middle">M</forename><surname>Uijttenbroek</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">R</forename><surname>Lodder</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Van Harmelen</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Frontiers in Artificial Intelligence and Applications</title>
		<imprint>
			<biblScope unit="volume">152</biblScope>
			<biblScope unit="page">61</biblScope>
			<date type="published" when="2006">2006</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">Crime knowledge extraction: An ontologydriven approach for detecting abstract terms in case law decisions</title>
		<author>
			<persName><forename type="first">S</forename><surname>Castano</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Ferrara</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Falduti</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Montanelli</surname></persName>
		</author>
		<idno type="DOI">10.1145/3322640.3326730</idno>
	</analytic>
	<monogr>
		<title level="m">Proc. of the 17th Int Conference on Artificial Intelligence and Law, ICAIL &apos;19</title>
				<meeting>of the 17th Int Conference on Artificial Intelligence and Law, ICAIL &apos;19<address><addrLine>New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2019">2019</date>
			<biblScope unit="page" from="179" to="183" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">BERT_LF: A similar case retrieval method based on legal facts</title>
		<author>
			<persName><forename type="first">W</forename><surname>Hu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Zhao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><surname>Zhao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Sun</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Hu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Guo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Cui</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Ma</surname></persName>
		</author>
		<idno type="DOI">10.1155/2022/2511147</idno>
		<ptr target="https://doi.org/10.1155/2022/2511147.doi:10.1155/2022/2511147" />
	</analytic>
	<monogr>
		<title level="j">Wireless Communications and Mobile Computing</title>
		<imprint>
			<biblScope unit="page" from="1" to="9" />
			<date type="published" when="2022">2022. 2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">Context-Aware Knowledge Extraction from Legal Documents through Zero-Shot Classification</title>
		<author>
			<persName><forename type="first">A</forename><surname>Ferrara</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Picascia</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Riva</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. of the 1st ER Int. Workshop on Digital Justice, Digital Law, and Conceptual Modeling (JUSMOD22)</title>
				<meeting>of the 1st ER Int. Workshop on Digital Justice, Digital Law, and Conceptual Modeling (JUSMOD22)<address><addrLine>Hyderabad, India</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2022">2022</date>
			<biblScope unit="page" from="81" to="90" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<monogr>
		<author>
			<persName><forename type="first">J</forename><surname>Devlin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M.-W</forename><surname>Chang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Lee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Toutanova</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1810.04805</idno>
		<title level="m">Bert: Pre-training of deep bidirectional transformers for language understanding</title>
				<imprint>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">Ontology learning from italian legal texts</title>
		<author>
			<persName><forename type="first">A</forename><surname>Lenci</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Montemagni</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Pirrelli</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Venturi</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Law, Ontologies and the Semantic Web</title>
				<imprint>
			<publisher>IOS Press</publisher>
			<date type="published" when="2009">2009</date>
			<biblScope unit="page" from="75" to="94" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<analytic>
		<title level="a" type="main">Unsupervised law article mining based on deep pre-trained language representation models with application to the italian civil code</title>
		<author>
			<persName><forename type="first">A</forename><surname>Tagarelli</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Simeri</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Artificial Intelligence and Law</title>
		<imprint>
			<biblScope unit="page" from="1" to="57" />
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<monogr>
		<author>
			<persName><forename type="first">D</forename><surname>Licari</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Comandè</surname></persName>
		</author>
		<title level="m">Italian-legal-bert: A pre-trained transformer language model for italian law</title>
				<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<monogr>
		<author>
			<persName><forename type="first">P</forename><surname>Kalamkar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Tiwari</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Agarwal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Karn</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Gupta</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Raghavan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Modi</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2201.13125</idno>
		<title level="m">Corpus for automatic structuring of legal documents</title>
				<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b21">
	<analytic>
		<title level="a" type="main">A decade of legal argumentation mining: Datasets and approaches</title>
		<author>
			<persName><forename type="first">G</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Nulty</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Lillis</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Natural Language Processing and Information Systems: 27th International Conference on Applications of Natural Language to Information Systems, NLDB 2022</title>
				<meeting><address><addrLine>Valencia, Spain</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2022">June 15-17, 2022. 2022</date>
			<biblScope unit="page" from="240" to="252" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b22">
	<analytic>
		<title level="a" type="main">Unsupervised text segmentation using semantic relatedness graphs</title>
		<author>
			<persName><forename type="first">G</forename><surname>Glavaš</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Nanni</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">P</forename><surname>Ponzetto</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Fifth Joint Conference on Lexical and Computational Semantics, Association for Computational Linguistics</title>
				<meeting>the Fifth Joint Conference on Lexical and Computational Semantics, Association for Computational Linguistics</meeting>
		<imprint>
			<date type="published" when="2016">2016</date>
			<biblScope unit="page" from="125" to="130" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b23">
	<monogr>
		<author>
			<persName><forename type="first">O</forename><surname>Koshorek</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Cohen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Mor</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Rotman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Berant</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1803.09337</idno>
		<title level="m">Text segmentation as a supervised learning task</title>
				<imprint>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b24">
	<monogr>
		<author>
			<persName><forename type="first">A</forename><surname>Solbiati</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Heffernan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Damaskinos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Poddar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Modi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Cali</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2106.12978</idno>
		<title level="m">Unsupervised topic segmentation of meetings with bert embeddings</title>
				<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b25">
	<analytic>
		<title level="a" type="main">Segmenting us court decisions into functional and issue specific parts</title>
		<author>
			<persName><forename type="first">J</forename><surname>Savelka</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><forename type="middle">D</forename><surname>Ashley</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">JURIX</title>
		<imprint>
			<biblScope unit="page" from="111" to="120" />
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b26">
	<analytic>
		<title level="a" type="main">Structural text segmentation of legal documents</title>
		<author>
			<persName><forename type="first">D</forename><surname>Aumiller</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Almasian</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Lackner</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Gertz</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Eighteenth International Conference on Artificial Intelligence and Law</title>
				<meeting>the Eighteenth International Conference on Artificial Intelligence and Law</meeting>
		<imprint>
			<date type="published" when="2021">2021</date>
			<biblScope unit="page" from="2" to="11" />
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
