<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">From Knowledge Management to Intelligence Engineering -A practical approach to building AI inside the law-firm using open-source Large Language Models</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author role="corresp">
							<persName><forename type="first">Uwais</forename><surname>Iqbal</surname></persName>
							<email>uwais.iqbal@simplexico.ai</email>
							<affiliation key="aff0">
								<orgName type="institution">Simplexico Limited</orgName>
								<address>
									<addrLine>33 Carnarvon Road</addrLine>
									<postCode>E10 6DW</postCode>
									<settlement>London</settlement>
									<country key="GB">United Kingdom</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">From Knowledge Management to Intelligence Engineering -A practical approach to building AI inside the law-firm using open-source Large Language Models</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">35688CC78457E6BE665F2BC76D2135D8</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-06-19T14:57+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Large Language Models</term>
					<term>Legal AI</term>
					<term>Natural Language Processing</term>
					<term>Generative AI</term>
					<term>Foundational Language Model</term>
					<term>Knowledge Management</term>
					<term>Intelligence Engineering</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Open-source foundational language models unlock a new opportunity for building AI inside the law firm. In this paper, we explore the different options in the AI build vs buy equation facing law firms and outline four postures across the spectrum of building AI. We motivate a particular posture that leverages open-source foundational models in a way that both mitigates data privacy and security concerns, while enabling customisation of these models with internal data. We explore the different ways in which these models can be fine-tuned and present a novel addition of intelligence engineering to the traditional knowledge management process that involves instruction fine-tuning language models to infinitely scale access to explicit knowledge. We provide a practical demonstration of this technical approach with a proof of concept using an open-source foundational model based on the GPT-3 architecture and an open-source dataset of contracts. We also provide a qualitative analysis of results.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>The recent wave of hype surrounding Generative AI and Large Language Models (LLMs) has captured the collective imagination of the legal industry. Following the launch of ChatGPT <ref type="bibr" target="#b0">[1]</ref>, GPT-3, and GPT-4, <ref type="bibr" target="#b1">[2]</ref>  <ref type="bibr" target="#b2">[3]</ref> there has been considerable interest in exploring how this new wave of AI could bring transformation to the legal industry. Reports and predictions estimate that up to 40% of legal work could be displaced by AI systems <ref type="bibr" target="#b3">[4]</ref>.</p><p>Despite the newfound enthusiasm of the legal sector around AI, the practical constraints around actually doing technology in the legal sector still remain as relevant as ever. In a conservative and risk-averse industry like legal, concerns around data privacy, security and confidentiality dictate the pace of adoption, transformation and innovation <ref type="bibr" target="#b4">[5]</ref>. There is a balance to be struck in driving innovation with technology in the sector while staying faithful to these justified concerns.</p><p>The scope of this paper is to explore whether Large Language Models could be used in law firms, and if so, how this may be realised technically. This paper attempts to provide a practical middle path to the future of AI in legal by facilitating the recent advancements around Generative AI and LLMs in a way that mitigates concerns around data privacy, security and confidentiality. In particular, we explore how open-source foundational language models can be brought inside the law firm to avoid the issues raised by sharing sensitive data with external vendors.</p><p>Rather than exploring the largest most performant models which demonstrate the state-of-the-art in terms of performance, we take a pragmatic approach and instead explore the state-of-the-feasible through models with parameter sizes that can be run on commodity hardware to set the floor. We test this approach using an open-source contracts dataset as a proxy for internal data within a firm and share results on this experiment.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">AI Build vs Buy -The Options</head><p>The build vs buy distinction is actually more of a spectrum when it comes to AI. With traditional software engineering there were two extremes -either build the product yourself or go buy it from the market. Developing software is a relatively simple process that doesn't involve lots of moving parts.</p><p>AI is fundamentally different. With AI there are a number of core components involved across three different contexts. There are three main components: a) the underlying code for the algorithm, b) the data and c) the compute resources where all of these elements are combined to create the model. The three parts come together to give rise to the derived product; the model. There are also three different stages to creating AI which are relevant to the different build vs buy options:</p><p>1. Pre-training -In the pre-training stage, the code for the algorithm is used with data in a compute resource to train a model. 2. Fine-tuning -In the fine-tuning stage, the pretrained model is further refined with additional data using the code for the algorithm in a compute resource to create a more refined model focused for a particular domain or for a particular task. 3. Serving -In the serving stage, the model is served in a compute environment and packaged in an API so it can be called on by other software services and the AI capability can be consumed.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1.">Buying AI</head><p>Buying AI can be understood as purchasing a specific AI-enabled application from a vendor. Just as one might purchase an electrical appliance like a toaster or a kettle from a supermarket. These AI-enabled applications perform particular tasks and fulfill a certain defined set of needs. Examples of such AI-enabled applications in the legal sector include offerings from vendors such as Harvey, Spellbook, Kira etc.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2.">The AI Build Spectrum</head><p>Building AI can be understood as going one level higher to interact with the underlying technology through APIs or actual code. AI can be built through four different postures across a spectrum that spans from one extreme of a consumer posture to another extreme of a creator posture. The four postures are as follows:</p><p>1. The Consumer Posture 2. The Consumer Customiser Posture 3. The Creator Customiser Posture 4. The Creator Posture Figure <ref type="figure" target="#fig_0">1</ref> outlines a visual representation of the postures across the AI build spectrum and how the different components of data, code, compute and model are distributed across open-source, vendor and internal management.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2.1.">The Consumer Posture</head><p>In the Consumer Posture, an organisation acts as a consumer of AI and the pre-training, fine-tuning and serving stages are all managed by a vendor. The Consumer Posture enables an organisation to just get on with consuming and integrating AI into their applications. The vendor's job is to worry about training and serving the AI system. The organisation just acts as a consumer and can build applications around it. A typical use case with The Consumer Posture is easy to get started with and no AI skills are needed. The services are all managed by the vendor so there are no technical or infrastructure concerns to worry about. However, under the Consumer Posture there is no control over the AI process so there is no effective way to mitigate bias and risk. The models available through such services tend to be too generic and general purpose to be effective and useful for specialised tasks in the legal domain.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>THE AI BUILD SPECTRUM DIAGRAM KEY</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Open</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2.2.">The Consumer Customiser Posture</head><p>The Consumer Customiser Posture is similar to the Consumer Posture. The vendor provides fine-tuning as a service offering and allows an organisation to customise and fine-tune the underlying model with their own data. The vendor still takes care of everything from pre-training, fine-tuning and serving. A typical use with the Consumer Customiser Posture would involve fine-tuning a model with an organisation's internal data through an API service like Open AI so the model is more familiar with specific domain language.</p><p>Much like the Consumer Posture, the Consumer Customiser Posture is easy to get started with and is offered as a managed service meaning all the technical infrastructure work is taken care of by the vendor. However, this posture involves sharing internal data with a vendor so raises risks and concerns around data privacy and security. It also raises concerns and questions around the IP ownership of the fine-tuned model managed by the vendor.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2.3.">The Creator Customiser Posture</head><p>In the Creator Customiser Posture, the posture changes from one of a consumer to a creator. In this posture, open-source pre-trained models are utilised for commercial purposes. The open-source base models are brought inside an organisation and further fine-tuned on a combination of open-source and internal data. Under this posture the pre-training is performed by another party who then makes their pre-trained model publicly available under an open-source license. The fine-tuning and serving phases are managed internally so infrastructure is needed to facilitate the compute resources in the form of either on-premise or cloud resources.</p><p>A typical use case for the Creator Customiser Posture would involve bringing an open-source foundational model in house and fine-tuning it on internal data and serving it internally for internal applications.</p><p>The Creator Customiser Posture is attractive in that it doesn't require the sharing of data with a third party. There is more control over the AI creation process so it becomes easier to mitigate for bias and risk as well as put in place controls for safety and governance. The IP of finetuned models is owned by the organisation. However, this posture requires infrastructure to be managed and maintained internally for the compute resources in the fine-tuning and serving phases. It also requires some level of specialised skill set around AI.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2.4.">The Creator Posture</head><p>The Creator Posture goes a step further than the Creator Customiser Posture. Instead of relying on a third party to perform the pre-training phase to create the pre-trained model, in the Creator Posture the pre-training phase is brought inside the organisation. There is no dependence on a third party and all three phases of pre-training, finetuning and serving are managed internally.</p><p>The Creator Posture is the polar opposite of the Consumer Posture and sits on the other end of the AI build spectrum. There is complete control over the process of creating AI so there the ability to have full provenance of data. There is a complete transparency and the digital footprint of the models can be traced back to their origins. Mitigating for bias, risk and putting in place controls for safety and governance is made much more accessible.</p><p>Just like the Creator Customiser Posture, a specialised skillset around AI is needed and the infrastructure for pretraining, fine-tuning and serving all have to be managed internally.</p><p>These four postures present the different ways an organisation can go about building AI. The options of buying AI and the four postures of building AI are not mutually exclusive. An organisation can develop a complimentary strategy across buying and the four postures of building to explore general purpose AI applications as well as building their own AI systems with internal data.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">The State of Affairs -Open-source Large Language Models</head><p>Over recent months, there has been something of a revolutionary movement in the open-source community.</p><p>Since the release of ChatGPT, the open-source community have been active in replicating the capabilities of closed-source models. The data and hardware requirements for creating foundational language models were previously significant barriers to entry. Only organisations in a privileged position could create such foundational language models and then proceed onto the later phase of development. The ability to create such models were in the hands of the few. The recent wave of activity in the open-source community has significantly changed this dynamic. With the release of open-source foundational model suites like LLaMa <ref type="bibr" target="#b5">[6]</ref>, pythia <ref type="bibr" target="#b6">[7]</ref>, Cerebras-GPT <ref type="bibr" target="#b7">[8]</ref>, StabilityLM <ref type="bibr" target="#b8">[9]</ref>, and MPT-7B <ref type="bibr" target="#b9">[10]</ref> these base models are now publicly accessible. The ability to customise models and continue through phases of development are in the hands of everyone. While some of the open-source base models have been released under licenses only permitting academic use, some models are available for commercial use. Table 1 outlines the licenses for some of these open-source foundational models and whether they are available for commercial use. Open-source foundational language models with their parameter sizes and licenses</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Name</head><p>The allure of open-source models for law firms is that they can be used with the Creator Customiser Posture. These models can be brought inside the law firm and developed further on internal data mitigating the risks around data privacy and security while still enabling access to cutting-edge technology. These developments around open-source foundational language models now mean that AI can be brought inside the law firm to build AI systems that power use cases inside the law firm.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Layers of Fine-tuning -Customising Open-source Language Models</head><p>An open-source foundational language model can be further fine-tuned to customise the model with domainspecific and internal data. There are three distinct layers of fine-tuning that can be performed with language models.</p><p>1. Unsupervised Fine-Tuning -With a domainspecific corpus of raw unstructured text, the language model can be fine-tuned to learn the particular nuances and quirks of legal language. This can be made even more specific by focusing on a particular practice area or a particular area of law. The data requirements for unsupervised finetuning are not restrictive. All that is needed is a corpus of raw text documents which should be relatively easy to find within a law firm. The data does not even need to be structured as only the raw text is needed. 2. Instruction Response Fine-tuning -Recasting structured data into tasks consisting of instruction response pairs results in language models being able to generalise across unseen tasks really well <ref type="bibr" target="#b10">[11]</ref>. Table <ref type="table">4</ref> in the appendix shows some examples of instruction response pairs. The base language model can be fine-tuned on pairs of instructions and responses to create an ability to follow particular instructions and commands <ref type="bibr" target="#b11">[12]</ref>. This creates an ability for the model to generalise across different tasks by learning how to complete the instruction on unseen data in a zero-shot fashion. There are some requirements around the data for instruction response fine-tuning since the data has to be structured from raw text into instruction response pairs. 3. Reinforcement Human Learning Feedback (RHLF) -The third phase involves creating a more human-like interface to the language model by using Reinforcement Learning to teach the model how to converse as a human <ref type="bibr" target="#b12">[13]</ref>. This creates a conversational layer with the language model that allows it to be interacted with as a chat bot.</p><p>The first two levels of fine-tuning create domainspecific functionality while the last is more aesthetic. These layers of fine-tuning can be performed with a combination of internal and open-source datasets for maximum learning. There are already a growing number of such open-source datasets for instruction response fine-tuning and RHLF available for commercial use <ref type="bibr" target="#b13">[14]</ref>. Open-source datasets can be combined with internal datasets and stacked in a modular fashion to create the desired intelligence capabilities within the language model.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1.">From Knowledge Management to Intelligence Engineering</head><p>The Instruction Response fine-tuning approach is of particular importance for law firms. Knowledge management (KM) can be defined as the "tools, techniques, and strategies to retain, analyse, organise, improve, and share business experience" <ref type="bibr" target="#b14">[15]</ref>. Within the context of a law firm, knowledge management involves "a firm's ability to identify, capture, and leverage the internal knowledge of individuals" to "enhance the ability of all law firm staff to create and share knowledge across the firm and to provide excellent client services and to compete in an increasingly aggressive professional legal services environment" <ref type="bibr" target="#b15">[16]</ref>. Knowledge management is based on three fundamental concepts: a) data, b) information, and c) knowledge <ref type="bibr" target="#b16">[17]</ref>. Data is understood as the raw resource without context. Information is understood as data with context that is able to provide value. Knowledge is understood as information combined with understanding and capability. The distinction between knowledge and information can be clarified as "knowledge being a personal subjective process emerging from previous experiences, while information is objective data about the environment" <ref type="bibr" target="#b17">[18]</ref>. Knowledge lives in the minds of people and is anthropomorphic while information is not.</p><p>Knowledge can further be broken down into two main types: a) Tacit Knowledge, and b) Explicit Knowledge <ref type="bibr" target="#b14">[15]</ref>. Tacit knowledge refers to personal knowledge embedded in individual experience while explicit knowledge refers to tacit knowledge that has been documented. One of the challenges in knowledge management is the difficulty in capturing tacit knowledge. One of the key functions of a knowledge management strategy is to make the tacit explicit so that it can be easily transferred and communicated from one individual to another.</p><p>We can introduce a fourth related concept of intelligence to the fundamental concepts of knowledge management. Intelligence can be defined as the ability to acquire knowledge and skills. As such, intelligence can be possessed by humans as natural intelligence and by machines in the form of artificial intelligence (AI). By including intelligence within the fundamental concepts of knowledge management, AI can be adopted to achieving and enhance the objectives of knowledge management.</p><p>Knowledge management within law firms has traditionally focused on capturing the tacit knowledge from the minds of highly skilled and experienced individuals into explicit knowledge in the form of written content. The explicit knowledge captured in content allows this expertise to be shared with and accessed by other individuals in the firm. However, one of the practical challenges with knowledge management is how to make this explicit knowledge easily accessible, retrievable and consumable to other individuals within the organisation.</p><p>Traditionally knowledge management seeks to take tacit knowledge from the mind of a human and capture it as explicit knowledge in the form of written content so that it can be consumed by another human. The process of knowledge transfer goes from human (tacit knowledge) to content (explicit knowledge) to human (tacit knowledge). With the advent of AI, large language models and instruction response fine-tuning, we propose intelligence engineering as an additional step in the process. Rather than going from human to content to human, we propose introducing a machine into the process; going from human (tacit knowledge) to content (explicit knowledge) to machine (intelligence) to human (tacit knowledge).</p><p>Practically, this would involve much of the same processes around knowledge management as before but with a few additional steps. Tacit knowledge from the mind of a human can be captured as explicit knowledge in the form of an instruction response dataset. Existing explicit knowledge content can easily be restructured into an instruction response format. Such a dataset can be used to fine-tune a large language model with the Creator Customiser posture to create protected and privileged intelligence within a law firm. Then, individuals in the firm can interface with the large language model to retrieve, access and query the intelligence captured. This additional step of intelligence engineering removes existing bottlenecks around accessing and retrieving explicit knowledge. In the context of knowledge management within a law firm, a large language model effectively creates infinite scale in providing access to explicit knowledge. While explicit knowledge has always been static in the form of content, large language models transform this content to intelligence that is dynamic, scalable and easily accessible.</p><p>We explore and evaluate this approach by layering unsupervised fine-tuning and instruction fine-tuning with the Creator Customiser Posture using an open-source foundational language model and a publicly available dataset of contracts.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Experimental Results and Findings</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.1.">Experimental Setup</head><p>To demonstrate how an open-source foundational language model can be leveraged inside a law firm we practically demonstrate how the Creator Customiser Posture can be used with layered unsupervised and instruction response fine-tuning. We first select a base foundational model. The Cerebras-GPT model suite contains 7 GPT-3 models ranging from 111M up to 13B in parameter size <ref type="bibr" target="#b7">[8]</ref>. The models were created by training on The Pile dataset <ref type="bibr" target="#b18">[19]</ref>. These models are licensed under the Apache 2.0 license and are available for commercial use.</p><p>Taking a pragmatic approach we focus on the smaller parameter size variants of the Cerebras-GPT model suite.</p><p>In particular we work with the 590M parameter variant. Our goal is to provide a pragmatic demonstration of the approach rather than optimising for performance.</p><p>In order to proxy law firm internal data, we use two open-source datasets related to contracts. Both are in the public domain and are available under the CC-BY 4.0 license. These datasets are as follows:</p><p>1. Contract Understanding Atticus Dataset (CUAD) <ref type="bibr" target="#b19">[20]</ref> -The CUAD dataset consists of 510 commercial legal contracts with over 13,000+ labels that have been manually labelled under the supervision of experienced lawyers. The annotations identify legal clauses that are considered important in contract review in connection with a corporate transaction, including mergers acquisitions, etc. 2. Merger Agreement Understanding Dataset (MAUD) <ref type="bibr" target="#b20">[21]</ref> -The MAUD dataset consists of 152 merger agreements with over 47,000+ labels that have been manually labeled under the supervision of experienced lawyers to identify 92 questions in each agreement used by the 2021 American Bar Association (ABA) Public Target Deal Points Study.</p><p>We combine the raw text from the contracts and agreements in the CUAD and MAUD datasets to create a dataset for unsupervised fine-tuning. The resulting combined dataset consists of 662 documents with 1.8M tokens. This dataset acts as a proxy for documents and unstructured text which may sit inside a document management system at a law firm.</p><p>Taking inspiration from FLAN <ref type="bibr" target="#b10">[11]</ref>, we mine the labels in the CUAD dataset to produce a collection of 8,000+ instruction response pairs. The pairs span a number of legal specific tasks including drafting, classification and extraction. This dataset acts as a proxy for internal data contained in a precedent bank or within explicit knowledge content and practice notes. With a little mining, any internal structured textual data can be reconstructed in the form of instruction response pairs for use with large language models. Table <ref type="table">4</ref> in the appendix shows examples of these instruction response pairs created from the CUAD dataset. We perform the experiments on commodity hardware in the form of a single Nvidia A100 GPU with 80GB of RAM.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.2.">Results</head><p>Using the base model, we perform two stages of finetuning:</p><p>1. Unsupervised Fine-Tuning -Using the dataset of 1.8M tokens created by combining the raw text from the CUAD and MAUD datasets, we fine-tune the model in an unsupervised manner. Through this stage the model learns the nuances of legal language. 2. Instruction Fine-Tuning -Using the 8,000+ instruction response pairs created from mining the CUAD data, we further fine-tune the model with these pairs. The model learns how to perform these legal specific tasks.</p><p>For both datasets we use an 80:20 split to create the training and testing sets. Using the 590M parameter variant from the Cerebras-GPT model suite as the base model, we run the unsupervised fine-tuning with the combined text from the CUAD and MAUD datasets.</p><p>We compare the quality of the fine-tuned language models by using perplexity as an intrinsic evaluation metric. The perplexity score captures the average number of words that can be encoded. In more concrete terms, a perplexity score of 4 means that when trying to guess the next word, the model is as confused as if it had to pick between 4 different words. A lower perplexity score means that the language model is more precise at predicting words and is better.</p><p>We report the following results after 3 epochs of unsupervised fine-tuning: Perplexity on Test Set (Before fine-tuning) </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Table 2</head><p>Results after performing unsupervised fine-tuning for 3 epochs using the combined text from the CUAD and MAUD datasets</p><p>We take the fine-tuned model and use it to further perform instruction fine-tuning using the 8,000+ instruction response pairs mined from the CUAD dataset. We report the following results after 3 epochs of instruction response fine-tuning: Perplexity on Test Set (Before fine-tuning) </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Table 3</head><p>Results after performing instruction response fine-tuning for 3 epochs using the 8,000+ instruction response pairs mined from the CUAD dataset.</p><p>Table <ref type="table">5</ref> in the appendix details some sample outputs from the fine-tuned model with their associated prompts.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.3.">Findings</head><p>The significant difference in perplexity scores before and after fine-tuning indicate that massive amounts of domain-specific data isn't needed to effectively fine-tune open-source models. This also demonstrates that finetuning can be performed with reasonable volumes of data at a reasonable cost in a reasonable time frame. We can estimate order of magnitude data requirements for the different fine-tuning layers. For unsupervised finetuning on the order of hundreds of documents are needed while for instruction response fine-tuning on the order of thousands of instruction response pairs are needed.</p><p>We posit that since the language in the domain is more specific and standardised, it is easier for the model to learn the nuances of legal language. As opposed to generic pre-training datasets like Pile <ref type="bibr" target="#b18">[19]</ref>, there is less variability in the language so it is easier for the model to learn. Further work is needed to evaluate and quantify the effectiveness of such fine-tuned large language models by way of extrinsic evaluation to quantify performance on downstream tasks. Further investigation is also needed to understand the effect of parameter size on the performance of fine-tuning on legal domain specific language and tasks.</p><p>The results also indicate that fine-tuning open-source foundational models is feasible and practical from a variety of operational perspectives: data, cost and time. The data requirements are not unreasonable -most law firms have access to thousands of documents that can be used for unsupervised fine-tuning. Following our approach, internal knowledge content, structured databases and practice notes can be mined to create instruction re-sponse pairs for instruction response fine-tuning. The cost of running such fine-tuning experiments is very cheap which means that the operational expense of experimentation is not a barrier to innovation. The time requirements for performing fine-tuning are not prohibitive to limit rapid prototyping and iterative development.</p><p>The approach outlined by taking the Creator Customiser Posture with open-source foundational models and performing unsupervised and instruction-response fine-tuning is a methodology that can be adopted for intelligence engineering to create an intelligence layer to make explicit knowledge more accessible so it can be readily be consumed by other humans. In effect, this methodology can immediately unlock value by advancing knowledge management objectives within law firms through intelligence engineering.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.">Conclusion</head><p>We have introduced the various options available when it comes to the buy vs build question for AI. We have outlined the four postures across the AI build spectrum and demonstrated that the Creator Customiser posture is the most appealing for law firms looking to leverage internal data while mitigating risks around data privacy and security. We have assessed the opportunities open-source foundational language models present and outlined the various ways in which these models can be further refined on internal data. We have also presented a practical way and method in which large language models can be introduced into knowledge management workflows in law firms through intelligence engineering.</p><p>We have carried out an early evaluation of this approach using an open-source foundational language model and an open-source contract dataset. While the experimental analysis in this paper is limited to a qualitative analysis of results, early findings indicate the value and technical feasibility of this approach in unlocking opportunities for building AI within the law firm. Further experimentation is required to quantitatively evaluate this approach on downstream tasks to provide an objective assessment of performance. The impact of larger parameter sizes on downstream performance is an area for further investigation.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>A. Examples of Instruction Response Data</head><p>The following table provides some examples from the instruction response dataset that was created by mining the CUAD dataset.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Instruction</head><p>Response Instruction Type Draft a renewal term clause for a marketing affiliate agreement between Birch First Global Investments Inc. ("Company") and Mount Knowledge Holdings Inc. ("Marketing Affiliate", "MA") with a renewal term of successive 1 year This agreement shall begin upon the date of its execution by MA and acceptance in writing by Company and shall remain in effect until the end of the current calendar year and shall be automatically renewed for successive one (1) year periods unless otherwise terminated according to the cancellation or termination provisions contained in paragraph 18 of this agreement.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Drafting</head><p>Classify the following clause as either a renewal term clause or a notice period to terminate renewal clause or a governing law clause .. or a covenant not to sue clause or a third party beneficiary clause: "This agreement may be terminated by either party at the expiration of its term or any renewal term upon thirty (30) days written notice to the other party. " This is a notice period to terminate renewal clause Classification Extract the jurisdiction of the governing law of the agreement from the following text: "This Agreement shall be governed by, and any dispute arising hereunder shall be determined in accordance with, the laws of State of New York"</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>New York Extraction</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Table 4</head><p>Examples of the instruction response data created from the CUAD dataset</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>B. Example Outputs of Model</head><p>The following table outlines examples of outputs from the models with the associated prompts:</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Model Input</head><p>Model Output Extract the jurisdiction of the governing law of the agreement from the following text: This Agreement shall be governed by, and construed in accordance with, the Laws of the State of New York, applicable to contracts executed in and to be performed entirely within that state.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>New York</head><p>Draft a notice period to terminate renewal clause for a videoon-demand content license agreement between Rogers Cable Communications Inc. ("rogers") and Euromedia Holdings Corp. ("licensor") with a notice period of 60 days Subject to the terms and conditions of this agreement, during the term hereof, and for a period of one (1) year thereafter, Rogers shall have the right to terminate this agreement at any time upon thirty (30) days written notice to the other party.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Table 5</head><p>Examples of generated output from the fine-tuned model</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: Visual demonstration of the various AI postures across the AI build spectrum</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1</head><label>1</label><figDesc></figDesc><table><row><cell></cell><cell cols="2">Provider Pa-</cell><cell>License</cell><cell>Com-</cell></row><row><cell></cell><cell></cell><cell>ram-</cell><cell></cell><cell>mer-</cell></row><row><cell></cell><cell></cell><cell>eters</cell><cell></cell><cell>cial</cell></row><row><cell></cell><cell></cell><cell></cell><cell></cell><cell>Use</cell></row><row><cell>LLaMa</cell><cell>Face-</cell><cell>65B</cell><cell>Aca-</cell><cell>No</cell></row><row><cell></cell><cell>book</cell><cell></cell><cell>demic</cell><cell></cell></row><row><cell></cell><cell>Research</cell><cell></cell><cell>Use Only</cell><cell></cell></row><row><cell>pythia</cell><cell cols="2">EleutherAI 70M-</cell><cell>Apache</cell><cell>Yes</cell></row><row><cell></cell><cell></cell><cell>12B</cell><cell>2.0</cell><cell></cell></row><row><cell>Cerebras-</cell><cell>Cerebras</cell><cell>111M-</cell><cell>Apache</cell><cell>Yes</cell></row><row><cell>GPT</cell><cell></cell><cell>13B</cell><cell>2.0</cell><cell></cell></row><row><cell>Stabil-</cell><cell>Stabil-</cell><cell>3B/7B</cell><cell>CC</cell><cell>Yes</cell></row><row><cell>ityLM</cell><cell>ityAI</cell><cell></cell><cell>BY-SA</cell><cell></cell></row><row><cell></cell><cell></cell><cell></cell><cell>4.0</cell><cell></cell></row><row><cell>MPT-7B</cell><cell>Mo-</cell><cell>7B</cell><cell>Apache</cell><cell>Yes</cell></row><row><cell></cell><cell>saicML</cell><cell></cell><cell>2.0</cell><cell></cell></row></table></figure>
		</body>
		<back>

			<div type="availability">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Data</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Code</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Compute</head><p>Base Model</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Summary of chatgpt/gpt-4 research and perspective towards the future of large language models</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Han</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Ma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Yang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Tian</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>He</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>He</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Wu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Zhu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Qiang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Shen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Ge</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">a r X i v</title>
		<imprint>
			<biblScope unit="volume">2</biblScope>
			<biblScope unit="issue">3</biblScope>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<monogr>
		<title/>
		<author>
			<persName><surname>Openai</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
	<note type="report_type">Gpt-4 technical report</note>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Language models are few-shot learners</title>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">B</forename><surname>Brown</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Mann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Ryder</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Subbiah</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Kaplan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Dhariwal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Neelakantan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Shyam</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Sastry</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Askell</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Agarwal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Herbert-Voss</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Krueger</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Henighan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Child</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Ramesh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">M</forename><surname>Ziegler</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Wu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Winter</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Hesse</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Sigler</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Litwin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Gray</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Chess</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Clark</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Berner</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Mccandlish</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Radford</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Sutskever</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Amodei</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">a r X i</title>
		<imprint>
			<biblScope unit="volume">2</biblScope>
			<biblScope unit="issue">0</biblScope>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<monogr>
		<title level="m" type="main">The potentially large effects of artificial intelligence on economic growth</title>
		<author>
			<persName><forename type="first">J</forename><surname>Hatzius</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Briggs</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Kodani</surname></persName>
		</author>
		<ptr target="https://www.key4biz.it/wp-content/uploads/2023/03/Global-Economics-Analyst_-ThePotentially-Large-Effects-of-Artificial-Intelligence-on-Economic-Growth-Briggs_Kodnani.pdf" />
		<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Innovation diffusion in the legal industry</title>
		<author>
			<persName><forename type="first">W</forename><forename type="middle">D</forename><surname>Henderson</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Dickinson L. Rev</title>
		<imprint>
			<biblScope unit="volume">122</biblScope>
			<biblScope unit="page">395</biblScope>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Llama: Open and efficient foundation language models</title>
		<author>
			<persName><forename type="first">H</forename><surname>Touvron</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Lavril</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Izacard</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Martinet</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M.-A</forename><surname>Lachaux</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Lacroix</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Rozière</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Goyal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Hambro</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Azhar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Rodriguez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Joulin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Grave</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Lample</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">a r X i</title>
		<imprint>
			<biblScope unit="volume">2</biblScope>
			<biblScope unit="issue">3</biblScope>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Pythia: A suite for analyzing large language models across training and scaling</title>
		<author>
			<persName><forename type="first">S</forename><surname>Biderman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Schoelkopf</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><surname>Anthony</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Bradley</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>O'brien</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Hallahan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">A</forename><surname>Khan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Purohit</surname></persName>
		</author>
		<author>
			<persName><forename type="first">U</forename><forename type="middle">S</forename><surname>Prashanth</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Raff</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Skowron</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Sutawika</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Van Der Wal</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">a r X i v</title>
		<imprint>
			<biblScope unit="volume">2</biblScope>
			<biblScope unit="issue">3</biblScope>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<monogr>
		<title level="m" type="main">Cerebras-GPT: Open Compute-Optimal Language Models Trained on the Cerebras Wafer-Scale Cluster</title>
		<author>
			<persName><forename type="first">N</forename><surname>Dey</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Gosal</surname></persName>
		</author>
		<author>
			<persName><surname>Zhiming</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Khachane</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Marshall</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Pathria</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Tom</surname></persName>
		</author>
		<author>
			<persName><surname>Hestness</surname></persName>
		</author>
		<idno type="DOI">10.48550/arXiv.2304.03208</idno>
		<idno type="arXiv">arXiv:2304.03208</idno>
		<ptr target="http://arxiv.org/abs/2304.03208.doi:10.48550/arXiv.2304.03208" />
		<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<monogr>
		<author>
			<persName><surname>Stabilityai</surname></persName>
		</author>
		<ptr target="https://stability.ai/blog/stability-ai-launches-the-first-of-its-stablelm-suite-of-language-models" />
		<title level="m">Stability ai launches the first of its stablelm suite of language models</title>
				<imprint>
			<date type="published" when="2023-05-17">2023. 2023-05-17</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<monogr>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">N</forename><surname>Team</surname></persName>
		</author>
		<ptr target="URL:www.mosaicml.com/blog/mpt-7b" />
		<title level="m">Introducing mpt-7b: A new standard for open-source, ly usable llms</title>
				<imprint>
			<date type="published" when="2023-05-17">2023. 2023-05-17</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<monogr>
		<title level="m" type="main">Finetuned Language Models Are Zero-Shot Learners</title>
		<author>
			<persName><forename type="first">J</forename><surname>Wei</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Bosma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><forename type="middle">Y</forename><surname>Zhao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Guu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">W</forename><surname>Yu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Lester</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Du</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">M</forename><surname>Dai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><forename type="middle">V</forename><surname>Le</surname></persName>
		</author>
		<idno type="DOI">10.48550/arXiv.2109.01652</idno>
		<idno type="arXiv">arXiv:2109.01652</idno>
		<ptr target="http://arxiv.org/abs/2109.01652.doi:10.48550/arXiv.2109." />
		<imprint>
			<date type="published" when="1652">2022. 0 1 6 5 2</date>
		</imprint>
	</monogr>
	<note>cs</note>
</biblStruct>

<biblStruct xml:id="b11">
	<monogr>
		<title level="m" type="main">Training language models to follow instructions with human feedback</title>
		<author>
			<persName><forename type="first">L</forename><surname>Ouyang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Wu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Jiang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Almeida</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">L</forename><surname>Wainwright</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Mishkin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Agarwal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Slama</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Ray</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Schulman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Hilton</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Kelton</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Miller</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Simens</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Askell</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Welinder</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Christiano</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Leike</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Lowe</surname></persName>
		</author>
		<idno type="DOI">10.48550/arXiv.2203.02155</idno>
		<idno type="arXiv">arXiv:2203.02155</idno>
		<ptr target="http://arxiv.org/abs/2203.02155.doi:10.48550/arXiv.2203." />
		<imprint>
			<date type="published" when="2022">2022. 0 2 1 5 5</date>
		</imprint>
	</monogr>
	<note>cs</note>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Training a helpful and harmless assistant with reinforcement learning from human feedback</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Bai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Jones</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Ndousse</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Askell</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Dassarma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Drain</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Fort</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Ganguli</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Henighan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Joseph</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Kadavath</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Kernion</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Conerly</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>El-Showk</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Elhage</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Hatfield-Dodds</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Hernandez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Hume</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Johnston</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Kravec</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Lovitt</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Nanda</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Olsson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Amodei</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Brown</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Clark</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Mccandlish</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Olah</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Mann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Kaplan</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">a r X i v</title>
		<imprint>
			<biblScope unit="volume">2</biblScope>
			<biblScope unit="issue">2</biblScope>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<monogr>
		<title level="m" type="main">Free dolly: Introducing the world&apos;s first truly open instruction-tuned llm</title>
		<author>
			<persName><surname>Databricks</surname></persName>
		</author>
		<ptr target="https://www.databricks.com/blog/2023/04/12/dolly-first-open-commercially-viable-instruction-tuned-llm" />
		<imprint>
			<date type="published" when="2023-05-17">2023. 2023-05-17</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<monogr>
		<author>
			<persName><forename type="first">T</forename><surname>Groff</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Jones</surname></persName>
		</author>
		<ptr target="https://books.google.co.uk/books?id=pmJv09aw4rMC" />
		<title level="m">Introduction to Knowledge Management: KM in Business</title>
				<imprint>
			<publisher>Taylor &amp; Francis Group</publisher>
			<date type="published" when="2003">2003</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">Knowledge management and legal practice</title>
		<author>
			<persName><forename type="first">T</forename><surname>Plessis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Toit</surname></persName>
		</author>
		<idno type="DOI">10.1016/j.ijinfomgt.2006.06.003</idno>
		<ptr target="https:" />
	</analytic>
	<monogr>
		<title level="j">International Journal of Information Management</title>
		<imprint>
			<biblScope unit="volume">26</biblScope>
			<biblScope unit="page" from="360" to="371" />
			<date type="published" when="2006">2006</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">Knowledge management and the law firm</title>
		<author>
			<persName><forename type="first">A</forename><surname>Terrett</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of knowledge management</title>
		<imprint>
			<date type="published" when="1998">1998</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<monogr>
		<author>
			<persName><forename type="first">J</forename><surname>Roos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Edvinsson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><forename type="middle">C</forename><surname>Dragonetti</surname></persName>
		</author>
		<title level="m">Intellectual capital: Navigating the new business landscape</title>
				<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="1997">1997</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<analytic>
		<title level="a" type="main">The pile: An 800gb dataset of diverse text for language modeling</title>
		<author>
			<persName><forename type="first">L</forename><surname>Gao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Biderman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Black</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Golding</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Hoppe</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Foster</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Phang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>He</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Thite</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Nabeshima</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Presser</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Leahy</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">a r X i</title>
		<imprint>
			<biblScope unit="volume">2</biblScope>
			<biblScope unit="issue">1</biblScope>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<analytic>
		<title level="a" type="main">Cuad: An expert-annotated nlp dataset for legal contract review</title>
		<author>
			<persName><forename type="first">D</forename><surname>Hendrycks</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Burns</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Ball</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">a r X i</title>
		<imprint>
			<biblScope unit="volume">2</biblScope>
			<biblScope unit="issue">1</biblScope>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<analytic>
		<title level="a" type="main">Maud: An expert-annotated legal nlp dataset for merger agreement understanding</title>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">H</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Scardigli</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Tang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Levkin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Ball</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Woodside</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Hendrycks</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">a r X i</title>
		<imprint>
			<biblScope unit="volume">2</biblScope>
			<biblScope unit="issue">3</biblScope>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
