<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Prior Case Retrieval using Evidence Extraction from Court Judgements</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Basit</forename><surname>Ali</surname></persName>
							<affiliation key="aff0">
								<orgName type="institution">TCS Research</orgName>
								<address>
									<settlement>Pune</settlement>
									<country key="IN">India</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Ravina</forename><surname>More</surname></persName>
							<affiliation key="aff0">
								<orgName type="institution">TCS Research</orgName>
								<address>
									<settlement>Pune</settlement>
									<country key="IN">India</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Sachin</forename><surname>Pawar</surname></persName>
							<affiliation key="aff0">
								<orgName type="institution">TCS Research</orgName>
								<address>
									<settlement>Pune</settlement>
									<country key="IN">India</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Girish</forename><forename type="middle">K</forename><surname>Palshikar</surname></persName>
							<affiliation key="aff0">
								<orgName type="institution">TCS Research</orgName>
								<address>
									<settlement>Pune</settlement>
									<country key="IN">India</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Prior Case Retrieval using Evidence Extraction from Court Judgements</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">09A463EC5B5D11D771D50D72322AC5A6</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T16:29+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Evidence Extraction</term>
					<term>Evidence Information Model</term>
					<term>Natural Language Processing</term>
					<term>Prior Case Retrieval</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>One of the key constituents of court case descriptions is Evidence description and observations. Along with witness testimonies, evidence plays a significant role in the final decision of the case. We propose a weakly supervised technique to automatically identify sentences containing evidences. We represent the information related to evidences in these sentences in a semantically rich structure -Evidence Structure defined as an Evidence Information Model. We show that witness testimony information can also be represented using the same model. We demonstrate the effectiveness of our Evidence Information Model for the prior case retrieval application by proposing a matching algorithm for computing semantic similarity between a query and a sentence in a court case description. To the best of our knowledge, this is the first paper to apply NLP techniques for the extraction of evidence information from court judgements and use it for retrieving relevant prior court cases.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>Evidences -typically based on documents (e.g., letter, receipt, report, agreements, affidavits) and physical objects (e.g., knife, guns, photos, phone call data records) -are often used by lawyers in their arguments during a court case. The observations made through these evidences may have a significant effect on the judges' final decision. In order to develop a deeper understanding of the past court cases, it is valuable to identify various Evidences discussed in these cases and the observations which are made about them or through them. Such information about evidences has several applications such as understanding and representing legal arguments, determining strengths and weaknesses of those arguments, identifying relevant past cases in which similar evidences were discussed, etc.</p><p>In this paper, we discuss Natural Language Processing (NLP) based techniques for extracting information regarding Evidences mentioned in court judgement documents. We propose to represent this information in a rich semantic structure -Evidence Structure defined as an Evidence Information Model. Along with Evidences, we also identify and represent Witness Testimonies using the same Information Model. Initially, we discuss a twostep approach for identifying evidence and testimony sentences. In the first step, linguistic rules are used to determine whether a sentence contains any evidence or <ref type="bibr">(ASAIL 2021)</ref>, June 25, 2021, São Paulo, Brazil. Envelope ali.basit@tcs.com (B. Ali); ravina.m@tcs.com (R. More); sachin7.p@tcs.com (S. Pawar); gk.palshikar@tcs.com (G. K. Palshikar) testimony information. Here, we use the rules proposed in Ghosh et al. <ref type="bibr" target="#b0">[1]</ref> for identification of witness testimonies and design new rules for identification of evidence sentences. In the second step, we train a Weakly Supervised Sentence Classifier whose training data is automatically created using the sentences identified by the linguistic rules. It is a multi-label classifier which predicts whether any sentence contains an Evidence or Witness Testimony or both. Once all the Evidence and Testimony sentences are identified from the corpus of court judgements, we propose a Semantic Role Labelling (SRL) <ref type="bibr" target="#b1">[2]</ref> based technique to automatically instantiate Evidence Structures for these sentences.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Proceedings of the Fifth Workshop on Automated Semantic Analysis of Information in Legal Text</head><p>To demonstrate effectiveness of the proposed Evidence Structure, we discuss its use in the prior case retrieval application. We propose a matching algorithm for computing semantic similarity between a query and a sentence in a court judgement document. This algorithm makes use of the proposed Evidence Structure in which both the query and the sentence are represented, resulting in a semantically sound similarity score between them.</p><p>Previously, Ghosh et al. <ref type="bibr" target="#b0">[1]</ref> identified witness testimonies from court case documents and used them for retrieving relevant prior cases. We propose that considering only witness testimonies leads to loss of key information regarding evidences mentioned in a case. Hence, we identify and use information about various evidences mentioned in the case documents leading to much better prior case retrieval performance as demonstrated in the experiments section. Moreover, Ghosh et al. <ref type="bibr" target="#b0">[1]</ref> use a much limited semantic structure to represent information regarding events mentioned in witness testimonies. This structure does not capture important semantic information like whether event is negated, what are the causes behind the event, the manner in which event takes place etc. We propose a richer semantic structure addressing these limitations and design a suitable semantic matching algorithm for that structure. To the best of our knowledge, this is the first paper to apply NLP techniques for the extraction of evidence information from court judgements and demonstrate its use for retrieving relevant prior court cases.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Evidence Information Model</head><p>The purpose of Evidence Information Model is to define a suitable structure to represent evidence information in court judgements. In this section, we describe Semantic Role Labelling in brief and how it is used to define our proposed Evidence Structure.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1.">Background: Semantic Role Labelling</head><p>Semantic Role Labelling (SRL) is a technique in Natural Language Processing that identifies verbs/predicates in a sentence, finds phrases connected to every predicate and assigns an appropriate semantic role to every phrase. By doing so, SRL helps machines to understand the roles of important words within a sentence. Following are some key semantic roles identified for a verb/predicate (often corresponding to an action or event) by SRL techniques: ARG0 : proto-agent or someone who performs the action denoted by the verb ARG1 : proto-patient or someone on whom the action is performed on ARGM-TMP : the time when the event took place ARGM-CAU : the cause of the action ARGM-PRP : the purpose of the action ARGM-LOC : the location where the event took place ARGM-MNR : the manner in which the action took place ARGM-NEG : the word indicating that the action did not take place Consider the following example sentence:</p><p>O n A u g u s t 2 5 , 1 9 6 5 , t h e b a n k d i s h o n o u r e d t h e c h e q u e d u e t o i n s u f f i c i e n t b a l a n c e .</p><p>The various semantic roles to the verb d i s h o n o u r e d are annotated as follows:</p><formula xml:id="formula_0">[ A R G M -T M P : O n A u g u s t 2 5 , 1 9 6 5 ] , [ A R G 0 : t h e b a n k ] [ V : d i s h o n o u r e d ] [ A R G 1 : t h e c h e q u e ] [ A R G M -C A U : d u e t o i n s u f f i c i e n t b a l a n c e ] .</formula><p>We use the predicates and corresponding arguments obtained from the pre-trained AllenSRL model <ref type="bibr" target="#b2">[3]</ref> to instantiate our Evidence Structure for the queries and candidate sentences.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2.">Evidence Structure</head><p>The Evidence Information Model represents every Evidence Sentence giving information about one or more Evidence Objects in an Evidence Structure. We define an Evidence Object as one of the objects presented by the counsels to the judge along with the information and findings about the crime. It is thus, a physical entity that can furnish some degree of support, contradiction or opposition to some legal arguments. Some examples of Evidence Objects are:</p><p>• Documents (autopsy • action took place (e.g., a s p e r t h e c h a l l a n , f r a u d u l e n t l y , w i l f u l l y ) Table <ref type="table">1</ref> shows examples of some Evidence Sentences along with the corresponding Evidence Structure Instances. In some cases, Observation Frame may be empty due to absence of ObservationVerb. In such cases, EvidenceObject may be present as a part of any argument in Evidence Frame. E.g., t h e c h e q u e is present as 𝐴 1 in the Evidence Frame of the first sentence in Table <ref type="table">1</ref>.</p><formula xml:id="formula_1">O F = [𝑂𝑉 = s h o w e d , 𝐸𝑂 = t h e r o u g h c a s h b o o k ] E F = [𝐸𝑉 = s e n t , 𝐴 0 = b y a p p e l l a n t G u p t a , 𝐴 1 = a s u m o f R s . 2 1 , 1 3 3 , 𝐴 2 = t o t h e T</formula><p>Information about named entities and their types present in various arguments of Observation or Evidence frame is important. Hence, the Observation Frame and Evidence Frame are also enriched by annotating entities such as</p><formula xml:id="formula_2">P E R S O N , O R G A N I S A T I O N , G E O -P O L I T I C A L E N T I T Y , L O C A T I O N , P R O D U C T , E V E N T , L A N G U A G E , D A T E , T I M E , P E R C E N T , M O N E Y , Q U A N T I T Y , O R D I N A L , C A R D I N A L , W E A P O N , S U B S T A N C E , D O C U M E N T , A R T I F A C T , W O R K _ O F _ A R T , W I T N E S S , B O D Y _ P A R T , and V E H I C L E present in the fields.</formula><p>Witness Information Model: Information in witness testimonies can also be represented using the same Evidence Structure. The statement verbs used in witness testimony sentences (e.g., s t a t e d , s a i d ) are treated similar to observation verbs and represented using Observation Frames. Similarly, other action/event verbs mentioned in witness testimony sentences are represented using Evidence Frames. Table <ref type="table">2</ref> shows examples of some Witness Sentences along with the corresponding Evidence Structure Instances. The advantage of representing information about evidences and witness testimonies in the same structure is that we can make use of both these sources of information seamlessly, for prior case retrieval.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Methodology</head><p>In this section, we describe our overall methodology which consists of two phases. In the first phase, we identify Evidence and Testimony Sentences using linguistic rules and weakly supervised sentence classifier. In the second phase, we instantiate the Evidence Structures for these identified sentences. For all our experiments, we use a corpus of 30,032 Indian Supreme Court judgements ranging from the year 1952 to 2012.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1.">Identification of Evidence and Testimony Sentences</head><p>We identify Evidence and Testimony sentences using a two-step approach. In the first step, we use linguistic rules to obtain Evidence and Testimony sentences. In the second step, we use these sentences to train a sentence classifier. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Table 2 Example Witness Testimony sentences with their Evidence Structure Instances</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>H e a d m i t t e d , h o w e v e r , t h a t S h r i B u c h h a d m e t h i m i n c o n n e c t i o n w i t h t h e c o v e n a n t , b u t h e d e n i e d t h a t h e h a d r e c e i v e d a n y l e t t e r E x h i b i t P -9 f r o m S h r i B u c h o r t h e l i s t s E</head><p>x h i b i t s P -1 0 t o P -1 2 r e g a r d i n g h i s p r i v a t e a n d S t a t e p r o p e r t i e s , w e r e a p a r t t h e r e o f .</p><formula xml:id="formula_3">• O F = [𝑂𝑉 = a d m i t t e d , 𝐴 0 = H e ] E F = [𝐸𝑉 = m e t , 𝐴 0 = S h r i B u c h , 𝐴 1 = h i m , 𝑇 𝑀𝑃 = i n c o n n e c t i o n w i t h t h e c o v e n a n t ] • O F = [𝑂𝑉 = d e n i e d , 𝐴 0 = H e ] E F = [𝐸𝑉 = r e c e i v e d , 𝐴 0 = h e , 𝐴 1 = a n y l e t t e r E x h i b i t P -9 , 𝐴 2 = f r o m S h r i B u c h ]</formula><p>Step I: Linguistic Rules based Approach: As there are no publicly annotated datasets for identification of Evidence and Testimony sentences, we rely on linguistic rules to identify these sentences with high precision as our first step. The linguistic rules for identifying Evidence sentences are described in detail in Table <ref type="table">3</ref>. These rules identified 62,310 sentences as Evidences from our corpus. As there is no annotated dataset, in order to estimate the precision of the linguistic rules we use random sampling strategy. We selected a set 100 random sentences identified as Evidence by the linguistic rule, and got them verified by a human expert. The precision turned out to be 85%. Similarly, we use the linguistic rules proposed in Ghosh et al. <ref type="bibr" target="#b0">[1]</ref> for identifying Testimony and non-Testimony sentences where the reported precision is around 85%. These rules identified 36,473 sentence as Testimony and 14,234 sentences as non-Testimony from the same corpus.</p><p>Step II: Weakly Supervised Sentence Classification:</p><p>We observed that although the linguistic rules identify Evidence and Testimony sentences with high precision, they may miss to identify some sentences which should have been identified as Evidence or Testimony (see examples in Table <ref type="table">4</ref>). Hence, we train a supervised sentence classifier to improve overall recall of identification of Evidence and Testimony sentences. The classifier used is a BiLSTM-based <ref type="bibr" target="#b3">[4]</ref> multi-label sentence classifier whose architecture is depicted in Figure <ref type="figure" target="#fig_4">1</ref>. This classifier is weakly supervised since its training data is automatically created using the sentences identified by the linguistic rules as follows:</p><p>• The classifier has two outputs -i) first output predicts a binary label indicating whether the sentence contains Evidence or not and ii) second output predicts a binary label indicating whether the sentence contains Testimony or not.</p><p>• 1824 sentences are labelled as Evidence and Testimony both. These sentences are identified as Evidence as well as Testimony by both the sets of linguistic rules.</p><p>• 60486 sentences are labelled as Evidence and non-Testimony. These sentences are identified as Evidence by the rules but not as Testimony.</p><p>• 34649 sentences are labelled as non-Evidence and Testimony. These sentences are identified as Testimony by the rules but not as Evidence.</p><p>• 14234 sentences are labelled as non-Evidence and non-Testimony. These sentences are identified as non-Testimony by the rules and not identified as Evidence.</p><p>After this classifier is trained, we use it to classify all the remaining sentences in the corpus. These sentences are neither identified Evidence by the Evidence rules nor as Testimony/non-Testimony by the Testimony rules. Using the prediction confidence, we selected top 10,000 sentences classified as Evidence and top 5,000 sentences classified as Testimony. Table <ref type="table">4</ref> shows some examples of sentences identified as Evidence by the classifier but not by the linguistic rules. To estimate the precision, we again employed the random sampling strategy. We selected 100 random sentences each from these high confidence Evidence and Testimony sentences and a human expert verified them. The precision of 72% is observed for Evidence sentences and 68% for Testimony sentences. The precision of the sentence classifier is lower as compared</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Table 3</head><p>Linguistic Rules for identifying Evidence Sentences Any sentence 𝑆 should satisfy the following conditions in order to be identified as an Evidence Sentence: E-R1 𝑆 should contain at least one Evidence Object as defined in Section 2.2. The list of words corresponding to evidence objects is created automatically by using WordNet hypernym structure. We create a list of all words for which the following WordNet synsets are ancestors in hypernym treea r t i f a c t (e.g., g u n , c l o t h e s ), d o c u m e n t (e.g. r e p o r t , l e t t e r ), s u b s t a n c e (e.g. k e r o s e n e , b l o o d ). This list is looked up to identify evidence objects in a sentence. E-R2 𝑆 should contain at least one action verb from a pre-defined set of verbs like t a m p e r , k i l l , s u s t a i n , f o r g e OR 𝑆 should contain at least one observation verb from a pre-defined set of verbs like r e p o r t , s h o w , f i n d . Both the pre-defined sets of verbs are prepared by observing multiple example sentences containing evidence objects. E-R3 In the dependency tree of 𝑆, the evidence object (identified by E-R1) should occur within the subtree rooted at the action or observation verb (identified by E-R2) AND there should not be any other verb (except auxiliary verbs like h a s , b e e n , w a s , w e r e , i s ) occurring between the two. This ensures that the evidence object always lies within the verb phrase headed by the action or observation verb.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Table 4</head><p>Example of Evidence Sentences Identified by the Classifier but not by the linguistic rules S 1 : R a j u P W 2   to the rules because, it is applied on a more difficult set of sentences for which the linguistic rules fail to identify any label. At the end of this two-step process (linguistic rules followed by the sentence classifier), we have 112,401 sentences identified either as Evidence or as Testimony.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>t o o k P r e e t i i n t o t h e b a t h r o o m a t t h e i n s t a n c e o f A c c u s e d N o . 1 w h o c u t a l e n g t h o f w i r e o f w a s h i n g m a c h i n e a n d u s e d i t t o c h o k e h e r t o d e a t h , w h</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>t t i n g h i s k n e e s o n h e r s t o m a c h a n d w h e n s h e w a s i m m o b i l i s e d t h i s w a y , t h e A c c u s e d N o . 1 g a v e h e r k n i f e b l o w s o n h e r n e c k w i t h t h e r e s u l t s h e a l s o d i e d . S 6 : A l m i r a h s f o u n d i n t h e f l a t w e r e e m p t i e d t o t h e e x t e n t t h e a c c u s e d c o u l d p u t a r t i c l e s a n d o t h e r c a</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.">Evidence Structure Instances</head><p>In this phase, we discuss the technique of instantiating Evidence Structures for sentences identified as Evidence or Testimony in the previous phase. We used Semantic Role Labelling <ref type="bibr" target="#b2">[3]</ref> to identify and fill the arguments of the Observation Frame and the Evidence Frame in the Evidence Structure Instance for every candidate sentence. This is demonstrated in Algorithm 1. We identify Observation Frames using Observation Cue Verbs. For each of these Observation Frames we identify the corresponding Evidence Objects and Evidence Frames. For identifying Evidence Objects, we first use Named Entity Recognition <ref type="bibr" target="#b4">[5]</ref> and WordNet based Entity Identification <ref type="bibr" target="#b5">[6]</ref> to identify the named entities in the sentence and annotate them in the Frames extracted. The Evidence Objects in a phrase are then obtained by selecting named entities annotated as one of the following types </p><formula xml:id="formula_4">-A R T I F A C T , V E H I C L E , W E A P O N , D O C U M E N T , W O R K _ O F _ A R T , S U B S T A N C E .</formula></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Prior Case Retrieval</head><p>In order to demonstrate effectiveness of the proposed Evidence Structure, we apply it for the task of prior case retrieval. This task is to create a relevance-based ranked list of court judgements (documents) in our corpus for a query. In order to retrieve prior cases for a query, we represent the query using an Evidence Structure Instance (𝐸𝑣𝑆𝑡𝑟𝑢𝑐𝑡 𝑄 ). We then compute the similarity of query instance 𝐸𝑣𝑆𝑡𝑟𝑢𝑐𝑡 𝑄 against each document instance 𝐸𝑣𝑆𝑡𝑟𝑢𝑐𝑡 𝐷 obtained from every Evidence or Testimony sentence in the corpus. Algorithm 2 shows the steps for computing similarity. We refer to this algorithm as 𝑆𝑒𝑚𝑀𝑎𝑡𝑐ℎ because of its semantic matching ability. We use cosine similarity between the phrase embeddings of corresponding arguments of the Evidence Structure Instances to compute similarity. For obtaining phrase embedding for any phrase (referred as 𝑃ℎ𝑟𝑎𝑠𝑒𝑉 𝑒𝑐 in Algorithm 2), we consider the average of GloVe word embeddings <ref type="bibr" target="#b6">[7]</ref> of the words in that phrase excluding stop words. We compute the similarity scores within corresponding arguments of both the frames. These scores across different arguments are combined to get a final similarity score between 𝐸𝑣𝑆𝑡𝑟𝑢𝑐𝑡 𝑄 and 𝐸𝑣𝑆𝑡𝑟𝑢𝑐𝑡 𝐷 . We multiply the final similarity score by a Sentence BERT <ref type="bibr" target="#b7">[8]</ref> based similarity score between the query and the sentence containing 𝐸𝑣𝑆𝑡𝑟𝑢𝑐𝑡 𝐷 . This is necessary because errors in the automated SRL tool may lead to imperfect Evidence Structure instances in some cases. A sentence similarity score which is not dependent on any such structure within the sentences provides a complementary view of capturing sentence similarity. Finally, the overall relevance score of the query with a document is the maximum score corresponding to any Evidence Structure Instance 𝐸𝑣𝑆𝑡𝑟𝑢𝑐𝑡 𝐷 obtained from the document. Table <ref type="table">5</ref> shows a running example of how a similarity score is computed between an Evidence Structure Instance (𝐸𝑣𝑆𝑡𝑟𝑢𝑐𝑡 𝑄 ) from a query and an Evidence Structure Instance (𝐸𝑣𝑆𝑡𝑟𝑢𝑐𝑡 𝐷 ) from a document in the corpus.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Related Work</head><p>While the task of evidence extraction from legal documents is related to several information retrieval and NLP tasks, there are no established baselines for the task. Bellot et al. <ref type="bibr" target="#b8">[9]</ref> and Cartright et al. <ref type="bibr" target="#b9">[10]</ref> have worked on Evidence Retrieval that identifies whole documents that contain an evidence. On the other hand, Rinott et al. <ref type="bibr" target="#b10">[11]</ref> use Context Dependent Evidence Detection to find evidence information present in a sentence on a phrase level. As compared to this, we identify both Evidence and Testimony sentences, represent them in a rich structure and also use that for prior case retrieval. This is a challenging task due to the inherently complex nature of legal texts and the finer granularity of matching involved. Ji et al. <ref type="bibr" target="#b11">[12]</ref> propose an Evidence Information Extraction system which captures evidence production paragraph, evidence cross-examination paragraph, evidence provider, evidence name, evidence content, crossexamination party and cross-examination opinion relating to an evidence presented in the court. While this technique may suit well for Chinese court records that follow a relatively structured representation, it does not suit well to the Indian Court Records that contain descriptive and varied formats of the court proceedings.</p><p>Gomes and Ladeira <ref type="bibr" target="#b12">[13]</ref> and Landthaler et al. <ref type="bibr" target="#b13">[14]</ref> performs full text search for legal document collection by obtaining word2vec word embeddings and then taking their average for computing similarity. However, computing the average of the embeddings gives a lossy representation where relative order of the words is lost. In contrast, we represent the sentences using the Evidence Structure Instances, where the structure itself takes care of the relative ordering. Gomes and Ladeira <ref type="bibr" target="#b12">[13]</ref> demonstrate BM25 and TF-IDF for Prior Case retrieval. In our results section, we demonstrate the comparative poor performance of BM25 and TF-IDF in handling corner cases.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.">Experimental Evaluation</head><p>In this section, we discuss our experiments including the dataset, baseline techniques, evaluation metrics and analysis of results.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.1.">Dataset</head><p>We use the Indian Supreme Court judgements from years 1952 to 2012 freely available at http://liiofindia.org/in/ cases/cen/INSC/. There are 30032 court (documents) containing 4,111,091 sentences where average sentence length is 31 words and standard deviation of 24.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.2.">Baselines</head><p>For the task of prior case retrieval, we implement two baseline techniques:</p><p>• BM25: It is a popular TF-IDF based relevance computation technique. We use the BM25+ variant<ref type="foot" target="#foot_0">1</ref> as described in Trotman et al. <ref type="bibr" target="#b14">[15]</ref>. This technique uses a bag-ofwords approach that ignores the sentence structure. We input : 𝑠 (sentence), 𝑆𝑅𝐿_𝑃 (set of semantic frames in 𝑠 as per any semantic role labeller, each frame 𝑃 consists of a predicate 𝑃.𝑉 and corresponding arguments 𝑃.𝐴𝑅𝐺 0 , 𝑃.𝐴𝑅𝐺 1 , 𝑃.𝐴𝑅𝐺 2 , 𝑃.𝐴𝑅𝐺𝑀-𝐿𝑂𝐶, etc.) output : 𝐸𝑣𝑆𝑡𝑟𝑢𝑐𝑡𝑠 = Evidence Structure Instances of the input sentence consisting of 𝑂𝑏𝑠𝑒𝑟𝑣𝑎𝑡𝑖𝑜𝑛𝐹 𝑟𝑎𝑚𝑒 (𝑂𝐹) and 𝐸𝑣𝑖𝑑𝑒𝑛𝑐𝑒𝐹 𝑟𝑎𝑚𝑒 (𝐸𝐹) parameter : 𝑂𝐵𝑆_𝑉 𝐸𝑅𝐵𝑆 = {accept, a d d , a d m i t , a g r e e , a l l e g e , a l l o w , a l t e r , a p p r i s e ,    <ref type="table">5</ref>). As we are not resolving co-references, we are missing a few relevant documents. E.g., 𝑆𝑀 𝑇 𝐸 does not assign a high score for the following document for query 𝑄 3 (see Table <ref type="table">6</ref>) - . This is because t h e m in the Evidence Structure instance for s h o t is not explicitly known to correspond to t h e p o l i c e in the previous sentence.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="7.">Conclusion and Future Work</head><p>In this paper, we discussed several NLP techniques for identifying evidence sentences, representing them in the semantically rich Evidence Structure and retrieving relevant prior cases by exploiting it. The proposed techniques are weakly supervised as they do not rely on any manually annotated training data, except for the human expertise in designing the linguistic rules. Keeping in mind the importance of witness testimonies in addition to evidences, we also extracted and represented the witness testimonies using the same Evidence Structure. For the application of prior case retrieval, we evaluated our proposed technique along with several competent baselines, on a dataset of 10 diverse queries. We demonstrated that our technique performs comparably for most of the queries and is the best considering the overall performance across all 10 queries. The results highlight the contribution of evidence and testimony information in improving prior case retrieval performance.</p><p>In future, we plan to apply advanced representation learning techniques for learning dense or embedded representation of an entire Evidence Structure instance. Also, we plan to automatically determine the best suited retrieval technique (BM25, Sentence-BERT or SemMatch) for any query based on its nature. We plan to explore ensemble of multiple retrieval techniques for improving prior case retrieval performance further.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head></head><label></label><figDesc>r e a s u r y , 𝐴𝑅𝐺 −𝑇 𝑀𝑃 = o n S e p t e m b e r 2 9 , 1 9 5 0 ] • O F = [𝑂𝑉 = s h o w e d , 𝐸𝑂 = t h e T r e a s u r y f i g u r e s i n t h e c h a l l a n ] E F = [𝐸𝑉 = d e p o s i t e d ,𝐴 0 = b y a p p e l l a n t G u p t a , 𝐴 1 = o n l y a s u m o f R s . 1 , 1 3 3 , 𝐴 2 = i n t o t h e T r e a s u r y , 𝑇 𝑀𝑃 = o n t h a t d a y ] • O F = [𝑂𝑉 = s h o w e d , 𝐸𝑂 = t h e T r e a s u r y f i g u r e s i n t h e c h a l l a n ] E F = [𝐸𝑉 = m i s a p p r o p r i a t e d , 𝐴 1 = a s u m o f R s . 2 0 , 0 0 0 , 𝑀𝑁 𝑅 = d i s h o n e s t l y ] d e c e a s e d , a c h e q u e o f R s . 3 , 2 0 0 , h i s w i f e ) • Location or LOC: location where the action took place (e.g., i n t h e b e d r o o m , a t t h e b a n k , i n M a l a y s i a ) • Time or TMP: timestamp of the action (e.g., a b o u t 1 2 h o u r s b a c k , i n t h e m o r n i n g , o n M o n d a y ) • Cause or CAU: cause of the action (e.g., d u e t o d o w r y , a s a r e s u l t o f t h e C B I e n q u i r y , o u t o f s h e e r s p i t e ) • Manner or MNR: manner in which the</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head></head><label></label><figDesc>H e h a s c a t e g o r i c a l l y s t a t e d t h a t b y r e a s o n o f e n m i t y , A 1 a n d A 2 t o g e t h e r h a v e m u r d e r e d h i s b r o t h e r -i n -l a w . • O F = [𝑂𝑉 = s t a t e d , 𝐴 0 = H e ] E F = [𝐸𝑉 = m u r d e r e d , 𝐴 0 = A 1 a n d A 2 t o g e t h e r , 𝐴 1 = h i s b r o t h e r -i n -l a w , 𝐶𝐴𝑈 = b y r e a s o n o f e n m i t y ] S h r i D h o l e y ( P W -6 ) r e i t e r a t e d a b o u t t h e d a c o i t y a n d c l a i m e d t h a t a p i s t o l w a s b r a n d i s h e d o n h i m b y o n e o f t h e a c c u s e d p e r s o n s . • O F = [𝑂𝑉 = c l a i m e d , 𝐴 0 = S h r i D h o l e y ( P W -6 ) ] E F = [𝐸𝑉 = b r a n d i s h e d , 𝐴 0 = b y o n e o f t h e a c c u s e d p e r s o n s , 𝐴 1 = a p i s t o l , 𝐶𝐴𝑈 = o n h i m ] T h o u g h h e s t a t e d i n t h e p o s t -m o r t e m r e p o r t t h a t d e a t h w o u l d h a v e o c c u r r e d a b o u t 1 2 h o u r s b a c k , h e c l a r i f i e d t h a t t h e r e w a s p o s s i b i l i t y o f i n j u r i e s b e i n g r e c e i v e d a t a b o u t 9 A . M . • O F = [𝑂𝑉 = s t a t e d , 𝐴 0 = h e , 𝐸𝑂 = t h e p o s t -m o r t e m r e p o r t ] E F = [𝐸𝑉 = o c c u r r e d , 𝐴 1 = d e a t h , 𝑇 𝑀𝑃 = a b o u t 1 2 h o u r s b a c k ] • E F = [𝑂𝑉 = c l a r i f i e d , 𝐴 0 = h e , 𝐴 1 = t h a t t h e r e w a s p o s s i b i l i t y o f i n j u r i e s b e i n g r e c e i v e d a t a b o u t 9 A . M . D e c e a s e d S a r i t K h a n n a w a s a g e d a b o u t 2 7 y e a r s ]</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head></head><label></label><figDesc>o h o w e v e r , s u r v i v e d . S 2 : R a j u P W 2 t o o k S a t y a b h a m a b a i S u t a r i n t h e k i t c h e n w h e r e t h e a c c u s e d N o . 1 h a d a l r e a d y r e a c h e d a n d w a s w a s h i n g t h e b l o o d s t a i n e d k n i f e . S 3 : H e m l a t a w a s a l s o k i l l e d b y i n f l i c t i n g k n i f e i n j u r i e s . S 4 : A c c u s e d N o . 2 a n d R a j u P W 2 t o o k t h e c h i l d i n t o t h e r o o m w h e r e M e e r a b a i w a s l y i n g d e a d i n t h e p o o l o f b l o o d . S 5 : A c c u s e d N o . 2 g a v e h e r b l o w s b y p u</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_3"><head></head><label></label><figDesc>s h a n d v a l u a b l e s i n t h e a i r -b a g o b t a i n e d f r o m t h e s a i d f l a t . S 7 : B l o o d s t a i n e d c l o t h e s o f A c c u s e d N o . 2 w e r e p u t i n t h e a i r -b a g a l o n g w i t h s t o l e n a r t i c l e s .</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_4"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: Architecture of the BiLSTM-based multi-label sentence classifier (T: Testimony, NT: Non-Testimony, E: Evidence, NE: Non-Evidence)</figDesc><graphic coords="5,89.29,406.97,203.37,122.02" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_5"><head></head><label></label><figDesc>a s s e r t , b r i e f , b u i l d , c h a l l e n g e , c l a i m , c l a r i f y , c o m p l a i n , c o n f i r m , c o r r o b o r a t e , d e c l i n e , d e m a n d , d e n y , d e p o s e , d e s c r i b e , d i s c l o s e , d i s m i s s , e x a m i n e , e x h i b i t , f i n d , i n c l u d e , i n d i c a t e , i n f o r m , m e n t i o n , n o t e , n o t i c e , o b s e r v e , o b t a i n , o c c u r , p o i n t , p r e p a r e , p r e s e n t , r e c e i v e , r e c o v e r , r e f u s e , r e j e c t , r e m e m b e r , r e p o r</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_6"><head>/ 5</head><label>5</label><figDesc>/ C o m p u t i n g s i m i l a r i t y b e t w e e n m a i n p r e d i c a t e s , u s i n g c o s i n e s i m i l a r i t y o f t h e i r w o r d e m b e d d i n g s 𝑠𝑖𝑚 𝐸 ∶= 𝐶𝑜𝑠𝑖𝑛𝑒𝑆𝑖𝑚(𝑊 𝑜𝑟𝑑𝑉 𝑒𝑐(𝐸𝑣𝑆𝑡𝑟𝑢𝑐𝑡 𝑄 .𝐸𝐹 .𝑉 ), 𝑊 𝑜𝑟𝑑𝑉 𝑒𝑐(𝐸𝑣𝑆𝑡𝑟𝑢𝑐𝑡 𝐷 .𝐸𝐹 .𝑉 )) / / C o m p u t i n g s i m i l a r i t y b e t w e e n c o r r e s p o n d i n g E v i d e n c e O b j e c t s , u s i n g c o s i n e s i m i l a r i t y o f t h e i r p h r a s e e m b e d d i n g s 𝑠𝑖𝑚 𝐸𝑂 ∶= 𝐶𝑜𝑠𝑖𝑛𝑒𝑆𝑖𝑚(𝑃ℎ𝑟𝑎𝑠𝑒𝑉 𝑒𝑐(𝐸𝑣𝑆𝑡𝑟𝑢𝑐𝑡 𝑄 .𝑂𝐹 .𝐸𝑂), 𝑃ℎ𝑟𝑎𝑠𝑒𝑉 𝑒𝑐(𝐸𝑣𝑆𝑡𝑟𝑢𝑐𝑡 𝐷 .𝑂𝐹 .𝐸𝑂)) / / C o m p u t i n g s i m i l a r i t y b e t w e e n o t h e r a r g u m e n t s , u s i n g c o s i n e s i m i l a r i t y o f t h e i r p h r a s e e m b e d d i n g s 𝑛𝑢𝑚 𝑎𝑟𝑔𝑠 ∶= 0 𝑠𝑖𝑚 𝑎𝑟𝑔𝑠 ∶= 0 foreach 𝑎𝑟𝑔 ∈ (𝐸𝑣𝑆𝑡𝑟𝑢𝑐𝑡 𝑄 .𝐸𝐹 .𝑎𝑟𝑔𝑢𝑚𝑒𝑛𝑡𝑠 − {𝑉 }) do if 𝐸𝑣𝑆𝑡𝑟𝑢𝑐𝑡 𝑄 .𝐸𝐹 .𝑎𝑟𝑔 exists then 𝑠𝑖𝑚 𝑎𝑟𝑔𝑠 ∶= 𝑠𝑖𝑚 𝑎𝑟𝑔𝑠 + 𝐶𝑜𝑠𝑖𝑛𝑒𝑆𝑖𝑚(𝑃ℎ𝑟𝑎𝑠𝑒𝑉 𝑒𝑐(𝐸𝑣𝑆𝑡𝑟𝑢𝑐𝑡 𝑄 .𝐸𝐹 .𝑎𝑟𝑔), 𝑃ℎ𝑟𝑎𝑠𝑒𝑉 𝑒𝑐(𝐸𝑣𝑆𝑡𝑟𝑢𝑐𝑡 𝐷 .𝐸𝐹 .𝑎𝑟𝑔)) 𝑛𝑢𝑚 𝑎𝑟𝑔𝑠 ∶= 𝑛𝑢𝑚 𝑎𝑟𝑔𝑠 + 1 𝑠𝑖𝑚 𝑎𝑟𝑔𝑠 ∶= 𝑠𝑖𝑚 𝑎𝑟𝑔𝑠 /𝑛𝑢𝑚 𝑎𝑟𝑔𝑠 / / C o m p u t i n g o v e r a l l s i m i l a r i t y 𝑠𝑖𝑚 𝑓 𝑖𝑛𝑎𝑙 ∶= 𝑠𝑖𝑚 𝐸 × 𝑠𝑖𝑚 𝑎𝑟𝑔𝑠 × 𝑠𝑖𝑚 𝐸𝑂 / / T h e o v e r a l l s i m i l a r i t y i s m u l t i p l i e d b y t h e S e n t e n c e -B E R T b a s e d s e n t e n c e s i m i l a r i t y b e t w e e n Q a n d D 𝑠𝑖𝑚 𝑓 𝑖𝑛𝑎𝑙 ∶= 𝑠𝑖𝑚 𝑓 𝑖𝑛𝑎𝑙 × 𝐶𝑜𝑠𝑖𝑛𝑒𝑆𝑖𝑚(𝑆𝑒𝑛𝑡𝑉 𝑒𝑐(𝑄), 𝑆𝑒𝑛𝑡𝑉 𝑒𝑐(𝐷)) return 𝑠𝑖𝑚 𝑓 𝑖𝑛𝑎𝑙 Algorithm 2: 𝑆𝑒𝑚𝑀𝑎𝑡𝑐ℎ: Algorithm for computing similarity between 𝐸𝑣𝑆𝑡𝑟𝑢𝑐𝑡 𝑄 and 𝐸𝑣𝑆𝑡𝑟𝑢𝑐𝑡 𝐷 Table Example of the proposed 𝑆𝑒𝑚𝑀𝑎𝑡𝑐ℎ algorithm in action Query: T h e a u t o p s y r e p o r t r e v e a l s t h a t s o m e p o i s o n o u s c o m p o u n d s a r e f o u n d i n t h e s t o m a c h o f t h e d e c e a s e d . 𝐸𝑣𝑆𝑡𝑟𝑢𝑐𝑡 𝑄 ∶ O F = [𝑂𝑉 = r e v e a l s , 𝐸𝑂 = T h e a u t o p s y r e p o r t ]; E F = [𝐸𝑉 = f o u n d , 𝐴 1 = s o m e p o i s o n o u s c o m p o u n d s , 𝐿𝑂𝐶 = i n t h e s t o m a c h o f t h e d e c e a s e d ] Sentence: T h e r e p o r t o f t h e C h e m i c a l E x a m i n e r s h o w e d t h a t a h e a v y c o n c e n t r a t i o n o f a r s e n i c w a s f o u n d i n t h e v i s c e r a . 𝐸𝑣𝑆𝑡𝑟𝑢𝑐𝑡 𝐷 ∶ O F = [𝑂𝑉 = s h o w e d , 𝐸𝑂 = T h e r e p o r t o f t h e C h e m i c a l E x a m i n e r ]; E F = [𝐸𝑉 = f o u n d , 𝐴 1 = a h e a v y c o n c e n t r a t i o n o f a r s e n i c , 𝐿𝑂𝐶 = i n t h e v i s c e r a ] • Similarity between main predicates, their arguments and evidence objects 𝑠𝑖𝑚 𝐸 ∶= 𝐶𝑜𝑠𝑖𝑛𝑒𝑆𝑖𝑚(𝑊 𝑜𝑟𝑑𝑉 𝑒𝑐(found), 𝑊 𝑜𝑟𝑑𝑉 𝑒𝑐(found)) = 1.0 𝑠𝑖𝑚 𝐴 1 ∶= 𝐶𝑜𝑠𝑖𝑛𝑒𝑆𝑖𝑚(𝑃ℎ𝑟𝑎𝑠𝑒𝑉 𝑒𝑐(some p o i s o n o u s c o m p o u n d s ), 𝑃ℎ𝑟𝑎𝑠𝑒𝑉 𝑒𝑐(a h e a v y c o n c e n t r a t i o n o f a r s e n i c )) = 0.5469 𝑠𝑖𝑚 𝐿𝑂𝐶 ∶= 𝐶𝑜𝑠𝑖𝑛𝑒𝑆𝑖𝑚(𝑃ℎ𝑟𝑎𝑠𝑒𝑉 𝑒𝑐(in t h e s t o m a c h o f t h e d e c e a s e d ), 𝑃ℎ𝑟𝑎𝑠𝑒𝑉 𝑒𝑐(in t h e v i s c e r a )) = 0.3173 𝑠𝑖𝑚 𝑎𝑟𝑔𝑠 ∶= (𝑠𝑖𝑚 𝐴 1 + 𝑠𝑖𝑚 𝐿𝑂𝐶 )/2.0 = 0.4321 𝑠𝑖𝑚 𝐸𝑂 ∶= 𝐶𝑜𝑠𝑖𝑛𝑒𝑆𝑖𝑚(𝑃ℎ𝑟𝑎𝑠𝑒𝑉 𝑒𝑐(The a u t o p s y r e p o r t ), 𝑃ℎ𝑟𝑎𝑠𝑒𝑉 𝑒𝑐(The r e p o r t o f t h e C h e m i c a l E x a m i n e r )) = 0.8641 • Final similarity 𝑠𝑖𝑚 𝑓 𝑖𝑛𝑎𝑙 ∶= 𝑠𝑖𝑚 𝐸 × 𝑠𝑖𝑚 𝑎𝑟𝑔𝑠 × 𝑠𝑖𝑚 𝐸𝑂 × 𝑠𝑖𝑚 𝑆𝐵𝐸𝑅𝑇 = 1.0 × 0.4321 × 0.8641 × 0.607 = 0.2266 (Ranked within top 10 relevant documents) t h r e e p i e c e s o f p e l l e t s w e r e f o u n d b y t h e d o c t o r i n t h e b o d y o f d e c e a s e d M o n u . Here, except the 𝐴 1 argument (some p o i s o n o u s c o m p o u n d s vs t h r e e p i e c e s o f p e l l e t s ) in Evidence Structure instances, other arguments are similar in meaning. We get cosine similarity of 0.36 between s o m e p o i s o n o u s c o m p o u n d s and t h r e e p i e c e s o f p e l l e t s which is misleading. It is not too low as compared to another case where there are semantically similar argument phrases (e.g., cosine similarity between s o m e p o i s o n o u s c o m p o u n d s and a h e a v y c o n c e n t r a t i o n o f a r s e n i c is just 0.55 as shown in Table</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_7"><head></head><label></label><figDesc>I n s t e a d o f s u r r e n d e r i n g b e f o r e t h e p o l i c e , t h e d e c e a s e d h a d a t t e m p t e d t o k i l l t h e p o l i c e . I n r e t a l i a t i o n , h e w a s s h o t b y t h e m .</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>EvidenceObject or EO: The Evidence Object in</head><label></label><figDesc>r e p o r t , , a f f i d a v i t , l e t t e r , c h e q u e , a g r e e m e n t , p e t i t i o n , F I R , s i g n a t u r e ) • Material objects (gun, b u l l e t , c l o t h e s , k e r o s e n e c a n ) • Substances (poison, a l c o h o l , k e r o s e n e ) In Indian court case documents, such Evidence Objects are also represented in the judgement document as E x h i b i t A , E x . 2 , E v i d e n c e 2 3 and so on. E F = [𝐸𝑉 = d i s h o n o u r e d , 𝐴 0 = T h e b a n k , 𝐴 1 = t h e c h e q u e , 𝐶𝐴𝑈 = d u e t o i n s u f f i c i e n t b a l a n c e ]</figDesc><table><row><cell>Table 1</cell><cell></cell></row><row><cell>Example Evidence sentences with their Evidence Structure Instances</cell><cell></cell></row><row><cell>T h e b a n k d i s h o n o u r e d t h e c h e q u e d u e t o i n s u f f i c i e n t b a l a n c e .</cell><cell></cell></row><row><cell cols="2">T h e r e p o r t r e v e a l e d t h a t o r g a n o -p h o s p h o r u s c o m p o u n d w a s f o u n d i n t h e s t o m a c h , s m a l l i n t e s t i n e s , l a r g e i n t e s t i n e s</cell></row><row><cell>, l i v e r , s p l e e n , k i d n e y a n d b r a i n o f t h e d e c e a s e d .</cell><cell></cell></row><row><cell>• O F = [𝑂𝑉 = r e v e a l e d , 𝐸𝑂 = T h e r e p o r t ]</cell><cell></cell></row><row><cell cols="2">E F = [𝐸𝑉 = f o u n d , 𝐿𝑂𝐶 = i n t h e s t o m a c h , s m a l l i n t e s t i n e s , l a r g e i n t e s t i n e s , l i v e r , s p l e e n , k i d n e y a n d b r a i n o f</cell></row><row><cell>t h e d e c e a s e d ]</cell><cell></cell></row><row><cell cols="2">T h e M a g i s t r a t e f o u n d p r i m a f a c i e e v i d e n c e t h a t t h e a p p e l l a n t h a d f r a u d u l e n t l y u s e d i n t h e C i v i l S u i t f o r g e d c h e q u e</cell></row><row><cell>a n d c o m m i t t e d h i m t o t h e S e s s i o n s f o r t r i a l</cell><cell></cell></row><row><cell>• O F = [𝑂𝑉 = f o u n d , 𝑂𝐴 = T h e M a g i s t r a t e , 𝐸𝑂 = p r i m a f a c i e e v i d e n c e ]</cell><cell>p o s t -m o r t e m</cell></row><row><cell cols="2">E F = [𝐸𝑉 = u s e d ,𝐴 0 = t h e a p p e l l a n t , 𝐴 1 = f o r g e d c h e q u e , 𝐿𝑂𝐶 = i n t h e C i v i l S u i t ] T h e p r o s e c u t i o n c a s e w a s t h a t t h o u g h t h e r o u g h c a s h b o o k s h o w e d t h a t o n S e p t e m b e r 2 9 , 1 9 5 0 a s u m o f R s . 2 1 , 1 3 3 w a s s e n t t o t h e T r e a s u r y b y a p p e l l a n t G u p t a , t h e T r e a s u r y f i g u r e s i n t h e c h a l l a n s h o w e d t h a t o n t h a t d a y o n l y a s u m r e p o r t On these lines, we define an Evidence Sentence as any o f R s . 1 , 1 3 3 w a s d e p o s i t e d i n t o t h e T r e a s u r y a n d t h u s a s u m o f R s . 2 0 , 0 0 0 w a s d i s h o n e s t l y m i s a p p r o p r i a t e d .</cell></row><row><cell cols="2">sentence containing one or more Evidence Objects rele-</cell></row><row><cell cols="2">vant to the current case but do not consist of</cell></row><row><cell cols="2">• any witness testimony which is not verifiable</cell></row><row><cell>• legal argumentation</cell><cell></cell></row><row><cell cols="2">• a reference to some prior case or some Act or</cell></row><row><cell>Section</cell><cell></cell></row><row><cell cols="2">• directions or instructions given by the court or</cell></row><row><cell>judge.</cell><cell></cell></row><row><cell cols="2">We now present a formal definition of the Evidence</cell></row><row><cell cols="2">Structure. For every evidence present in an Evidence</cell></row><row><cell cols="2">Sentence, the structure consists of an optional Observation</cell></row><row><cell cols="2">Frame and a mandatory Evidence Frame. The Observation</cell></row><row><cell cols="2">Frame represents the source of the information and the</cell></row><row><cell cols="2">agent disclosing it. This information is optional as it may</cell></row><row><cell cols="2">or may not be explicitly stated in a sentence. It consists</cell></row><row><cell>of the following arguments:</cell><cell></cell></row><row><cell cols="2">• ObserverVerb or OV: The verb indicating</cell></row><row><cell cols="2">the observation/discovery/disclosure (e.g., f o u n d ,</cell></row><row><cell>r e v e a l e d , s t a t e d )</cell><cell></cell></row><row><cell cols="2">• ObserverAgent or A 0 : The source disclosing</cell></row><row><cell cols="2">the information (e.g., p e r s o n , a g e n c y , a u t h o r i t y )</cell></row><row><cell cols="2">• focus (e.g., p o s t -m o r t e m r e p o r t , F I R , l e t t e r )</cell></row><row><cell cols="2">The Evidence Frame captures details about the evidence</cell></row><row><cell>itself through the following arguments:</cell><cell></cell></row><row><cell cols="2">• EvidenceVerb or EV: the main verb of any ac-</cell></row><row><cell cols="2">tion, event or fact mentioned in a sentence or re-</cell></row><row><cell cols="2">vealed by the Evidence Object (e.g., k i l l e d , f o r g e d ,</cell></row><row><cell>e s c a p e d )</cell><cell></cell></row><row><cell cols="2">• Agent or A 0 : someone who initiates the action</cell></row><row><cell cols="2">indicated by the EvidenceVerb (e.g., t h e a c c u s e d ,</cell></row><row><cell>R a m , A B C P v t . L t d . )</cell><cell></cell></row><row><cell cols="2">• Patient or A 1 : someone who undergoes the ac-</cell></row><row><cell cols="2">tion indicated by the EvidenceVerb. (e.g., t h e</cell></row></table><note>•</note></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head></head><label></label><figDesc>This corresponds to the 𝑔𝑒𝑡_𝑒𝑣𝑖𝑑𝑒𝑛𝑐𝑒_𝑜𝑏𝑗𝑒𝑐𝑡 function used in Algorithm 1. Observation Frames that do not contain a corresponding Evidence Frame are redesigned as stand alone Evidence Frames. We finally combine the Evidence Frame and the Observation Frame into an Evidence Structure Instance.We measured the accuracy of 260 Evidence Structure Instances obtained from 100 random Evidence and Testimony sentences. The accuracy of the Observation Frame extraction is 86% and that of Evidence Frame extraction is 88%. We observed that most of the incorrect extractions were due to parsing error in the SRL model.</figDesc><table /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head></head><label></label><figDesc>t , r e v e a l , s a y , s h o w , s t a t e , s u b m i t , s u g g e s t , t e l l , w i t h d r a w }, 𝑁 𝐸𝐺_𝑊 𝑂𝑅𝐷𝑆 = {no, n o t , n e i t h e r , n o r , n e v e r } 𝑠 foreach 𝑃 ∈ 𝑆𝑅𝐿_𝑃 such that 𝑃.𝑉 ∈ 𝑂𝐵𝑆_𝑉 𝐸𝑅𝐵𝑆 do 𝑂𝐹 ∶= Create empty Observation Frame 𝑂𝐹 .𝑉 ∶= 𝑃.𝑉 𝑂𝐹 .𝑁 𝐸𝐺 ∶= 𝑃.𝐴𝑅𝐺𝑀-𝑁 𝐸𝐺 𝑂𝐹 .𝐴 0 ∶= 𝑃.𝐴𝑅𝐺 0 𝑂𝐹 .𝐴 1 ∶= 𝑃.𝐴𝑅𝐺 1 // If any of the arguments of the predicate starts with a negative word, then we negate the verb. if 𝑂𝐹 .𝐴 0 or 𝑂𝐹 .𝐴 1 starts with any word from 𝑁 𝐸𝐺_𝑊 𝑂𝑅𝐷𝑆 then 𝑂𝐹 .𝑁 𝐸𝐺 ∶= 𝑇 𝑟𝑢𝑒 𝑂𝐹 .𝐸𝑂 ∶= 𝑔𝑒𝑡_𝑒𝑣𝑖𝑑𝑒𝑛𝑐𝑒_𝑜𝑏𝑗𝑒𝑐𝑡(𝑃.𝐴𝑅𝐺 0 ) ∪ 𝑔𝑒𝑡_𝑒𝑣𝑖𝑑𝑒𝑛𝑐𝑒_𝑜𝑏𝑗𝑒𝑐𝑡(𝑃.𝐴𝑅𝐺𝑀-𝐿𝑂𝐶) 𝑂𝐹 𝑠 ∶= 𝑂𝐹 𝑠 ∪ {𝑂𝐹 } / / O b t a i n c o r r e s p o n d i n g E v i d e n c e F r a m e s f o r e v e r y O b s e r v a t i o n F r a m e foreach 𝑂𝐹 ∈ 𝑂𝐹 𝑠 do 𝐹 𝑜𝑢𝑛𝑑𝐸𝐹 ∶= 𝐹 𝑎𝑙𝑠𝑒 foreach 𝑃 ∈ 𝑆𝑅𝐿_𝑃 such that 𝑃.𝑉 occurs within the span of 𝑂𝐹 .𝐴 1 do if 𝑃.𝑉 is a copula verb and any of 𝑃.𝐴𝑅𝐺 0 or 𝑃.𝐴𝑅𝐺 1 does not exist then continue 𝐸𝐹 ∶= Create empty Evidence Frame 𝐸𝐹 .𝑉 ∶= 𝑃.𝑉 𝐸𝐹 .𝑁 𝐸𝐺 ∶= 𝑃.𝐴𝑅𝐺𝑀-𝑁 𝐸𝐺 // If any of the arguments of the predicate starts with a negative word, then we negate the verb if 𝑂𝐹 .𝐴 0 or 𝑂𝐹 .𝐴 1 starts with any word from 𝑁 𝐸𝐺_𝑊 𝑂𝑅𝐷𝑆 then 𝐸𝐹 .𝑁 𝐸𝐺 ∶= 𝑇 𝑟𝑢𝑒 𝐸𝑣𝑆𝑡𝑟𝑢𝑐𝑡 𝑄 : Evidence Structure Instance from a query sentence 𝑄 𝐸𝑣𝑆𝑡𝑟𝑢𝑐𝑡 𝐷 : Evidence Structure Instance from a sentence 𝐷 in the corpus output : Similarity score between 𝐸𝑣𝑆𝑡𝑟𝑢𝑐𝑡 𝑄 and 𝐸𝑣𝑆𝑡𝑟𝑢𝑐𝑡 𝐷 / / C h e c k i n g f o r n e g a t i o n if 𝐸𝑣𝑆𝑡𝑟𝑢𝑐𝑡 𝑄 .𝑂𝐹 .𝑁 𝐸𝐺 ≠ 𝐸𝑣𝑆𝑡𝑟𝑢𝑐𝑡 𝐷 .𝑂𝐹 .𝑁 𝐸𝐺 then return 0 if 𝐸𝑣𝑆𝑡𝑟𝑢𝑐𝑡 𝑄 .𝐸𝐹 .𝑁 𝐸𝐺 ≠ 𝐸𝑣𝑆𝑡𝑟𝑢𝑐𝑡 𝐷 .𝐸𝐹 .𝑁 𝐸𝐺 then return 0</figDesc><table><row><cell>input :</cell></row><row><cell>𝐸𝑣𝑆𝑡𝑟𝑢𝑐𝑡𝑠 ∶= ∅</cell></row><row><cell>𝑂𝐹 𝑠 ∶= ∅</cell></row><row><cell>/ / O b t a i n O b s e r v a t i o n F r a m e s i n t h e s e n t e n c e foreach argument 𝐴𝑅𝐺 ∈ 𝑃.𝑎𝑟𝑔𝑢𝑚𝑒𝑛𝑡𝑠 do</cell></row><row><cell>𝐸𝐹 .𝐴𝑅𝐺 ∶= 𝑃.𝐴𝑅𝐺</cell></row><row><cell>delete(𝑂𝐹 .𝐴 1 )</cell></row><row><cell>𝐸𝑣𝑆𝑡𝑟𝑢𝑐𝑡 ∶= {(𝑂𝐹 , 𝐸𝐹 )}</cell></row><row><cell>𝐸𝑣𝑆𝑡𝑟𝑢𝑐𝑡𝑠 ∶= 𝐸𝑣𝑆𝑡𝑟𝑢𝑐𝑡𝑠 ∪ 𝐸𝑣𝑆𝑡𝑟𝑢𝑐𝑡</cell></row><row><cell>𝐹 𝑜𝑢𝑛𝑑𝐸𝐹 ∶= 𝑇 𝑟𝑢𝑒</cell></row><row><cell>/ / I f n o E v i d e n c e F r a m e e x i s t s f o r a n O b s e r v a t i o n F r a m e , t r a n s f e r t h e O b s e r v a t i o n F r a m e t o t h e E v i d e n c e</cell></row><row><cell>F r a m e</cell></row><row><cell>if FoundEF == False then</cell></row><row><cell>𝐸𝐹 ∶= Create empty Evidence Frame</cell></row><row><cell>𝐸𝐹 .𝑉 ∶= 𝑂𝐹 .𝑉</cell></row><row><cell>𝑃 ∶= 𝑃 ′ ∈ 𝑆𝑅𝐿_𝑃 such that 𝑃 ′ .𝑉 = 𝑂𝐹 .𝑉</cell></row><row><cell>foreach argument 𝐴𝑅𝐺 ∈ 𝑃.𝑎𝑟𝑔𝑢𝑚𝑒𝑛𝑡𝑠 do</cell></row><row><cell>𝐸𝐹 .𝐴𝑅𝐺 ∶= 𝑃.𝐴𝑅𝐺</cell></row><row><cell>clear(𝑂𝐹)</cell></row><row><cell>𝑂𝐹 .𝐸𝑂 ∶= 𝑔𝑒𝑡_𝑒𝑣𝑖𝑑𝑒𝑛𝑐𝑒_𝑜𝑏𝑗𝑒𝑐𝑡(𝑃.𝐴𝑅𝐺 0 ) ∪ 𝑔𝑒𝑡_𝑒𝑣𝑖𝑑𝑒𝑛𝑐𝑒_𝑜𝑏𝑗𝑒𝑐𝑡(𝑃.𝐴𝑅𝐺𝑀-𝐿𝑂𝐶)</cell></row><row><cell>//Add all the required arguments to Evidence Frame</cell></row><row><cell>𝐸𝑣𝑆𝑡𝑟𝑢𝑐𝑡 ∶= {(𝑂𝐹 , 𝐸𝐹 )}</cell></row><row><cell>𝐸𝑣𝑆𝑡𝑟𝑢𝑐𝑡𝑠 ∶= 𝐸𝑣𝑆𝑡𝑟𝑢𝑐𝑡𝑠 ∪ 𝐸𝑣𝑆𝑡𝑟𝑢𝑐𝑡</cell></row><row><cell>return(𝐸𝑣𝑆𝑡𝑟𝑢𝑐𝑡𝑠)</cell></row><row><cell>Algorithm 1: 𝑔𝑒𝑡_𝑒𝑣𝑖𝑑𝑒𝑛𝑐𝑒_𝑠𝑡𝑟𝑢𝑐𝑡𝑢𝑟𝑒_𝑖𝑛𝑠𝑡𝑎𝑛𝑐𝑒𝑠: Algorithm for instantiating Evidence Structure for a sentence</cell></row></table></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_0">https://pypi.org/project/rank-bm25/</note>
		</body>
		<back>
			<div type="annex">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Table 6</head><p>Evaluation of various techniques for the task of prior case retrieval. All entries are of the form (R-Prec; Avg. Precision). (Note: Our proposed approach 𝑆𝑒𝑚𝑀𝑎𝑡𝑐ℎ is referred as 𝑆𝑀. Underlines indicate the best performing results for each query across multiple techniques) W h a t a r e t h e c a s e s w h e r e ⋯ 𝑄  Siamese-BERT networks to obtain more meaningful sentence embeddings as compared to vanilla BERT <ref type="bibr" target="#b15">[16]</ref>. We used the pre-trained model b e r t -b a s e -n l i -s t s b -m e a nt o k e n s to obtain sentence embeddings for sentences. Following Ghosh et al. <ref type="bibr" target="#b0">[1]</ref>, we use the pre-trained model as it is and did not fine-tune it further. This is because such fine-tuning needs annotated sentence pairs with labels indicating whether the sentences in the pair are semantically similar or not. Such annotated dataset is expensive to create and our aim is to avoid any dependence on manually annotated training data. Similar to Ghosh et al. <ref type="bibr" target="#b0">[1]</ref>, we used sentence embeddings obtained by Sentence-BERT to compute cosine similarity between a query sentence and a candidate sentence in a document. The overall similarity of a document with a query is the maximum cosine similarity obtained for any of its sentences with the query sentence. We use 3 settings considering different sentences in each document:</p><p>• 𝑆𝐵 𝑇 𝐸 : Only Testimony or Evidence sentences • 𝑆𝐵 𝑇 : Only Testimony sentences • 𝑆𝐵 𝐸 : Only Evidence sentences</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.3.">Evaluation</head><p>All the baseline techniques and our proposed technique are evaluated using a set of queries and using certain evaluation metrics to evaluate and compare the ranked lists produced by each of these techniques.</p><p>Queries: We chose 10 queries (shown in Table <ref type="table">6</ref>) which represent cases and evidence objects of diverse nature (domestic violence, financial fraud etc.).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Ground Truth:</head><p>We created a set of gold-standard relevant documents for each query using the standard pooling technique <ref type="bibr" target="#b16">[17]</ref>. We ran the following techniques to produce a ranked list of documents for each query -𝐵𝑀25 𝑎𝑙𝑙 , 𝐵𝑀25 𝑇 𝐸 , 𝑆𝐵 𝑇 𝐸 , and our proposed technique 𝑆𝑒𝑚𝑀𝑎𝑡𝑐ℎ 𝑇 𝐸 . We chose top 10 documents from the ranked list produced by each technique. Human experts verified the relevance of each document for the query. Finally, after discarding all the irrelevant documents, we got a set of gold-standard relevant documents for each query 2 . Metrics: We used R-Precision and Average Precision as our evaluation metrics <ref type="bibr" target="#b16">[17]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">R-Precision (R-Prec):</head><p>This calculates the the number of relevant documents observed at 𝑅.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Average Precision (AP):</head><p>This captures the joint effect of Precision and Recall. It computes precision at each rank of the predicted ranked list and then computes mean of these precision values.</p><p>2 This dataset can be obtained from the authors on request</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.4.">Results</head><p>Table <ref type="table">6</ref> shows comparative evaluation results for various baselines and our proposed technique. Average performance of 𝐵𝑀25 𝑇 𝐸 is better than 𝐵𝑀25 𝑎𝑙𝑙 indicating that considering only Evidence and Testimony sentences for representing any document, results in better prior case retrieval performance. Other two baselines 𝑆𝐵 (Sentence-BERT) and 𝑆𝑀 (our proposed technique 𝑆𝑒𝑚𝑀𝑎𝑡𝑐ℎ) also consider only Evidence and Testimony sentences rather than considering all the sentences in a document. All the baselines which consider only Testimony sentences, perform poorly as compared to the corresponding techniques using both Testimony and Evidence sentences. This highlights the importance of evidence information as compared to using only witness testimony information for prior case retrieval as done in Ghosh et al. <ref type="bibr" target="#b0">[1]</ref>.</p><p>Considering the average performance across all the 10 queries, our proposed technique 𝑆𝑀 𝑇 𝐸 is the best performing technique in terms of both R-Prec and AP. The performance of 𝑆𝑀 𝑇 𝐸 is the most consistent across the diverse queries. It achieves minimum R-Prec of 0.24 (for 𝑄 1 ) as compared to other baselines like 𝐵𝑀25 𝑎𝑙𝑙 , 𝐵𝑀25 𝑇 𝐸 and 𝑆𝐵 𝑇 𝐸 which have minimum R-Prec of 0 for some queries. As described in Algorithm 2, 𝑆𝑀 uses Sentence-BERT based similarity within sentences for producing an enhanced matching score. We experimented with a variant of 𝑆𝑀 which does not rely on Sentence-BERT based similarity. This variant resulted in average R-Prec of 0.36 and MAP of 0.30 across all the 10 queries. Although this is lower than 𝑆𝑀 𝑇 𝐸 performance, the R-Prec is still comparable with 𝐵𝑀25 𝑇 𝐸 (avg R-Prec of 0.36) and better than that of 𝑆𝐵 𝑇 𝐸 (avg R-Prec of 0.28).</p><p>For some queries, it is important to have some semantic understanding at sentence-level. For example, 𝑄 4 , which contains "negation", 𝑆𝐵 and 𝑆𝑀 can capture the query's meaning in a better way. 𝑆𝑀 such negations in a more principled manner as the Evidence Structure Instance captures negation as one of its arguments.</p><p>For 𝑆𝑀, the maximum matching score achieved for any Evidence Structure Instance in a document, is considered as the overall matching score with the whole document. In contrast, 𝐵𝑀25 based techniques directly compute matching score for the whole document as they do not rely on sentence structure. This is one limitation of 𝑆𝑀 which we plan to address as a future work. However, as 𝑆𝑀 computes matching scores for individual Evidence Structure instances, it is able to provide better interpretation for each relevant document in terms of the actual sentences which provided the maximum matching score. Analysis of errors: We analyzed cases where 𝑆𝑀 𝑇 𝐸 was assigned a lower score to a relevant document or a higher score to a non-relevant document. We discovered 3 main reasons -missing or incorrect arguments within Evidence Structure instances, misleading high similarity between argument phrases and presence of co-references. Consider the following sentence for which 𝑆𝑀 𝑇 𝐸 incorrectly assigns a high score for query 𝑄 5 (see Table <ref type="table">6</ref>) -T h e p o l i c e r e p o r t a l s o r e v e a l s t h a t</p></div>			</div>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Retrieval of prior court cases using witness testimonies</title>
		<author>
			<persName><forename type="first">K</forename><surname>Ghosh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Pawar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Palshikar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Bhattacharyya</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Varma</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">JURIX</title>
		<imprint>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">The proposition bank: An annotated corpus of semantic roles</title>
		<author>
			<persName><forename type="first">M</forename><surname>Palmer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Gildea</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Kingsbury</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Computational linguistics</title>
		<imprint>
			<biblScope unit="volume">31</biblScope>
			<biblScope unit="page" from="71" to="106" />
			<date type="published" when="2005">2005</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<monogr>
		<title level="m" type="main">Simple BERT models for relation extraction and semantic role labeling</title>
		<author>
			<persName><forename type="first">P</forename><surname>Shi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Lin</surname></persName>
		</author>
		<idno>CoRR abs/1904.05255</idno>
		<ptr target="http://arxiv.org/abs/1904.05255.arXiv:1904.05255" />
		<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Long short-term memory</title>
		<author>
			<persName><forename type="first">S</forename><surname>Hochreiter</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Schmidhuber</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Neural computation</title>
		<imprint>
			<biblScope unit="volume">9</biblScope>
			<biblScope unit="page" from="1735" to="1780" />
			<date type="published" when="1997">1997</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<monogr>
		<title level="m" type="main">spaCy: Industrial-strength Natural Language Processing in Python</title>
		<author>
			<persName><forename type="first">M</forename><surname>Honnibal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Montani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Van Landeghem</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Boyd</surname></persName>
		</author>
		<idno type="DOI">10.5281/zenodo.1212303</idno>
		<ptr target=":10.5281/zenodo.1212303" />
		<imprint>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Wordnet: a lexical database for english</title>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">A</forename><surname>Miller</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Communications of the ACM</title>
		<imprint>
			<biblScope unit="volume">38</biblScope>
			<biblScope unit="page" from="39" to="41" />
			<date type="published" when="1995">1995</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Glove: Global vectors for word representation</title>
		<author>
			<persName><forename type="first">J</forename><surname>Pennington</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Socher</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">D</forename><surname>Manning</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2014 conference on empirical methods in natural language processing (EMNLP)</title>
				<meeting>the 2014 conference on empirical methods in natural language processing (EMNLP)</meeting>
		<imprint>
			<date type="published" when="2014">2014</date>
			<biblScope unit="page" from="1532" to="1543" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Sentence-BERT: Sentence Embeddings using Siamese BERT-networks</title>
		<author>
			<persName><forename type="first">N</forename><surname>Reimers</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Gurevych</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP)</title>
				<meeting>the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP)</meeting>
		<imprint>
			<date type="published" when="2019">2019</date>
			<biblScope unit="page" from="3973" to="3983" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Overview of INEX</title>
		<author>
			<persName><forename type="first">P</forename><surname>Bellot</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Doucet</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Geva</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Gurajada</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Kamps</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Kazai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Koolen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Mishra</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Moriceau</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Mothe</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Preminger</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Sanjuan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Schenkel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Tannier</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Theobald</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Trappett</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><surname>Wang</surname></persName>
		</author>
		<idno type="DOI">10.1007/978-3-642-40802-1_27</idno>
		<idno>doi:</idno>
		<ptr target="10.1007/978-3-642-40802-1\_27" />
	</analytic>
	<monogr>
		<title level="m">Information Access Evaluation. Multilinguality, Multimodality, and Visualization -4th International Conference of the CLEF Initiative, CLEF 2013</title>
		<title level="s">Lecture Notes in Computer Science</title>
		<editor>
			<persName><forename type="first">P</forename><surname>Forner</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">H</forename><surname>Müller</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">R</forename><surname>Paredes</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">P</forename><surname>Rosso</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">B</forename><surname>Stein</surname></persName>
		</editor>
		<meeting><address><addrLine>Valencia, Spain</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2013-09-23">2013. September 23-26, 2013. 2013</date>
			<biblScope unit="volume">8138</biblScope>
			<biblScope unit="page" from="269" to="281" />
		</imprint>
	</monogr>
	<note>Proceedings</note>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Evidence finding using a collection of books</title>
		<author>
			<persName><forename type="first">M.-A</forename><surname>Cartright</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><forename type="middle">A</forename><surname>Feild</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Allan</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 4th ACM workshop on Online books, complementary social media and crowdsourcing</title>
				<meeting>the 4th ACM workshop on Online books, complementary social media and crowdsourcing</meeting>
		<imprint>
			<date type="published" when="2011">2011</date>
			<biblScope unit="page" from="11" to="18" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Show me your evidence-an automatic method for context dependent evidence detection</title>
		<author>
			<persName><forename type="first">R</forename><surname>Rinott</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Dankin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Alzate</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">M</forename><surname>Khapra</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Aharoni</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Slonim</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2015 conference on empirical methods in natural language processing</title>
				<meeting>the 2015 conference on empirical methods in natural language processing</meeting>
		<imprint>
			<date type="published" when="2015">2015</date>
			<biblScope unit="page" from="440" to="450" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">An end-to-end joint model for evidence information extraction from court record document</title>
		<author>
			<persName><forename type="first">D</forename><surname>Ji</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Tao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Fei</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Ren</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Information Processing &amp; Management</title>
		<imprint>
			<biblScope unit="volume">57</biblScope>
			<biblScope unit="page">102305</biblScope>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">A new conceptual framework for enhancing legal information retrieval at the brazilian superior court of justice</title>
		<author>
			<persName><forename type="first">T</forename><surname>Gomes</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Ladeira</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 12th International Conference on Management of Digital EcoSystems</title>
				<meeting>the 12th International Conference on Management of Digital EcoSystems</meeting>
		<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="26" to="29" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">Extending full text search for legal document collections using word embeddings</title>
		<author>
			<persName><forename type="first">J</forename><surname>Landthaler</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Waltl</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Holl</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Matthes</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">JURIX</title>
		<imprint>
			<biblScope unit="page" from="73" to="82" />
			<date type="published" when="2016">2016</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">Improvements to bm25 and language models examined</title>
		<author>
			<persName><forename type="first">A</forename><surname>Trotman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Puurula</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Burgess</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2014 Australasian Document Computing Symposium</title>
				<meeting>the 2014 Australasian Document Computing Symposium</meeting>
		<imprint>
			<date type="published" when="2014">2014</date>
			<biblScope unit="page" from="58" to="65" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<monogr>
		<author>
			<persName><forename type="first">J</forename><surname>Devlin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M.-W</forename><surname>Chang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Lee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Toutanova</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1810.04805</idno>
		<title level="m">BERT: Pre-training of deep bidirectional transformers for language understanding</title>
				<imprint>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">Introduction to information retrieval</title>
		<author>
			<persName><forename type="first">C</forename><surname>Manning</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Raghavan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Schutze</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Natural Language Engineering</title>
		<imprint>
			<biblScope unit="volume">16</biblScope>
			<biblScope unit="page" from="100" to="103" />
			<date type="published" when="2010">2010</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
