=Paper= {{Paper |id=Vol-2888/paper1 |storemode=property |title=Prior Case Retrieval using Evidence Extraction from Court Judgements |pdfUrl=https://ceur-ws.org/Vol-2888/paper1.pdf |volume=Vol-2888 |authors=Basit Ali,Ravina More,Sachin Pawar,Girish K. Palshikar |dblpUrl=https://dblp.org/rec/conf/icail/AliMPP21 }} ==Prior Case Retrieval using Evidence Extraction from Court Judgements== https://ceur-ws.org/Vol-2888/paper1.pdf
Prior Case Retrieval using Evidence Extraction from Court
Judgements
Basit Ali, Ravina More, Sachin Pawar and Girish K. Palshikar
TCS Research, Pune, India.


                                          Abstract
                                          One of the key constituents of court case descriptions is Evidence description and observations. Along with witness testimonies,
                                          evidence plays a significant role in the final decision of the case. We propose a weakly supervised technique to automatically
                                          identify sentences containing evidences. We represent the information related to evidences in these sentences in a semantically
                                          rich structure – Evidence Structure defined as an Evidence Information Model. We show that witness testimony information
                                          can also be represented using the same model. We demonstrate the effectiveness of our Evidence Information Model for the
                                          prior case retrieval application by proposing a matching algorithm for computing semantic similarity between a query and a
                                          sentence in a court case description. To the best of our knowledge, this is the first paper to apply NLP techniques for the
                                          extraction of evidence information from court judgements and use it for retrieving relevant prior court cases.

                                          Keywords
                                          Evidence Extraction, Evidence Information Model, Natural Language Processing, Prior Case Retrieval



1. Introduction                                                                                                    testimony information. Here, we use the rules proposed
                                                                                                                   in Ghosh et al. [1] for identification of witness testimonies
Evidences - typically based on documents (e.g., letter, re-                                                        and design new rules for identification of evidence sen-
ceipt, report, agreements, affidavits) and physical objects                                                        tences. In the second step, we train a Weakly Supervised
(e.g., knife, guns, photos, phone call data records) - are                                                         Sentence Classifier whose training data is automatically
often used by lawyers in their arguments during a court                                                            created using the sentences identified by the linguistic
case. The observations made through these evidences                                                                rules. It is a multi-label classifier which predicts whether
may have a significant effect on the judges’ final decision.                                                       any sentence contains an Evidence or Witness Testimony
In order to develop a deeper understanding of the past                                                             or both. Once all the Evidence and Testimony sentences
court cases, it is valuable to identify various Evidences                                                          are identified from the corpus of court judgements, we
discussed in these cases and the observations which are                                                            propose a Semantic Role Labelling (SRL) [2] based tech-
made about them or through them. Such information                                                                  nique to automatically instantiate Evidence Structures
about evidences has several applications such as under-                                                            for these sentences.
standing and representing legal arguments, determining                                                                To demonstrate effectiveness of the proposed Evidence
strengths and weaknesses of those arguments, identify-                                                             Structure, we discuss its use in the prior case retrieval ap-
ing relevant past cases in which similar evidences were                                                            plication. We propose a matching algorithm for comput-
discussed, etc.                                                                                                    ing semantic similarity between a query and a sentence
   In this paper, we discuss Natural Language Process-                                                             in a court judgement document. This algorithm makes
ing (NLP) based techniques for extracting information                                                              use of the proposed Evidence Structure in which both
regarding Evidences mentioned in court judgement doc-                                                              the query and the sentence are represented, resulting in
uments. We propose to represent this information in a                                                              a semantically sound similarity score between them.
rich semantic structure – Evidence Structure defined as                                                               Previously, Ghosh et al. [1] identified witness testi-
an Evidence Information Model. Along with Evidences,                                                               monies from court case documents and used them for
we also identify and represent Witness Testimonies using                                                           retrieving relevant prior cases. We propose that con-
the same Information Model. Initially, we discuss a two-                                                           sidering only witness testimonies leads to loss of key
step approach for identifying evidence and testimony                                                               information regarding evidences mentioned in a case.
sentences. In the first step, linguistic rules are used to                                                         Hence, we identify and use information about various
determine whether a sentence contains any evidence or                                                              evidences mentioned in the case documents leading to
Proceedings of the Fifth Workshop on Automated Semantic Analysis of                                                much better prior case retrieval performance as demon-
Information in Legal Text (ASAIL 2021), June 25, 2021, São Paulo,                                                  strated in the experiments section. Moreover, Ghosh et
Brazil.                                                                                                            al. [1] use a much limited semantic structure to repre-
Envelope-Open ali.basit@tcs.com (B. Ali); ravina.m@tcs.com (R. More);                                              sent information regarding events mentioned in witness
sachin7.p@tcs.com (S. Pawar); gk.palshikar@tcs.com
                                                                                                                   testimonies. This structure does not capture important
(G. K. Palshikar)
                                    © 2021 Copyright for this paper by its authors. Use permitted under Creative   semantic information like whether event is negated, what
                                    Commons License Attribution 4.0 International (CC BY 4.0).
 CEUR
 Workshop
 Proceedings
               http://ceur-ws.org
               ISSN 1613-0073
                                    CEUR Workshop Proceedings (CEUR-WS.org)                                        are the causes behind the event, the manner in which
event takes place etc. We propose a richer semantic struc-       2.2. Evidence Structure
ture addressing these limitations and design a suitable
                                                                 The Evidence Information Model represents every Evi-
semantic matching algorithm for that structure. To the
                                                                 dence Sentence giving information about one or more
best of our knowledge, this is the first paper to apply
                                                                 Evidence Objects in an Evidence Structure. We define an
NLP techniques for the extraction of evidence informa-
                                                                 Evidence Object as one of the objects presented by the
tion from court judgements and demonstrate its use for
                                                                 counsels to the judge along with the information and
retrieving relevant prior court cases.
                                                                 findings about the crime. It is thus, a physical entity
                                                                 that can furnish some degree of support, contradiction
2. Evidence Information Model                                    or opposition to some legal arguments. Some examples
                                                                 of Evidence Objects are:
The purpose of Evidence Information Model is to define                    • Documents (a u t o p s y r e p o r t ,                   post-mortem
a suitable structure to represent evidence information in                   report, affidavit, letter, cheque, agreement,
court judgements. In this section, we describe Semantic                     petition, FIR, signature)
Role Labelling in brief and how it is used to define our                  • Material objects (g u n , b u l l e t , c l o t h e s , k e r o s e n e
proposed Evidence Structure.                                                can)
                                                                          • Substances (p o i s o n , a l c o h o l , k e r o s e n e )
2.1. Background: Semantic Role                                   In Indian court case documents, such Evidence Objects are
                                                                 also represented in the judgement document as E x h i b i t
     Labelling                                                   A , E x . 2 , E v i d e n c e 2 3 and so on.
Semantic Role Labelling (SRL) is a technique in Natural              On these lines, we define an Evidence Sentence as any
Language Processing that identifies verbs/predicates in a        sentence containing one or more Evidence Objects rele-
sentence, finds phrases connected to every predicate and         vant to the current case but do not consist of
assigns an appropriate semantic role to every phrase. By                  • any witness testimony which is not verifiable
doing so, SRL helps machines to understand the roles of                   • legal argumentation
important words within a sentence. Following are some                     • a reference to some prior case or some Act or
key semantic roles identified for a verb/predicate (often                   Section
corresponding to an action or event) by SRL techniques:                   • directions or instructions given by the court or
ARG0 : proto-agent or someone who performs the action                       judge.
denoted by the verb                                                  We now present a formal definition of the Evidence
ARG1 : proto-patient or someone on whom the action               Structure. For every evidence present in an Evidence
is performed on                                                  Sentence, the structure consists of an optional Observation
ARGM-TMP : the time when the event took place                    Frame and a mandatory Evidence Frame. The Observation
ARGM-CAU : the cause of the action                               Frame represents the source of the information and the
ARGM-PRP : the purpose of the action                             agent disclosing it. This information is optional as it may
ARGM-LOC : the location where the event took place               or may not be explicitly stated in a sentence. It consists
ARGM-MNR : the manner in which the action took                   of the following arguments:
place                                                                     • ObserverVerb or OV: The verb indicating
ARGM-NEG : the word indicating that the action did                          the observation/discovery/disclosure (e.g., f o u n d ,
not take place                                                              revealed, stated)
   Consider the following example sentence:                               • ObserverAgent or A0 : The source disclosing
On August 25, 1965, the bank dishonoured the cheque                         the information (e.g., p e r s o n , a g e n c y , a u t h o r i t y )
due to insufficient balance.                                              • EvidenceObject or EO: The Evidence Object in
  The various semantic roles to the verb d i s h o n o u r e d              focus (e.g., p o s t - m o r t e m r e p o r t , F I R , l e t t e r )
are annotated as follows:                                        The Evidence Frame captures details about the evidence
[ARGM-TMP: On August 25 , 1965] , [ARG0: the bank]               itself through the following arguments:
[V: dishonoured] [ARG1: the cheque] [ARGM-CAU: due                        • EvidenceVerb or EV: the main verb of any ac-
to insufficient balance].                                                   tion, event or fact mentioned in a sentence or re-
  We use the predicates and corresponding arguments                         vealed by the Evidence Object (e.g., k i l l e d , f o r g e d ,
obtained from the pre-trained AllenSRL model [3] to                         escaped)
instantiate our Evidence Structure for the queries and                    • Agent or A0 : someone who initiates the action
candidate sentences.                                                        indicated by the EvidenceVerb (e.g., t h e a c c u s e d ,
                                                                            Ram, ABC Pvt. Ltd.)
                                                                          • Patient or A1 : someone who undergoes the ac-
                                                                            tion indicated by the EvidenceVerb. (e.g., t h e
Table 1
Example Evidence sentences with their Evidence Structure Instances
   The bank dishonoured the cheque due to insufficient balance.
   • E F = [𝐸𝑉 = d i s h o n o u r e d , 𝐴0 = T h e b a n k , 𝐴1 = t h e c h e q u e , 𝐶𝐴𝑈 = d u e t o i n s u f f i c i e n t b a l a n c e ]
   The report revealed that organo-phosphorus compound was found in the stomach , small intestines , large intestines
   , liver , spleen , kidney and brain of the deceased .
   • O F = [𝑂𝑉 = r e v e a l e d , 𝐸𝑂 = T h e r e p o r t ]
   E F = [𝐸𝑉 = f o u n d , 𝐿𝑂𝐶 = i n t h e s t o m a c h , s m a l l i n t e s t i n e s , l a r g e i n t e s t i n e s , l i v e r , s p l e e n , k i d n e y a n d b r a i n o f
   the deceased]
   The Magistrate found prima facie evidence that the appellant had fraudulently used in the Civil Suit forged cheque
   and committed him to the Sessions for trial
   • O F = [𝑂𝑉 = f o u n d , 𝑂𝐴 = T h e M a g i s t r a t e , 𝐸𝑂 = p r i m a f a c i e e v i d e n c e ]
   E F = [𝐸𝑉 = u s e d ,𝐴0 = t h e a p p e l l a n t , 𝐴1 = f o r g e d c h e q u e , 𝐿𝑂𝐶 = i n t h e C i v i l S u i t ]
   The prosecution case was that though the rough cash book showed that on September 29, 1950 a sum of Rs. 21,133 was
   sent to the Treasury by appellant Gupta , the Treasury figures in the challan showed that on that day only a sum
   of Rs. 1,133 was deposited into the Treasury and thus a sum of Rs.20,000 was dishonestly misappropriated .
   • O F = [𝑂𝑉 = s h o w e d , 𝐸𝑂 = t h e r o u g h c a s h b o o k ]
   E F = [𝐸𝑉 = s e n t , 𝐴0 = b y a p p e l l a n t G u p t a , 𝐴1 = a s u m o f R s . 2 1 , 1 3 3 , 𝐴2 = t o t h e T r e a s u r y , 𝐴𝑅𝐺 −𝑇 𝑀𝑃 = o n S e p t e m b e r 2 9 , 1 9 5 0 ]
   • O F = [𝑂𝑉 = s h o w e d , 𝐸𝑂 = t h e T r e a s u r y f i g u r e s i n t h e c h a l l a n ]
   E F = [𝐸𝑉 = d e p o s i t e d ,𝐴0 = b y a p p e l l a n t G u p t a , 𝐴1 = o n l y a s u m o f R s . 1 , 1 3 3 , 𝐴2 = i n t o t h e T r e a s u r y , 𝑇 𝑀𝑃 = o n t h a t d a y ]
   • O F = [𝑂𝑉 = s h o w e d , 𝐸𝑂 = t h e T r e a s u r y f i g u r e s i n t h e c h a l l a n ]
   E F = [𝐸𝑉 = m i s a p p r o p r i a t e d , 𝐴1 = a s u m o f R s . 2 0 , 0 0 0 , 𝑀𝑁 𝑅 = d i s h o n e s t l y ]



             deceased, a cheque of Rs. 3,200, his wife)                                               testimony sentences (e.g., s t a t e d , s a i d ) are treated similar
         • Location or LOC: location where the action                                                 to observation verbs and represented using Observation
             took place (e.g., i n t h e b e d r o o m , a t t h e b a n k , i n                      Frames. Similarly, other action/event verbs mentioned
             Malaysia)                                                                                in witness testimony sentences are represented using
         • Time or TMP: timestamp of the action                                                       Evidence Frames. Table 2 shows examples of some Wit-
             (e.g., a b o u t 1 2 h o u r s b a c k , i n t h e m o r n i n g , o n                   ness Sentences along with the corresponding Evidence
             Monday)                                                                                  Structure Instances. The advantage of representing in-
         • Cause or CAU: cause of the action (e.g., d u e t o                                         formation about evidences and witness testimonies in
             dowry, as a result of the CBI enquiry, out of                                            the same structure is that we can make use of both these
             sheer spite)                                                                             sources of information seamlessly, for prior case retrieval.
         • Manner or MNR: manner in which the
             action took place (e.g., a s p e r t h e c h a l l a n ,
             fraudulently, wilfully)                                                                  3. Methodology
    Table 1 shows examples of some Evidence Sentences
                                                                                                      In this section, we describe our overall methodology
along with the corresponding Evidence Structure Instances.
                                                                                                      which consists of two phases. In the first phase, we
In some cases, Observation Frame may be empty due to
                                                                                                      identify Evidence and Testimony Sentences using linguis-
absence of ObservationVerb. In such cases, EvidenceOb-
                                                                                                      tic rules and weakly supervised sentence classifier. In the
ject may be present as a part of any argument in Evidence
                                                                                                      second phase, we instantiate the Evidence Structures for
Frame. E.g., t h e c h e q u e is present as 𝐴1 in the Evidence
                                                                                                      these identified sentences. For all our experiments, we
Frame of the first sentence in Table 1.
                                                                                                      use a corpus of 30,032 Indian Supreme Court judgements
    Information about named entities and their types
                                                                                                      ranging from the year 1952 to 2012.
present in various arguments of Observation or Evidence
frame is important. Hence, the Observation Frame and
Evidence Frame are also enriched by annotating enti-                                                  3.1. Identification of Evidence and
ties such as P E R S O N , O R G A N I S A T I O N , G E O - P O L I T I C A L E N T I T Y ,               Testimony Sentences
LOCATION, PRODUCT, EVENT, LANGUAGE, DATE, TIME, PERCENT,
MONEY, QUANTITY, ORDINAL, CARDINAL, WEAPON, SUBSTANCE,
                                                                                                      We identify Evidence and Testimony sentences using a
D O C U M E N T , A R T I F A C T , W O R K _ O F _ A R T , W I T N E S S , B O D Y _ P A R T , and
                                                                                                      two-step approach. In the first step, we use linguistic
V E H I C L E present in the fields.
                                                                                                      rules to obtain Evidence and Testimony sentences. In the
Witness Information Model: Information in witness                                                     second step, we use these sentences to train a sentence
testimonies can also be represented using the same Ev-                                                classifier.
idence Structure. The statement verbs used in witness
Table 2
Example Witness Testimony sentences with their Evidence Structure Instances
 He has categorically stated that by reason of enmity , A1 and A2 together have murdered his brother-in-law .
 • O F = [𝑂𝑉 = s t a t e d , 𝐴0 = H e ]
 E F = [𝐸𝑉 = m u r d e r e d , 𝐴0 = A 1 a n d A 2 t o g e t h e r , 𝐴1 = h i s b r o t h e r - i n - l a w , 𝐶𝐴𝑈 = b y r e a s o n o f e n m i t y ]
 Shri Dholey ( PW-6 ) reiterated about the dacoity and claimed that a pistol was brandished on him by one of the
 accused persons .
 • O F = [𝑂𝑉 = c l a i m e d , 𝐴0 = S h r i D h o l e y ( P W - 6 ) ]
 E F = [𝐸𝑉 = b r a n d i s h e d , 𝐴0 = b y o n e o f t h e a c c u s e d p e r s o n s , 𝐴1 = a p i s t o l , 𝐶𝐴𝑈 = o n h i m ]
 Though he stated in the post-mortem report that death would have occurred about 12 hours back , he clarified that
 there was possibility of injuries being received at about 9 A.M.
 • O F = [𝑂𝑉 = s t a t e d , 𝐴0 = h e , 𝐸𝑂 = t h e p o s t - m o r t e m r e p o r t ]
 E F = [𝐸𝑉 = o c c u r r e d , 𝐴1 = d e a t h , 𝑇 𝑀𝑃 = a b o u t 1 2 h o u r s b a c k ]
 • E F = [𝑂𝑉 = c l a r i f i e d , 𝐴0 = h e , 𝐴1 = t h a t t h e r e w a s p o s s i b i l i t y o f i n j u r i e s b e i n g r e c e i v e d a t a b o u t 9 A . M . D e c e a s e d
 Sarit Khanna was aged about 27 years]
 He admitted , however , that Shri Buch had met him in connection with the covenant , but he denied that he had
 received any letter Exhibit P-9 from Shri Buch or the lists Exhibits P- 10 to P- 12 regarding his private and
 State properties , were a part thereof .
 • O F = [𝑂𝑉 = a d m i t t e d , 𝐴0 = H e ]
 E F = [𝐸𝑉 = m e t , 𝐴0 = S h r i B u c h , 𝐴1 = h i m , 𝑇 𝑀𝑃 = i n c o n n e c t i o n w i t h t h e c o v e n a n t ]
 • O F = [𝑂𝑉 = d e n i e d , 𝐴0 = H e ]
 E F = [𝐸𝑉 = r e c e i v e d , 𝐴0 = h e , 𝐴1 = a n y l e t t e r E x h i b i t P - 9 , 𝐴2 = f r o m S h r i B u c h ]



Step I: Linguistic Rules based Approach: As there                                          • The classifier has two outputs - i) first output predicts
are no publicly annotated datasets for identification of                                   a binary label indicating whether the sentence contains
Evidence and Testimony sentences, we rely on linguistic                                    Evidence or not and ii) second output predicts a binary
rules to identify these sentences with high precision                                      label indicating whether the sentence contains Testimony
as our first step. The linguistic rules for identifying                                    or not.
Evidence sentences are described in detail in Table 3.                                     • 1824 sentences are labelled as Evidence and Testimony
These rules identified 62,310 sentences as Evidences                                       both. These sentences are identified as Evidence as well
from our corpus. As there is no annotated dataset,                                         as Testimony by both the sets of linguistic rules.
in order to estimate the precision of the linguistic                                       • 60486 sentences are labelled as Evidence and non-
rules we use random sampling strategy. We selected                                         Testimony. These sentences are identified as Evidence
a set 100 random sentences identified as Evidence by                                       by the rules but not as Testimony.
the linguistic rule, and got them verified by a human                                      • 34649 sentences are labelled as non-Evidence and Testi-
expert. The precision turned out to be 85%. Similarly,                                     mony. These sentences are identified as Testimony by the
we use the linguistic rules proposed in Ghosh et al. [1]                                   rules but not as Evidence.
for identifying Testimony and non-Testimony sentences                                      • 14234 sentences are labelled as non-Evidence and
where the reported precision is around 85%. These                                          non-Testimony. These sentences are identified as non-
rules identified 36,473 sentence as Testimony and 14,234                                   Testimony by the rules and not identified as Evidence.
sentences as non-Testimony from the same corpus.                                              After this classifier is trained, we use it to classify all
                                                                                           the remaining sentences in the corpus. These sentences
Step II: Weakly Supervised Sentence Classification:                                        are neither identified Evidence by the Evidence rules nor
We observed that although the linguistic rules identify                                    as Testimony/non-Testimony by the Testimony rules. Us-
Evidence and Testimony sentences with high precision,                                      ing the prediction confidence, we selected top 10,000
they may miss to identify some sentences which should                                      sentences classified as Evidence and top 5,000 sentences
have been identified as Evidence or Testimony (see exam-                                   classified as Testimony. Table 4 shows some examples
ples in Table 4). Hence, we train a supervised sentence                                    of sentences identified as Evidence by the classifier but
classifier to improve overall recall of identification of Ev-                              not by the linguistic rules. To estimate the precision,
idence and Testimony sentences. The classifier used is a                                   we again employed the random sampling strategy. We
BiLSTM-based [4] multi-label sentence classifier whose                                     selected 100 random sentences each from these high con-
architecture is depicted in Figure 1. This classifier is                                   fidence Evidence and Testimony sentences and a human
weakly supervised since its training data is automatically                                 expert verified them. The precision of 72% is observed for
created using the sentences identified by the linguistic                                   Evidence sentences and 68% for Testimony sentences. The
rules as follows:                                                                          precision of the sentence classifier is lower as compared
Table 3
Linguistic Rules for identifying Evidence Sentences
  Any sentence 𝑆 should satisfy the following conditions in order to be identified as an Evidence Sentence:
   E-R1 𝑆 should contain at least one Evidence Object as defined in Section 2.2. The list of words corresponding to evidence
         objects is created automatically by using WordNet hypernym structure. We create a list of all words for which the
         following WordNet synsets are ancestors in hypernym tree – a r t i f a c t (e.g., g u n , c l o t h e s ), d o c u m e n t (e.g. r e p o r t , l e t t e r ),
         s u b s t a n c e (e.g. k e r o s e n e , b l o o d ). This list is looked up to identify evidence objects in a sentence.
   E-R2 𝑆 should contain at least one action verb from a pre-defined set of verbs like t a m p e r , k i l l , s u s t a i n , f o r g e OR 𝑆 should
         contain at least one observation verb from a pre-defined set of verbs like r e p o r t , s h o w , f i n d . Both the pre-defined sets
         of verbs are prepared by observing multiple example sentences containing evidence objects.
   E-R3 In the dependency tree of 𝑆, the evidence object (identified by E-R1) should occur within the subtree rooted at the
         action or observation verb (identified by E-R2) AND there should not be any other verb (except auxiliary verbs like
         h a s , b e e n , w a s , w e r e , i s ) occurring between the two. This ensures that the evidence object always lies within the verb
         phrase headed by the action or observation verb.



Table 4
Example of Evidence Sentences Identified by the Classifier but not by the linguistic rules
 S1: Raju PW2 took Preeti into the bath room at the instance of Accused No.1 who cut a length of wire of washing
 machine and used it to choke her to death, who however, survived.
 S2: Raju PW2 took Satyabhamabai Sutar in the kitchen where the accused No.1 had already reached and was washing
 the blood stained knife.
 S3: Hemlata was also killed by inflicting knife injuries.
 S4: Accused No.2 and Raju PW2 took the child into the room where Meerabai was lying dead in the pool of blood.
 S5: Accused No.2 gave her blows by putting his knees on her stomach and when she was immobilised this way , the
 Accused No.1 gave her knife blows on her neck with the result she also died.
 S6: Almirahs found in the flat were emptied to the extent the accused could put articles and other cash and
 valuables in the air-bag obtained from the said flat.
 S7: Blood stained clothes of Accused No.2 were put in the air-bag along with stolen articles.



                                                                                   3.2. Evidence Structure Instances
                                                                In this phase, we discuss the technique of instantiating
                                                                Evidence Structures for sentences identified as Evidence
                                                                or Testimony in the previous phase. We used Semantic
                                                                Role Labelling [3] to identify and fill the arguments of
                                                                the Observation Frame and the Evidence Frame in the
                                                                Evidence Structure Instance for every candidate sentence.
                                                                This is demonstrated in Algorithm 1. We identify Obser-
                                                                vation Frames using Observation Cue Verbs. For each of
                                                                these Observation Frames we identify the corresponding
                                                                Evidence Objects and Evidence Frames. For identifying Ev-
                                                                idence Objects, we first use Named Entity Recognition [5]
Figure 1: Architecture of the BiLSTM-based multi-label sen- and WordNet based Entity Identification [6] to identify
tence classifier (T: Testimony, NT: Non-Testimony, E: Evidence, the named entities in the sentence and annotate them in
NE: Non-Evidence)                                               the Frames extracted. The Evidence Objects in a phrase are
                                                                then obtained by selecting named entities annotated as
                                                                one of the following types - A R T I F A C T , V E H I C L E , W E A P O N ,
to the rules because, it is applied on a more difficult set D O C U M E N T , W O R K _ O F _ A R T , S U B S T A N C E . This corresponds to
of sentences for which the linguistic rules fail to identify the 𝑔𝑒𝑡_𝑒𝑣𝑖𝑑𝑒𝑛𝑐𝑒_𝑜𝑏𝑗𝑒𝑐𝑡 function used in Algorithm 1. Ob-
any label. At the end of this two-step process (linguistic servation Frames that do not contain a corresponding
rules followed by the sentence classifier), we have 112,401 Evidence Frame are redesigned as stand alone Evidence
sentences identified either as Evidence or as Testimony.        Frames. We finally combine the Evidence Frame and the
                                                                Observation Frame into an Evidence Structure Instance.
                                                                   We measured the accuracy of 260 Evidence Structure
Instances obtained from 100 random Evidence and Testi-       use Context Dependent Evidence Detection to find evidence
mony sentences. The accuracy of the Observation Frame        information present in a sentence on a phrase level. As
extraction is 86% and that of Evidence Frame extraction is   compared to this, we identify both Evidence and Testi-
88%. We observed that most of the incorrect extractions      mony sentences, represent them in a rich structure and
were due to parsing error in the SRL model.                  also use that for prior case retrieval. This is a challenging
                                                             task due to the inherently complex nature of legal texts
                                                             and the finer granularity of matching involved.
4. Prior Case Retrieval                                         Ji et al. [12] propose an Evidence Information Ex-
                                                             traction system which captures evidence production
In order to demonstrate effectiveness of the proposed
                                                             paragraph, evidence cross-examination paragraph, evi-
Evidence Structure, we apply it for the task of prior case
                                                             dence provider, evidence name, evidence content, cross-
retrieval. This task is to create a relevance-based ranked
                                                             examination party and cross-examination opinion relat-
list of court judgements (documents) in our corpus for
                                                             ing to an evidence presented in the court. While this
a query. In order to retrieve prior cases for a query,
                                                             technique may suit well for Chinese court records that
we represent the query using an Evidence Structure In-
                                                             follow a relatively structured representation, it does not
stance (𝐸𝑣𝑆𝑡𝑟𝑢𝑐𝑡𝑄 ). We then compute the similarity of
                                                             suit well to the Indian Court Records that contain de-
query instance 𝐸𝑣𝑆𝑡𝑟𝑢𝑐𝑡𝑄 against each document instance
                                                             scriptive and varied formats of the court proceedings.
𝐸𝑣𝑆𝑡𝑟𝑢𝑐𝑡𝐷 obtained from every Evidence or Testimony sen-
                                                                Gomes and Ladeira [13] and Landthaler et al. [14] per-
tence in the corpus. Algorithm 2 shows the steps for com-
                                                             forms full text search for legal document collection by
puting similarity. We refer to this algorithm as 𝑆𝑒𝑚𝑀𝑎𝑡𝑐ℎ
                                                             obtaining word2vec word embeddings and then taking
because of its semantic matching ability. We use cosine
                                                             their average for computing similarity. However, com-
similarity between the phrase embeddings of correspond-
                                                             puting the average of the embeddings gives a lossy rep-
ing arguments of the Evidence Structure Instances to
                                                             resentation where relative order of the words is lost. In
compute similarity. For obtaining phrase embedding
                                                             contrast, we represent the sentences using the Evidence
for any phrase (referred as 𝑃ℎ𝑟𝑎𝑠𝑒𝑉 𝑒𝑐 in Algorithm 2),
                                                             Structure Instances, where the structure itself takes care
we consider the average of GloVe word embeddings [7]
                                                             of the relative ordering. Gomes and Ladeira [13] demon-
of the words in that phrase excluding stop words. We
                                                             strate BM25 and TF-IDF for Prior Case retrieval. In our
compute the similarity scores within corresponding argu-
                                                             results section, we demonstrate the comparative poor
ments of both the frames. These scores across different
                                                             performance of BM25 and TF-IDF in handling corner
arguments are combined to get a final similarity score
                                                             cases.
between 𝐸𝑣𝑆𝑡𝑟𝑢𝑐𝑡𝑄 and 𝐸𝑣𝑆𝑡𝑟𝑢𝑐𝑡𝐷 . We multiply the final
similarity score by a Sentence BERT [8] based similarity
score between the query and the sentence containing          6. Experimental Evaluation
𝐸𝑣𝑆𝑡𝑟𝑢𝑐𝑡𝐷 . This is necessary because errors in the auto-
mated SRL tool may lead to imperfect Evidence Structure      In this section, we discuss our experiments including
instances in some cases. A sentence similarity score         the dataset, baseline techniques, evaluation metrics and
which is not dependent on any such structure within the      analysis of results.
sentences provides a complementary view of capturing
sentence similarity. Finally, the overall relevance score    6.1. Dataset
of the query with a document is the maximum score cor-
responding to any Evidence Structure Instance 𝐸𝑣𝑆𝑡𝑟𝑢𝑐𝑡𝐷      We use the Indian Supreme Court judgements from years
obtained from the document. Table 5 shows a running          1952 to 2012 freely available at http://liiofindia.org/in/
example of how a similarity score is computed between        cases/cen/INSC/. There are 30032 court (documents)
an Evidence Structure Instance (𝐸𝑣𝑆𝑡𝑟𝑢𝑐𝑡𝑄 ) from a query     containing 4,111,091 sentences where average sentence
and an Evidence Structure Instance (𝐸𝑣𝑆𝑡𝑟𝑢𝑐𝑡𝐷 ) from a       length is 31 words and standard deviation of 24.
document in the corpus.
                                                             6.2. Baselines
5. Related Work                                              For the task of prior case retrieval, we implement two
                                                             baseline techniques:
While the task of evidence extraction from legal docu- • BM25: It is a popular TF-IDF based relevance computa-
ments is related to several information retrieval and NLP tion technique. We use the BM25+ variant1 as described
tasks, there are no established baselines for the task. Bel- in Trotman et al. [15]. This technique uses a bag-of-
lot et al. [9] and Cartright et al. [10] have worked on words approach that ignores the sentence structure. We
Evidence Retrieval that identifies whole documents that
contain an evidence. On the other hand, Rinott et al. [11]       1
                                                                   https://pypi.org/project/rank-bm25/
 input         : 𝑠 (sentence), 𝑆𝑅𝐿_𝑃 (set of semantic frames in 𝑠 as per any semantic role labeller, each frame 𝑃 consists of a
                 predicate 𝑃.𝑉 and corresponding arguments 𝑃.𝐴𝑅𝐺0 , 𝑃.𝐴𝑅𝐺1 , 𝑃.𝐴𝑅𝐺2 , 𝑃.𝐴𝑅𝐺𝑀-𝐿𝑂𝐶, etc.)
 output        : 𝐸𝑣𝑆𝑡𝑟𝑢𝑐𝑡𝑠 = Evidence Structure Instances of the input sentence consisting of 𝑂𝑏𝑠𝑒𝑟𝑣𝑎𝑡𝑖𝑜𝑛𝐹 𝑟𝑎𝑚𝑒 (𝑂𝐹) and
                 𝐸𝑣𝑖𝑑𝑒𝑛𝑐𝑒𝐹 𝑟𝑎𝑚𝑒 (𝐸𝐹)
 parameter : 𝑂𝐵𝑆_𝑉 𝐸𝑅𝐵𝑆 = {a c c e p t , a d d , a d m i t , a g r e e , a l l e g e , a l l o w , a l t e r , a p p r i s e , a s s e r t , b r i e f , b u i l d , c h a l l e n g e ,
                 claim, clarify, complain, confirm, corroborate, decline, demand, deny, depose, describe, disclose, dismiss,
                 examine, exhibit, find, include, indicate, inform, mention, note, notice, observe, obtain, occur, point,
                 prepare, present, receive, recover, refuse, reject, remember, report, reveal, say, show, state, submit,
                 s u g g e s t , t e l l , w i t h d r a w }, 𝑁 𝐸𝐺_𝑊 𝑂𝑅𝐷𝑆 = {n o , n o t , n e i t h e r , n o r , n e v e r }
 𝐸𝑣𝑆𝑡𝑟𝑢𝑐𝑡𝑠 ∶= ∅
 𝑂𝐹 𝑠 ∶= ∅
 // Obtain Observation Frames in the sentence 𝑠
 foreach 𝑃 ∈ 𝑆𝑅𝐿_𝑃 such that 𝑃.𝑉 ∈ 𝑂𝐵𝑆_𝑉 𝐸𝑅𝐵𝑆 do
     𝑂𝐹 ∶= Create empty Observation Frame
     𝑂𝐹 .𝑉 ∶= 𝑃.𝑉
     𝑂𝐹 .𝑁 𝐸𝐺 ∶= 𝑃.𝐴𝑅𝐺𝑀-𝑁 𝐸𝐺
     𝑂𝐹 .𝐴0 ∶= 𝑃.𝐴𝑅𝐺0
     𝑂𝐹 .𝐴1 ∶= 𝑃.𝐴𝑅𝐺1
     // If any of the arguments of the predicate starts with a negative word, then we negate the verb.
     if 𝑂𝐹 .𝐴0 or 𝑂𝐹 .𝐴1 starts with any word from 𝑁 𝐸𝐺_𝑊 𝑂𝑅𝐷𝑆 then
           𝑂𝐹 .𝑁 𝐸𝐺 ∶= 𝑇 𝑟𝑢𝑒
     𝑂𝐹 .𝐸𝑂 ∶= 𝑔𝑒𝑡_𝑒𝑣𝑖𝑑𝑒𝑛𝑐𝑒_𝑜𝑏𝑗𝑒𝑐𝑡(𝑃.𝐴𝑅𝐺0 ) ∪ 𝑔𝑒𝑡_𝑒𝑣𝑖𝑑𝑒𝑛𝑐𝑒_𝑜𝑏𝑗𝑒𝑐𝑡(𝑃.𝐴𝑅𝐺𝑀-𝐿𝑂𝐶)
     𝑂𝐹 𝑠 ∶= 𝑂𝐹 𝑠 ∪ {𝑂𝐹 }

 // Obtain corresponding Evidence Frames for every Observation Frame
 foreach 𝑂𝐹 ∈ 𝑂𝐹 𝑠 do
     𝐹 𝑜𝑢𝑛𝑑𝐸𝐹 ∶= 𝐹 𝑎𝑙𝑠𝑒
     foreach 𝑃 ∈ 𝑆𝑅𝐿_𝑃 such that 𝑃.𝑉 occurs within the span of 𝑂𝐹 .𝐴1 do
         if 𝑃.𝑉 is a copula verb and any of 𝑃.𝐴𝑅𝐺0 or 𝑃.𝐴𝑅𝐺1 does not exist then
              continue
             𝐸𝐹 ∶= Create empty Evidence Frame
             𝐸𝐹 .𝑉 ∶= 𝑃.𝑉
             𝐸𝐹 .𝑁 𝐸𝐺 ∶= 𝑃.𝐴𝑅𝐺𝑀-𝑁 𝐸𝐺
             // If any of the arguments of the predicate starts with a negative word, then we negate the verb
             if 𝑂𝐹 .𝐴0 or 𝑂𝐹 .𝐴1 starts with any word from 𝑁 𝐸𝐺_𝑊 𝑂𝑅𝐷𝑆 then
                   𝐸𝐹 .𝑁 𝐸𝐺 ∶= 𝑇 𝑟𝑢𝑒
              foreach argument 𝐴𝑅𝐺 ∈ 𝑃.𝑎𝑟𝑔𝑢𝑚𝑒𝑛𝑡𝑠 do
                  𝐸𝐹 .𝐴𝑅𝐺 ∶= 𝑃.𝐴𝑅𝐺
              delete(𝑂𝐹 .𝐴1 )
              𝐸𝑣𝑆𝑡𝑟𝑢𝑐𝑡 ∶= {(𝑂𝐹 , 𝐸𝐹 )}
              𝐸𝑣𝑆𝑡𝑟𝑢𝑐𝑡𝑠 ∶= 𝐸𝑣𝑆𝑡𝑟𝑢𝑐𝑡𝑠 ∪ 𝐸𝑣𝑆𝑡𝑟𝑢𝑐𝑡
              𝐹 𝑜𝑢𝑛𝑑𝐸𝐹 ∶= 𝑇 𝑟𝑢𝑒

       // If no Evidence Frame exists for an Observation Frame, transfer the Observation Frame to the Evidence
          Frame
       if FoundEF == False then
            𝐸𝐹 ∶= Create empty Evidence Frame
            𝐸𝐹 .𝑉 ∶= 𝑂𝐹 .𝑉
            𝑃 ∶= 𝑃 ′ ∈ 𝑆𝑅𝐿_𝑃 such that 𝑃 ′ .𝑉 = 𝑂𝐹 .𝑉
            foreach argument 𝐴𝑅𝐺 ∈ 𝑃.𝑎𝑟𝑔𝑢𝑚𝑒𝑛𝑡𝑠 do
                 𝐸𝐹 .𝐴𝑅𝐺 ∶= 𝑃.𝐴𝑅𝐺
             clear(𝑂𝐹)
             𝑂𝐹 .𝐸𝑂 ∶= 𝑔𝑒𝑡_𝑒𝑣𝑖𝑑𝑒𝑛𝑐𝑒_𝑜𝑏𝑗𝑒𝑐𝑡(𝑃.𝐴𝑅𝐺0 ) ∪ 𝑔𝑒𝑡_𝑒𝑣𝑖𝑑𝑒𝑛𝑐𝑒_𝑜𝑏𝑗𝑒𝑐𝑡(𝑃.𝐴𝑅𝐺𝑀-𝐿𝑂𝐶)
             //Add all the required arguments to Evidence Frame
             𝐸𝑣𝑆𝑡𝑟𝑢𝑐𝑡 ∶= {(𝑂𝐹 , 𝐸𝐹 )}
             𝐸𝑣𝑆𝑡𝑟𝑢𝑐𝑡𝑠 ∶= 𝐸𝑣𝑆𝑡𝑟𝑢𝑐𝑡𝑠 ∪ 𝐸𝑣𝑆𝑡𝑟𝑢𝑐𝑡
 return(𝐸𝑣𝑆𝑡𝑟𝑢𝑐𝑡𝑠)
Algorithm 1: 𝑔𝑒𝑡_𝑒𝑣𝑖𝑑𝑒𝑛𝑐𝑒_𝑠𝑡𝑟𝑢𝑐𝑡𝑢𝑟𝑒_𝑖𝑛𝑠𝑡𝑎𝑛𝑐𝑒𝑠: Algorithm for instantiating Evidence Structure for a sentence
  input : 𝐸𝑣𝑆𝑡𝑟𝑢𝑐𝑡𝑄 : Evidence Structure Instance from a query sentence 𝑄
           𝐸𝑣𝑆𝑡𝑟𝑢𝑐𝑡𝐷 : Evidence Structure Instance from a sentence 𝐷 in the corpus
  output : Similarity score between 𝐸𝑣𝑆𝑡𝑟𝑢𝑐𝑡𝑄 and 𝐸𝑣𝑆𝑡𝑟𝑢𝑐𝑡𝐷
  // Checking for negation
  if 𝐸𝑣𝑆𝑡𝑟𝑢𝑐𝑡𝑄 .𝑂𝐹 .𝑁 𝐸𝐺 ≠ 𝐸𝑣𝑆𝑡𝑟𝑢𝑐𝑡𝐷 .𝑂𝐹 .𝑁 𝐸𝐺 then return 0

  if 𝐸𝑣𝑆𝑡𝑟𝑢𝑐𝑡𝑄 .𝐸𝐹 .𝑁 𝐸𝐺 ≠ 𝐸𝑣𝑆𝑡𝑟𝑢𝑐𝑡𝐷 .𝐸𝐹 .𝑁 𝐸𝐺 then return 0


  // Computing similarity between main predicates, using cosine similarity of their word embeddings
  𝑠𝑖𝑚𝐸 ∶= 𝐶𝑜𝑠𝑖𝑛𝑒𝑆𝑖𝑚(𝑊 𝑜𝑟𝑑𝑉 𝑒𝑐(𝐸𝑣𝑆𝑡𝑟𝑢𝑐𝑡𝑄 .𝐸𝐹 .𝑉 ), 𝑊 𝑜𝑟𝑑𝑉 𝑒𝑐(𝐸𝑣𝑆𝑡𝑟𝑢𝑐𝑡𝐷 .𝐸𝐹 .𝑉 ))

  // Computing similarity between corresponding Evidence Objects, using cosine similarity of their phrase
       embeddings
  𝑠𝑖𝑚𝐸𝑂 ∶= 𝐶𝑜𝑠𝑖𝑛𝑒𝑆𝑖𝑚(𝑃ℎ𝑟𝑎𝑠𝑒𝑉 𝑒𝑐(𝐸𝑣𝑆𝑡𝑟𝑢𝑐𝑡𝑄 .𝑂𝐹 .𝐸𝑂), 𝑃ℎ𝑟𝑎𝑠𝑒𝑉 𝑒𝑐(𝐸𝑣𝑆𝑡𝑟𝑢𝑐𝑡𝐷 .𝑂𝐹 .𝐸𝑂))

  // Computing similarity between other arguments, using cosine similarity of their phrase embeddings
  𝑛𝑢𝑚𝑎𝑟𝑔𝑠 ∶= 0
  𝑠𝑖𝑚𝑎𝑟𝑔𝑠 ∶= 0
  foreach 𝑎𝑟𝑔 ∈ (𝐸𝑣𝑆𝑡𝑟𝑢𝑐𝑡𝑄 .𝐸𝐹 .𝑎𝑟𝑔𝑢𝑚𝑒𝑛𝑡𝑠 − {𝑉 }) do
      if 𝐸𝑣𝑆𝑡𝑟𝑢𝑐𝑡𝑄 .𝐸𝐹 .𝑎𝑟𝑔 exists then
          𝑠𝑖𝑚𝑎𝑟𝑔𝑠 ∶= 𝑠𝑖𝑚𝑎𝑟𝑔𝑠 + 𝐶𝑜𝑠𝑖𝑛𝑒𝑆𝑖𝑚(𝑃ℎ𝑟𝑎𝑠𝑒𝑉 𝑒𝑐(𝐸𝑣𝑆𝑡𝑟𝑢𝑐𝑡𝑄 .𝐸𝐹 .𝑎𝑟𝑔), 𝑃ℎ𝑟𝑎𝑠𝑒𝑉 𝑒𝑐(𝐸𝑣𝑆𝑡𝑟𝑢𝑐𝑡𝐷 .𝐸𝐹 .𝑎𝑟𝑔))
          𝑛𝑢𝑚𝑎𝑟𝑔𝑠 ∶= 𝑛𝑢𝑚𝑎𝑟𝑔𝑠 + 1

  𝑠𝑖𝑚𝑎𝑟𝑔𝑠 ∶= 𝑠𝑖𝑚𝑎𝑟𝑔𝑠 /𝑛𝑢𝑚𝑎𝑟𝑔𝑠

  // Computing overall similarity
  𝑠𝑖𝑚𝑓 𝑖𝑛𝑎𝑙 ∶= 𝑠𝑖𝑚𝐸 × 𝑠𝑖𝑚𝑎𝑟𝑔𝑠 × 𝑠𝑖𝑚𝐸𝑂

  // The overall similarity is multiplied by the Sentence-BERT based sentence similarity between Q and D
  𝑠𝑖𝑚𝑓 𝑖𝑛𝑎𝑙 ∶= 𝑠𝑖𝑚𝑓 𝑖𝑛𝑎𝑙 × 𝐶𝑜𝑠𝑖𝑛𝑒𝑆𝑖𝑚(𝑆𝑒𝑛𝑡𝑉 𝑒𝑐(𝑄), 𝑆𝑒𝑛𝑡𝑉 𝑒𝑐(𝐷))

  return 𝑠𝑖𝑚𝑓 𝑖𝑛𝑎𝑙
      Algorithm 2: 𝑆𝑒𝑚𝑀𝑎𝑡𝑐ℎ: Algorithm for computing similarity between 𝐸𝑣𝑆𝑡𝑟𝑢𝑐𝑡𝑄 and 𝐸𝑣𝑆𝑡𝑟𝑢𝑐𝑡𝐷



Table 5
Example of the proposed 𝑆𝑒𝑚𝑀𝑎𝑡𝑐ℎ algorithm in action
 Query: T h e a u t o p s y r e p o r t r e v e a l s t h a t s o m e p o i s o n o u s c o m p o u n d s a r e f o u n d i n t h e s t o m a c h o f t h e d e c e a s e d .
 𝐸𝑣𝑆𝑡𝑟𝑢𝑐𝑡𝑄 ∶ O F = [𝑂𝑉 = r e v e a l s , 𝐸𝑂 = T h e a u t o p s y r e p o r t ]; E F = [𝐸𝑉 = f o u n d , 𝐴1 = s o m e p o i s o n o u s c o m p o u n d s , 𝐿𝑂𝐶 = i n t h e
 stomach of the deceased]
 Sentence:                The report of the Chemical Examiner showed that a heavy concentration of arsenic was found in the
 viscera.
 𝐸𝑣𝑆𝑡𝑟𝑢𝑐𝑡𝐷 ∶ O F = [𝑂𝑉 = s h o w e d , 𝐸𝑂 = T h e r e p o r t o f t h e C h e m i c a l E x a m i n e r ]; E F = [𝐸𝑉 = f o u n d , 𝐴1 = a h e a v y c o n c e n t r a t i o n o f
 a r s e n i c , 𝐿𝑂𝐶 = i n t h e v i s c e r a ]
 • Similarity between main predicates, their arguments and evidence objects
 𝑠𝑖𝑚𝐸 ∶= 𝐶𝑜𝑠𝑖𝑛𝑒𝑆𝑖𝑚(𝑊 𝑜𝑟𝑑𝑉 𝑒𝑐(f o u n d ), 𝑊 𝑜𝑟𝑑𝑉 𝑒𝑐(f o u n d )) = 1.0
 𝑠𝑖𝑚𝐴1 ∶= 𝐶𝑜𝑠𝑖𝑛𝑒𝑆𝑖𝑚(𝑃ℎ𝑟𝑎𝑠𝑒𝑉 𝑒𝑐(s o m e p o i s o n o u s c o m p o u n d s ), 𝑃ℎ𝑟𝑎𝑠𝑒𝑉 𝑒𝑐(a h e a v y c o n c e n t r a t i o n o f a r s e n i c )) = 0.5469
 𝑠𝑖𝑚𝐿𝑂𝐶 ∶= 𝐶𝑜𝑠𝑖𝑛𝑒𝑆𝑖𝑚(𝑃ℎ𝑟𝑎𝑠𝑒𝑉 𝑒𝑐(i n t h e s t o m a c h o f t h e d e c e a s e d ), 𝑃ℎ𝑟𝑎𝑠𝑒𝑉 𝑒𝑐(i n t h e v i s c e r a )) = 0.3173
 𝑠𝑖𝑚𝑎𝑟𝑔𝑠 ∶= (𝑠𝑖𝑚𝐴1 + 𝑠𝑖𝑚𝐿𝑂𝐶 )/2.0 = 0.4321
 𝑠𝑖𝑚𝐸𝑂 ∶= 𝐶𝑜𝑠𝑖𝑛𝑒𝑆𝑖𝑚(𝑃ℎ𝑟𝑎𝑠𝑒𝑉 𝑒𝑐(T h e a u t o p s y r e p o r t ), 𝑃ℎ𝑟𝑎𝑠𝑒𝑉 𝑒𝑐(T h e r e p o r t o f t h e C h e m i c a l E x a m i n e r )) = 0.8641
 • Final similarity
 𝑠𝑖𝑚𝑓 𝑖𝑛𝑎𝑙 ∶= 𝑠𝑖𝑚𝐸 × 𝑠𝑖𝑚𝑎𝑟𝑔𝑠 × 𝑠𝑖𝑚𝐸𝑂 × 𝑠𝑖𝑚𝑆𝐵𝐸𝑅𝑇 = 1.0 × 0.4321 × 0.8641 × 0.607 = 0.2266 (Ranked within top 10 relevant documents)
Table 6
Evaluation of various techniques for the task of prior case retrieval. All entries are of the form (R-Prec; Avg. Precision). (Note:
Our proposed approach 𝑆𝑒𝑚𝑀𝑎𝑡𝑐ℎ is referred as 𝑆𝑀. Underlines indicate the best performing results for each query across
multiple techniques)
  What are the cases where⋯
  𝑄1 : b l o o d s t a i n s w e r e f o u n d o n c l o t h e s o f t h e d e c e a s e d .
  𝑄2 : t h e d e c e a s e d h a d a t t a c k e d s o m e p e r s o n w i t h s t i c k s .
  𝑄3 : t h e p o l i c e h a s m u r d e r e d t h e d e c e a s e d .
  𝑄4 : s o m e e v i d e n c e s h o w s t h a t t h e e x h i b i t e d g u n w a s n o t u s e d .
  𝑄5 : t h e a u t o p s y r e p o r t r e v e a l s t h a t s o m e p o i s o n o u s c o m p o u n d s a r e f o u n d i n t h e s t o m a c h o f t h e d e c e a s e d .
  𝑄6 : t h e d e c e a s e d i s a t t a c k e d w i t h a k n i f e .
  𝑄7 : a l e t t e r b y t h e d e c e a s e d r e v e a l t h a t d o w r y w a s d e m a n d e d .
  𝑄8 : a c h e q u e w a s d i s h o n o u r e d d u e t o i n s u f f i c i e n t f u n d s .
  𝑄9 : b r i b e w a s d e m a n d e d b y p o l i c e .
  𝑄10 : a s i g n a t u r e w a s f o r g e d o n a n a f f i d a v i t .

  Query         𝐵𝑀25𝑎𝑙𝑙          𝐵𝑀25𝑇           𝐵𝑀25𝐸           𝐵𝑀25𝑇 𝐸             𝑆𝐵𝑇                𝑆𝐵𝐸          𝑆𝐵𝑇 𝐸             𝑆𝑀𝑇              𝑆𝑀𝐸             𝑆𝑀𝑇 𝐸
    𝑄1         0.24; 0.26       0.06; 0.02      0.59; 0.49       0.59; 0.52       0.00; 0.01         0.24; 0.15    0.18; 0.14       0.00; 0.01       0.24; 0.16       0.24; 0.14
    𝑄2         0.25; 0.43       0.00; 0.05      0.00; 0.04       0.00; 0.06       0.00; 0.01         0.00; 0.00    0.00; 0.00       0.25; 0.14       0.25; 0.25       0.50; 0.30
    𝑄3         0.00; 0.01       0.00; 0.03      0.33; 0.33       0.33; 0.35       0.33; 0.12         0.00; 0.00    0.00; 0.09       0.33; 0.12       0.00; 0.00       0.33; 0.12
    𝑄4         0.17; 0.06       0.00; 0.01      0.00; 0.02       0.00; 0.04       0.00; 0.01         0.42; 0.25    0.42; 0.22       0.08; 0.04       0.25; 0.27       0.33; 0.29
    𝑄5         0.30; 0.43       0.10; 0.05      0.40; 0.35       0.40; 0.37       0.20; 0.15         0.70; 0.80    0.70; 0.80       0.00; 0.02       0.40; 0.40       0.40; 0.40
    𝑄6         0.31; 0.42       0.33; 0.28      0.38; 0.35       0.46; 0.52       0.23; 0.14         0.33; 0.38    0.36; 0.40       0.20; 0.18       0.28; 0.27       0.41; 0.42
    𝑄7         0.25; 0.35       0.00; 0.08      0.50; 0.54       0.50; 0.33       0.00; 0.04         0.00; 0.12    0.00; 0.09       0.25; 0.06       0.00; 0.00       0.25; 0.06
    𝑄8         0.48; 0.46       0.01; 0.09      0.67; 0.71       0.71; 0.73       0.05; 0.02         0.62; 0.67    0.62; 0.67       0.00; 0.00       0.57; 0.63       0.57; 0.64
    𝑄9         0.20; 0.23       0.20; 0.17      0.20; 0.21       0.40; 0.31       0.40; 0.39         0.20; 0.21    0.50; 0.51       0.40; 0.41       0.10; 0.12       0.50; 0.48
   𝑄10         0.50; 0.52       0.00; 0.11      0.25; 0.16       0.25; 0.21       0.00; 0.01         0.00; 0.04    0.00; 0.03       0.25; 0.13       0.50; 0.61       0.50; 0.61
   Avg         0.27; 0.32       0.08; 0.09      0.33; 0.32       0.36; 0.34       0.12; 0.09         0.25; 0.26    0.28; 0.30       0.18; 0.11       0.26; 0.27       0.40; 0.35



use 4 settings considering different sentences in each                                   6.3. Evaluation
document:
                                                                                         All the baseline techniques and our proposed technique
• 𝐵𝑀25𝑎𝑙𝑙 : All sentences
                                                                                         are evaluated using a set of queries and using certain
• 𝐵𝑀25𝑇 𝐸 : Only Testimony or Evidence sentences
                                                                                         evaluation metrics to evaluate and compare the ranked
• 𝐵𝑀25𝑇 : Only Testimony sentences
                                                                                         lists produced by each of these techniques.
• 𝐵𝑀25𝐸 : Only Evidence sentences
                                                                                         Queries: We chose 10 queries (shown in Table 6) which
• Sentence-BERT [8]: This technique is based on
                                                                                         represent cases and evidence objects of diverse nature
Siamese-BERT networks to obtain more meaningful sen-
                                                                                         (domestic violence, financial fraud etc.).
tence embeddings as compared to vanilla BERT [16].
                                                                                         Ground Truth: We created a set of gold-standard rele-
We used the pre-trained model b e r t - b a s e - n l i - s t s b - m e a n -
                                                                                         vant documents for each query using the standard pooling
t o k e n s to obtain sentence embeddings for sentences. Fol-
                                                                                         technique [17]. We ran the following techniques to pro-
lowing Ghosh et al. [1], we use the pre-trained model
                                                                                         duce a ranked list of documents for each query – 𝐵𝑀25𝑎𝑙𝑙 ,
as it is and did not fine-tune it further. This is because
                                                                                         𝐵𝑀25𝑇 𝐸 , 𝑆𝐵𝑇 𝐸 , and our proposed technique 𝑆𝑒𝑚𝑀𝑎𝑡𝑐ℎ𝑇 𝐸 .
such fine-tuning needs annotated sentence pairs with
                                                                                         We chose top 10 documents from the ranked list pro-
labels indicating whether the sentences in the pair are
                                                                                         duced by each technique. Human experts verified the
semantically similar or not. Such annotated dataset is
                                                                                         relevance of each document for the query. Finally, after
expensive to create and our aim is to avoid any depen-
                                                                                         discarding all the irrelevant documents, we got a set of
dence on manually annotated training data. Similar to
                                                                                         gold-standard relevant documents for each query2 .
Ghosh et al. [1], we used sentence embeddings obtained
                                                                                         Metrics: We used R-Precision and Average Precision as
by Sentence-BERT to compute cosine similarity between
                                                                                         our evaluation metrics [17].
a query sentence and a candidate sentence in a docu-
                                                                                              1. R-Precision (R-Prec): This calculates the the
ment. The overall similarity of a document with a query
                                                                                                  number of relevant documents observed at 𝑅.
is the maximum cosine similarity obtained for any of its
                                                                                              2. Average Precision (AP): This captures the joint
sentences with the query sentence. We use 3 settings
                                                                                                  effect of Precision and Recall. It computes preci-
considering different sentences in each document:
                                                                                                  sion at each rank of the predicted ranked list and
• 𝑆𝐵𝑇 𝐸 : Only Testimony or Evidence sentences
                                                                                                  then computes mean of these precision values.
• 𝑆𝐵𝑇 : Only Testimony sentences
• 𝑆𝐵𝐸 : Only Evidence sentences
                                                                                               2
                                                                                                   This dataset can be obtained from the authors on request
6.4. Results                                                similarity between argument phrases and presence
                                                            of co-references. Consider the following sentence for
Table 6 shows comparative evaluation results for various
                                                            which 𝑆𝑀𝑇 𝐸 incorrectly assigns a high score for query
baselines and our proposed technique. Average perfor-
                                                            𝑄5 (see Table 6) – T h e p o l i c e r e p o r t a l s o r e v e a l s t h a t
mance of 𝐵𝑀25𝑇 𝐸 is better than 𝐵𝑀25𝑎𝑙𝑙 indicating that
                                                            three pieces of pellets were found by the doctor in the
considering only Evidence and Testimony sentences for
                                                            b o d y o f d e c e a s e d M o n u . Here, except the 𝐴1 argument
representing any document, results in better prior case
                                                            (s o m e p o i s o n o u s c o m p o u n d s vs t h r e e p i e c e s o f p e l l e t s )
retrieval performance. Other two baselines 𝑆𝐵 (Sentence-
                                                            in Evidence Structure instances, other arguments
BERT) and 𝑆𝑀 (our proposed technique 𝑆𝑒𝑚𝑀𝑎𝑡𝑐ℎ) also
                                                            are similar in meaning. We get cosine similarity of
consider only Evidence and Testimony sentences rather
                                                            0.36 between s o m e p o i s o n o u s c o m p o u n d s and t h r e e p i e c e s
than considering all the sentences in a document. All
                                                            o f p e l l e t s which is misleading. It is not too low as
the baselines which consider only Testimony sentences,
                                                            compared to another case where there are semantically
perform poorly as compared to the corresponding tech-
                                                            similar argument phrases (e.g., cosine similarity between
niques using both Testimony and Evidence sentences. This
                                                            s o m e p o i s o n o u s c o m p o u n d s and a h e a v y c o n c e n t r a t i o n o f
highlights the importance of evidence information as
                                                            a r s e n i c is just 0.55 as shown in Table 5). As we are not
compared to using only witness testimony information
                                                            resolving co-references, we are missing a few relevant
for prior case retrieval as done in Ghosh et al. [1].
                                                            documents. E.g., 𝑆𝑀𝑇 𝐸 does not assign a high score for
   Considering the average performance across all the
                                                            the following document for query 𝑄3 (see Table 6) –
10 queries, our proposed technique 𝑆𝑀𝑇 𝐸 is the best per-
                                                            Instead of surrendering before the police, the deceased
forming technique in terms of both R-Prec and AP. The
                                                            had attempted to kill the police. In retaliation, he was
performance of 𝑆𝑀𝑇 𝐸 is the most consistent across the
                                                            s h o t b y t h e m . . This is because t h e m in the Evidence
diverse queries. It achieves minimum R-Prec of 0.24 (for
                                                            Structure instance for s h o t is not explicitly known to
𝑄1 ) as compared to other baselines like 𝐵𝑀25𝑎𝑙𝑙 , 𝐵𝑀25𝑇 𝐸
                                                            correspond to t h e p o l i c e in the previous sentence.
and 𝑆𝐵𝑇 𝐸 which have minimum R-Prec of 0 for some
queries. As described in Algorithm 2, 𝑆𝑀 uses Sentence-
BERT based similarity within sentences for producing 7. Conclusion and Future Work
an enhanced matching score. We experimented with a
variant of 𝑆𝑀 which does not rely on Sentence-BERT In this paper, we discussed several NLP techniques for
based similarity. This variant resulted in average R-Prec identifying evidence sentences, representing them in the
of 0.36 and MAP of 0.30 across all the 10 queries. Al- semantically rich Evidence Structure and retrieving rel-
though this is lower than 𝑆𝑀𝑇 𝐸 performance, the R-Prec evant prior cases by exploiting it. The proposed tech-
is still comparable with 𝐵𝑀25𝑇 𝐸 (avg R-Prec of 0.36) and niques are weakly supervised as they do not rely on any
better than that of 𝑆𝐵𝑇 𝐸 (avg R-Prec of 0.28).             manually annotated training data, except for the human
   For some queries, it is important to have some semantic expertise in designing the linguistic rules. Keeping in
understanding at sentence-level. For example, 𝑄4 , which mind the importance of witness testimonies in addition
contains “negation”, 𝑆𝐵 and 𝑆𝑀 can capture the query’s to evidences, we also extracted and represented the wit-
meaning in a better way. 𝑆𝑀 handles such negations ness testimonies using the same Evidence Structure. For
in a more principled manner as the Evidence Structure the application of prior case retrieval, we evaluated our
Instance captures negation as one of its arguments.         proposed technique along with several competent base-
   For 𝑆𝑀, the maximum matching score achieved for any lines, on a dataset of 10 diverse queries. We demonstrated
Evidence Structure Instance in a document, is considered that our technique performs comparably for most of the
as the overall matching score with the whole document. queries and is the best considering the overall perfor-
In contrast, 𝐵𝑀25 based techniques directly compute mance across all 10 queries. The results highlight the
matching score for the whole document as they do not contribution of evidence and testimony information in
rely on sentence structure. This is one limitation of 𝑆𝑀 improving prior case retrieval performance.
which we plan to address as a future work. However, as           In future, we plan to apply advanced representation
𝑆𝑀 computes matching scores for individual Evidence learning techniques for learning dense or embedded
Structure instances, it is able to provide better interpre- representation of an entire Evidence Structure instance.
tation for each relevant document in terms of the actual Also, we plan to automatically determine the best suited
sentences which provided the maximum matching score. retrieval technique (BM25, Sentence-BERT or SemMatch)
Analysis of errors: We analyzed cases where 𝑆𝑀𝑇 𝐸 for any query based on its nature. We plan to explore
was assigned a lower score to a relevant document or a ensemble of multiple retrieval techniques for improving
higher score to a non-relevant document. We discovered prior case retrieval performance further.
3 main reasons - missing or incorrect arguments
within Evidence Structure instances, misleading high
References                                                                           detection, in: Proceedings of the 2015 conference
                                                                                     on empirical methods in natural language process-
 [1] K. Ghosh, S. Pawar, G. Palshikar, P. Bhattacharyya,                             ing, 2015, pp. 440–450.
     V. Varma, Retrieval of prior court cases using wit-                        [12] D. Ji, P. Tao, H. Fei, Y. Ren, An end-to-end joint
     ness testimonies, JURIX (2020).                                                 model for evidence information extraction from
 [2] M. Palmer, D. Gildea, P. Kingsbury, The proposi-                                court record document, Information Processing &
     tion bank: An annotated corpus of semantic roles,                               Management 57 (2020) 102305.
     Computational linguistics 31 (2005) 71–106.                                [13] T. Gomes, M. Ladeira, A new conceptual frame-
 [3] P. Shi, J. Lin, Simple BERT models for rela-                                    work for enhancing legal information retrieval at
     tion extraction and semantic role labeling, CoRR                                the brazilian superior court of justice, in: Proceed-
     abs/1904.05255 (2019). URL: http://arxiv.org/abs/                               ings of the 12th International Conference on Man-
     1904.05255. a r X i v : 1 9 0 4 . 0 5 2 5 5 .                                   agement of Digital EcoSystems, 2020, pp. 26–29.
 [4] S. Hochreiter, J. Schmidhuber, Long short-term                             [14] J. Landthaler, B. Waltl, P. Holl, F. Matthes, Extend-
     memory, Neural computation 9 (1997) 1735–1780.                                  ing full text search for legal document collections
 [5] M. Honnibal, I. Montani, S. Van Landeghem,                                      using word embeddings., in: JURIX, 2016, pp. 73–82.
     A. Boyd, spaCy: Industrial-strength Natural Lan-                           [15] A. Trotman, A. Puurula, B. Burgess, Improvements
     guage Processing in Python, 2020. URL: https://doi.                             to bm25 and language models examined, in: Pro-
     org/10.5281/zenodo.1212303. doi:1 0 . 5 2 8 1 / z e n o d o .                   ceedings of the 2014 Australasian Document Com-
     1212303.                                                                        puting Symposium, 2014, pp. 58–65.
 [6] G. A. Miller, Wordnet: a lexical database for english,                     [16] J. Devlin, M.-W. Chang, K. Lee, K. Toutanova,
     Communications of the ACM 38 (1995) 39–41.                                      BERT: Pre-training of deep bidirectional transform-
 [7] J. Pennington, R. Socher, C. D. Manning, Glove:                                 ers for language understanding, arXiv preprint
     Global vectors for word representation, in: Pro-                                arXiv:1810.04805 (2018).
     ceedings of the 2014 conference on empirical meth-                         [17] C. Manning, P. Raghavan, H. Schutze, Introduc-
     ods in natural language processing (EMNLP), 2014,                               tion to information retrieval, Natural Language
     pp. 1532–1543.                                                                  Engineering 16 (2010) 100–103.
 [8] N. Reimers, I. Gurevych, Sentence-BERT: Sentence
     Embeddings using Siamese BERT-networks, in:
     Proceedings of the 2019 Conference on Empiri-
     cal Methods in Natural Language Processing and
     the 9th International Joint Conference on Natural
     Language Processing (EMNLP-IJCNLP), 2019, pp.
     3973–3983.
 [9] P. Bellot, A. Doucet, S. Geva, S. Gurajada, J. Kamps,
     G. Kazai, M. Koolen, A. Mishra, V. Moriceau,
     J. Mothe, M. Preminger, E. SanJuan, R. Schenkel,
     X. Tannier, M. Theobald, M. Trappett, Q. Wang,
     Overview of INEX 2013,                                    in: P. Forner,
     H. Müller, R. Paredes, P. Rosso, B. Stein (Eds.),
     Information Access Evaluation. Multilinguality,
     Multimodality, and Visualization - 4th Interna-
     tional Conference of the CLEF Initiative, CLEF
     2013, Valencia, Spain, September 23-26, 2013.
     Proceedings, volume 8138 of Lecture Notes in
     Computer Science, Springer, 2013, pp. 269–281.
     URL: https://doi.org/10.1007/978-3-642-40802-1_27.
     doi:1 0 . 1 0 0 7 / 9 7 8 - 3 - 6 4 2 - 4 0 8 0 2 - 1 \ _ 2 7 .
[10] M.-A. Cartright, H. A. Feild, J. Allan, Evidence find-
     ing using a collection of books, in: Proceedings of
     the 4th ACM workshop on Online books, comple-
     mentary social media and crowdsourcing, 2011, pp.
     11–18.
[11] R. Rinott, L. Dankin, C. Alzate, M. M. Khapra,
     E. Aharoni, N. Slonim, Show me your evidence-an
     automatic method for context dependent evidence