<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Intelligence for Legal Assistance</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Paheli Bhattacharya</string-name>
          <email>paheli@cse.iitkgp.ac.in</email>
          <xref ref-type="aff" rid="aff4">4</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Parth Mehta</string-name>
          <email>parth.mehta126@gmail.com</email>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Kripabandhu Ghosh</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Saptarshi Ghosh</string-name>
          <email>saptarshi@cse.iitkgp.ac.in</email>
          <xref ref-type="aff" rid="aff4">4</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Arindam Pal</string-name>
          <email>arindamp@gmail.com</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Arnab Bhattacharya</string-name>
          <email>arnabb@cse.iitk.ac.in</email>
          <xref ref-type="aff" rid="aff3">3</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Prasenjit Majumder</string-name>
          <email>prasenjit.majumdar@gmail.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Legal NLP</string-name>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>DA-IICT Gandhinagar</institution>
          ,
          <country country="IN">India</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Data61, CSIRO and Cyber Security CRC</institution>
          ,
          <addr-line>Sydney, New South Wales</addr-line>
          ,
          <country country="AU">Australia</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Indian Institute of Science Education and Research (IISER)</institution>
          ,
          <addr-line>Kolkata</addr-line>
          ,
          <country country="IN">India</country>
        </aff>
        <aff id="aff3">
          <label>3</label>
          <institution>Indian Institute of Technology Kanpur</institution>
          ,
          <country country="IN">India</country>
        </aff>
        <aff id="aff4">
          <label>4</label>
          <institution>Indian Institute of Technology Kharagpur</institution>
          ,
          <country country="IN">India</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2020</year>
      </pub-date>
      <abstract>
        <p>The FIRE 2020 AILA track focused on two tasks - (i) Retrieving relevant Prior cases and Statutes given a factual description, and (ii) Rhetorical labelling of sentences in a legal case document, where the rhetorical roles are - Facts of the case, Ruling by the Lower Court, Argument, Statute, Precedent, Ratio of the decision and Ruling by the Present Court. Both the tasks were based on publicly available case documents from the Indian Supreme Court judiciary. Legal data analytics, Prior case retrieval, Statute retrieval, Legal facts, Rhetorical role labelling, Legal IR, Common Law system, which is followed by most countries (e.g., UK, USA, Canada, Australia, India etc.) has two primary sources - Precedents and Statutes. Precedents are the prior cases decided in the Courts of law. Statutes are bodies of written law, such as the Constitution of a country.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>rOcid</p>
      <p>Also, a case document from the Indian judiciary is usually very long and unstructured, lacking
section and paragraph headings. This makes it dificult for a reader of the document to identify
where the facts of the case are written, which sentences mention the arguments made in the
case, what was the final judgement and so on. To address this research problem, we introduce
Task 2: Rhetorical Role Labeling for Legal Judgements.</p>
      <p>
        AILA 2020 witnessed a participation of 15 teams with 14 of them submitting the working
notes. We received a total of 74 runs across the three tasks. Apart from teams from India, we
had teams from China, Botswana, Italy, Austria and Canada. Also we had teams from both
academic institutions as well as the industry.
1.1. Task 1: Precedent and Statute Retrieval
In this task, the aim is to retrieve relevant prior-cases and statutes for a given factual scenario,
from a large pool of prior-case documents and statutes. Further details of the task are given in
the overview paper of AILA-2019 [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ].
      </p>
      <p>
        For training, we provide the dataset of AILA-2019 [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. For the test/evaluation data, we provide
a set of additional 10 queries, that describe factual scenarios in natural English language. The
pool / candidate statutes from which the relevant ones are to be retrieved, is the same as that in
AILA-2019 dataset [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. A set of 343 documents were added to the existing pool of prior-case
documents in AILA-2019 [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. Out of these 343 documents, 43 documents were the relevant prior
cases for the 10 queries in the test set. The remaining 300 documents were sampled based on
the text similarity (cosine similarity between the document vectors were in [0.3, 0.5]) between
the existing documents in the prior-case pool. Hence, for the present task, the resultant number
of prior-case documents is 3, 257 (from which, prior cases relevant to a given query need to be
retrieved).
      </p>
      <p>Also we provide a set of 197 statutes (Sections of Acts) from Indian law, that are relevant to
some of the queries stated above. We provide the participants with the title and description of
these statutes.</p>
      <p>For each query, the task was to retrieve the most similar/relevant precedent case documents
and statutes with respect to the situation in the given query.</p>
      <p>
        Similar research has been done on Chinese legal case documents [
        <xref ref-type="bibr" rid="ref2 ref3">2, 3</xref>
        ] that dealt with the
task of retrieving statutes for a given fact. Note that, in addition to retrieving statutes, we also
consider retrieving prior-cases for the query. Also, constructing a large dataset for the task to
train supervised models as in [
        <xref ref-type="bibr" rid="ref2 ref3">2, 3</xref>
        ] is not practical in the context of Indian legal documents.
This is because, unlike Chinese legal documents, Indian legal documents are not well-structured
and it is dificult to extract the facts of the cases automatically [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] for creating the dataset.
1.2. Task 2: Rhetorical Role Labeling for Legal Judgements
Since Indian legal case documents are unstructured, there is a need to design systems that
can automatically segment these documents into coherent, meaningful parts. This can not
only enhance the readability of the documents but also has applications in downstream tasks
such as summarization, case-law analysis, semantic search and so on. We introduce the task
of rhetorical role labeling of sentences in legal case judgements in AILA-2020. The task is to
assign one of the following labels to each sentence in a legal case document. We consider the
following seven (07) rhetorical labels/semantic segments [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]:
• Facts: legal situation that led to filing the case
• Ruling by Lower Court: since we consider documents from the Supreme Court of India,
there was some preliminary ruling given at the lower courts, e.g., High Court, Tribunal,
etc.
• Arguments: arguments made by the contending parties
• Precedents: citation to relevant prior cases
• Statutes: citation to relevant statutes
• Ratio of the decision: reasoning behind the final judgement
• Ruling by Present Court: final judgement given by the Supreme Court of India
A state-of-the-art model that addresses this task for Indian legal documents is [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ], where a
neural model is used for the segmentation.
      </p>
    </sec>
    <sec id="sec-2">
      <title>2. Dataset</title>
      <p>We consider case documents from the Supreme Court of India and Statutes from the Indian
judiciary.</p>
      <p>
        Task 1: The training dataset for Task 1 was the AILA-2019 dataset [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] (available at https:
//github.com/Law-AI/aila-2019-dataset). There were 50 queries and 197 statutes.
      </p>
      <p>
        For the present task (AILA-2020), the pool of prior-cases was extended to having 3, 257
documents, as mentioned in Section 1.1. The test dataset of 10 queries was created in the same
way as described in [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ].
      </p>
      <p>
        Task 2: The dataset made publicly available by [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] was used as the training dataset1. There
were 50 documents containing 9, 308 sentences in total across all the documents. The rhetorical
labels were assigned by law experts from a reputed law school in India.
      </p>
      <p>
        As the test set, we consider a set of 10 additional case documents. We randomly selected
2 documents from each of the 5 law domains mentioned in [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. These documents were then
given to a law expert for annotating every sentence with one of the rhetorical labels. There are
a total of 1, 905 sentences in the test set.
      </p>
    </sec>
    <sec id="sec-3">
      <title>3. Evaluation</title>
      <p>
        For both the Tasks, evaluation was done on the test dataset. For Task 1, the same evaluation
metrics as in [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] were used – Mean Average Precision (MAP), Precision@10 (P10), BPREF and
      </p>
      <sec id="sec-3-1">
        <title>1https://github.com/Law-AI/semantic-segmentation</title>
        <p>Reciprocal rank (recip_rank). The trec_eval tool 2 was used for computing the metrics stated
above. We choose MAP as the primary measure since it incorporates both Precision and Recall.</p>
        <p>For Task 2, we use the standard Recall, Precision and F1-Scores. The documents have
a considerable variation in their size. Moreover, even within a document, there is a class
imbalance among the 7 categories / rhetorical roles. Hence we use macro-averaging at both
document-level and category-level. The scores were calculated as below:
1. Recall, Precision and F-score were computed for each category of labels within each
document.
2. The score for each document in a run were computed by averaging the scores for all
seven categories in that document.
3. Finally, the overall scores for a run are computed by averaging the scores for each
document.</p>
        <p>For task-2 we additionally report the overall Accuracy for each submitted run. Accuracy is
micro-averaged across documents and classes, i.e., it is measured as the fraction of sentences
correctly classified out of the 1, 905 test sentences.
4. Methodologies for Task 1: Precedent Retrieval and Statute</p>
        <p>Retrieval
For the first task of retrieving relevant prior/ precedent cases (Task 1a) , we received a total of
26 runs from 10 participating teams. For the second task of retrieving relevant statutes (Task
1b) , we received a total of 27 runs from 12 participating teams. The comparative results are in
Table 1 (for Task 1a) and Table 2 (for Task 1b). We briefly describe below the methodologies used
by each team in each of their runs. Details can be found in the working notes of the respective
submissions.</p>
        <p>
          Task 1a : Precedent Retrieval : We briefly describe the methodologies submitted by the
various teams for the task:
• UB [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ] : The team was from the University of Botswana. In their first submitted run UB-1,
they weighed the terms in the query and documents using TF-IDF. In their second run
(UB-2), they extracted key concepts and used TF-IDF for retrieval. The best performance
in terms of MAP, BPREF and recip_rank for the task was by their third run, UB-3, where
they use Terrier 4.2 KL divergence model.
• double_liu_2020 [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ] : This team is afiliated to the Heilongjiang Institute of Technology,
China. They extract the top 50% of the words based on their IDF scores as the search
keywords in their first and third runs. In the second run, they used all the words as
search keywords. In terms of MAP, BPREF in Task 1a, their third run (double_liu_2020_3)
performed the second best.
experiment with diferent methods for stopword removal as the preprocessing step.
• SSNCSE_NLP [10]: This team is from Sri Siva Subramaniya Nadar College of Engineering,
        </p>
        <p>India. They use BM-25 and TF-IDF for the task.
• fs_hit_1 [11] : This team is from the Foshan University, China. They used BM25 and
TF-IDF similarity in their first and second runs. Their second run was the second best
• fs_hit_2 [12] : This team from the Foshan University, China and Heilongjiang Institute
of Technology, China, explored diferent Language models for the task. In their first run,
they use the language model assorting algorithm of Indri. In the second run, language
model assorting algorithm of Lucene is used. In the third run, they use language Model
with Dirichlet Similarity.
this task in brief. See Table 2 for a comparison among the performances of the methodologies.
• scnu 3 : This team has its members from the South China Normal University. It was
3Working note not submitted.</p>
        <p>the only team that modelled the the task as a supervised task and performed training
using the training dataset provided. Their first run that uses BERT is the best performing
method in the statute retrieval task in terms of MAP, BPREF and P@10. They are second
best in terms of recip_rank. They have also experimented with diferent supervised
methods in their second and third runs.
• SSNCSE_NLP [10]: This team is from Sri Siva Subramaniya Nadar College of Engineering.</p>
        <p>
          They use BM-25 and TF-IDF for the task. They get the second best MAP scores for the
task.
• IMS_UNIPD [13]: This team is afiliated to the Information Management System (IMS)
Group, University of Padua. They used BM-25 on the lemma forms of the words in the
query and candidate case documents for the bm25_lemma run; used TF-IDF weighting
on the lemma forms in tfidf_lemma run and TF-IDF weighting on the stemmed words in
the tfidf_stem run. They perform second best in terms of P@10.
• UB [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ] : The team was from the University of Botswana. In their first submitted run
they weighed the terms in the query and documents using Tf-IDF. In the second run,
they extracted key concepts and used TF-IDF for retrieval. This method acheives the best
recip_rank for the task. In their third run they use Terrier 4.2 KL divergence model.
• SSN_NLP [16]: The team has its members from the SSN College Of Engineering. They
use BM-25 for the statute retrieval task.
• Lawnics [14]: This team is from Lawnics Technologies, India. They use BM25 in their
ifrst run. They explore Law2Vec embedings in their second run. They achieve second
best BPREF scores for the task.
• TUW_Informatics [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ]: This team from TU Wien experiment with diferent stopword
lists. In the first run, basic, preprocessing is performed on the documents and retrieval is
through BM-25 algorithm. In their next two runs, word_count and false_friends, they
experiment with diferent methods for stopword removal as the preprocessing step.
• Uottawa_NLP [15]: This team is from the University of Ottawa. After preprocessing the
documents, they used Glove, Doc2Vec and TF-IDF based methods, for the first, second
and third runs respectively.
• fs_hu [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ] : This team from the Foshan University, China used TF-IDF and Jaccard to
compute similarity.
• fs_hit_1 [11] : This team is from the Foshan University, China. They experiment with
diferent Language Models with Dirichlet smoothing and JM Smoothing with diferent
hyper-parameters.
• fs_hit_2 [12] : This team, also from the Foshan University, China, explored diferent
        </p>
        <p>Language models for the task tuned with diferent hyper-parameters.
• nlpninjas 4 : This team is from Deloitte USI. They combined unigrams and bigrams.</p>
        <p>They used BM25 for retrieval.</p>
        <p>For Task 1a (Precedent retrieval) the best performing run achieves a MAP of 0.1573. The
methods submitted were all unsupervised in nature. The training data provided was mainly
used to tune the hyperparameters of the retrieval models (eg. BM-25). Embedding methods like
Doc2Vec, Glove were also used but the performance was not good. This is probably because
these methods require a huge amount of data to be trained on.</p>
        <p>For Task 1b (Statute retrieval) the best performing run achieved a MAP of 0.3851. The method
used BERT for extracting deep semantic features. Law2Vec embeddings have also been explored
for the task. Only 1 team (scnu) used training data to train supervised models – BERT, LSTM,
and a Neural Ranking model. Other teams used the training data to tune the hyperparameters
of unsupervised methods.
5. Methodologies for Task 2: Rhetorical Role Labeling for Legal</p>
        <p>Judgements
We received 21 runs from 9 teams for the task. Table 3 compares the performance of the various
runs. Brief descriptions of the methods are as follows. Details can be found in the working
notes of the respective submissions.</p>
        <p>• ju_nlp [17] : The team from Jadavpur University, India was the best performing team
in terms of F-Score and Recall. They used a state-of-the-art transformer architecture
ROBERTA along with BiLSTM for rhetorical role classification. The diferent runs are for
diferent epochs of the model training.
• heu_gjm [18] : The team has its members from Harbin Engineering University Harbin,
and Foshan University, China. They combine TF-IDF features and deep semantic features
using BERT. Logistic regression, linear kernel SVM and AdaBoost are used as classifiers.</p>
        <p>
          The BERT model with LogisticRegression gave the best precision for the task.
• double_liu [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ] : This team is afiliated to the Heilongjiang Institute of Technology, China.
        </p>
        <p>
          They experiment with bag-of-words based features along with SVM and Adaboost as
classifiers. They also use BERT for the task. The BERT model gave the best accuracy.
• spectre [19] : This team from BITS Pilani, India used ROBERTA and a fully connected
layer for classification.
• LAWNICS [14] : This team from Lawnics Technologies, India experimented with BERT
for extracting features. For classification they use a fully connected layer and a max
pooling layer for the diferent submitted runs.
• fs_hu [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ]: The team has its members from the Foshan University, China. They have
experimented with BERT using diferent random seeds for the runs.
        </p>
      </sec>
      <sec id="sec-3-2">
        <title>4Working note not submitted</title>
        <p>documents comprising of 1, 905 sentences. Numbers in bold and underline indicate the best and the
second-best performing methods corresponding to the evaluation metrics. Rows are sorted in decreasing
order of FScore (primary measure).</p>
        <p>Team
ju_nlp
ju_nlp
heu_gjm
heu_gjm
ju_nlp
double_liu
heu_gjm
spectre
LAWNICS
fs_hu
fs_hit_1
fs_hit_2
fs_hit_1
fs_hit_2
fs_hit_1
double_liu
SSNCSE_NLP
SSNCSE_NLP
double_liu</p>
        <p>fs_hu
LAWNICS</p>
        <p>Run
ju_nlp_2
ju_nlp_3
heu_gjm_1
heu_gjm_2
ju_nlp_1
double_liu_3
heu_gjm_3
spectre_1
lawnics_2
fs_hu_1
fs_hit_1_3
fs_hit_2_1
fs_hit_1_2
fs_hit_2_2
fs_hit_1_1
double_liu_2
ssncse_nlp_2
ssncse_nlp_1
double_liu_1
fs_hu_2
lawnics_1
by BERT with diferent random seeds.
• fs_hit_2 [12]: This team from the Foshan University, China and Heilongjiang Institute
of Technology, China, experiment with both TF-IDF features and BERT-based features
for the task.</p>
        <p>aspect.
• SSNCSE_NLP [10] : This team is afiliated to the Sri Siva Subramaniya Nadar College
of Engineering, India. They experiment with FastText and TF-IDF from the feature
engineering aspect, and Multi-layer perpceptron and Random Forest from the classifier
We find that the best performing method which achieved an FScore of
0.468 used ROBERTA
which is a state-of-the-art deep learning model. We observe that BERT was applied by almost
all the teams, with diferent classifiers (LR, FC etc.) Traditional Machine Learning approaches
(SVM, Random Forest etc.) were also tried out and Bag-of-Words based feature-vector was also
explored. Deep Learning methods that could extract deep semantic features were shown to
perform much better than traditional feature based approaches.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>6. Concluding Discussions</title>
      <p>The FIRE 2020 AILA track has created benchmark datasets for two important tasks in the field of
legal data analytics. We retained AILA 2019’s task of retrieving relevant statutes and precedents
for a query. We created a new task on rhetorical role labelling of sentences in Indian legal
documents. For the precedent retrieval task, we conclude from the results that it is a challenging
task, mainly because of the diference in length of the query and a prior-case document. For
the statute retrieval task, training BERT using very little amount of data (50 queries and their
corresponding gold standard statutes), shows promising results. For the rhetorical role labelling
task, participants have used state-of-the-art deep learning techniques like ROBERTA and BERT,
which gives good results. In the future, we plan to extend the dataset of both the tasks.
Acknowledgements: The track organizers thank all the participants for their interest in
this track. We also thank the FIRE 2020 organizers for their support in organizing the track.
The research is partially supported by SERB, Government of India, through a project titled
“NYAYA: A Legal Assistance System for Legal Experts and the Common Man in India”. Paheli
Bhattacharya is supported by a PhD Fellowship from Tata Consultancy Services.
domain knowledge, in: Proceedings of FIRE 2020 - Forum for Information Retrieval
Evaluation, 2020.
[10] N. N. A. Balaji, B. Bharathi, J. Bhuvana, Legal information retrieval and rhetorical role
labelling for legal judgements, in: Proceedings of FIRE 2020 - Forum for Information
Retrieval Evaluation, 2020.
[11] M. Wu, Z. Wu, W. Xiangyu, Z. Han, Retrieval model and classification model for aila2020,
in: Proceedings of FIRE 2020 - Forum for Information Retrieval Evaluation, 2020.
[12] Y. Xu, T. Li, Z. Han, The language model for legal retrieval and bert-based model for
rhetorical role labeling for legal judgments, in: Proceedings of FIRE 2020 - Forum for
Information Retrieval Evaluation, 2020.
[13] G. M. Di Nunzio, A study on lemma vs stem for legal information retrieval using r tidyverse,
in: Proceedings of FIRE 2020 - Forum for Information Retrieval Evaluation, 2020.
[14] J. Arora, T. Patankar, A. Shah, S. Joshi, Artificial intelligence as legal research assistant, in:</p>
      <p>Proceedings of FIRE 2020 - Forum for Information Retrieval Evaluation, 2020.
[15] I. Almuslim, D. Inkpen, Document level embeddings for identifying similar legal cases
and laws (aila 2020 shared task), in: Proceedings of FIRE 2020 - Forum for Information
Retrieval Evaluation, 2020.
[16] K. S., T. D., C. Aravindan, Best matching algorithm to identify and rank the relevant
statutes, in: Proceedings of FIRE 2020 - Forum for Information Retrieval Evaluation, 2020.
[17] S. B. Majumder, D. Das, Rhetorical role labelling for legal judgements using roberta, in:</p>
      <p>Proceedings of FIRE 2020 - Forum for Information Retrieval Evaluation, 2020.
[18] J. Gao, H. Ning, Z. Han, L. Kong, H. Qi, Legal text classification model based on text
statistical features and deep semantic features, in: Proceedings of FIRE 2020 - Forum for
Information Retrieval Evaluation, 2020.
[19] R. Jain, A. Agarwal, Y. Sharma, Spectre@aila-fire2020: Supervised rhetorical role
labeling for legal judgments using transformers, in: Proceedings of FIRE 2020 - Forum for
Information Retrieval Evaluation, 2020.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>P.</given-names>
            <surname>Bhattacharya</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Ghosh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Ghosh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Pal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Mehta</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Bhattacharya</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Majumder</surname>
          </string-name>
          ,
          <article-title>Overview of the FIRE 2019 AILA track: Artificial Intelligence for Legal Assistance</article-title>
          ,
          <source>in: Proceedings of FIRE 2019 - Forum for Information Retrieval Evaluation</source>
          ,
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>P.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Niu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , L. Zhang, S. Niu,
          <article-title>Modeling dynamic pairwise attention for crime classification over legal articles</article-title>
          ,
          <source>in: Proceedings of ACM SIGIR Conference on Research &amp; Development in Information Retrieval, SIGIR '18</source>
          ,
          <year>2018</year>
          , pp.
          <fpage>485</fpage>
          -
          <lpage>494</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>P.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Fan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Niu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , J. Guo,
          <article-title>Hierarchical matching network for crime classification</article-title>
          ,
          <source>in: Proceedings of ACM SIGIR Conference on Research and Development in Information Retrieval, SIGIR'19</source>
          ,
          <year>2019</year>
          , pp.
          <fpage>325</fpage>
          -
          <lpage>334</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>P.</given-names>
            <surname>Bhattacharya</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Hiware</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Rajgaria</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Pochhi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Ghosh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Ghosh</surname>
          </string-name>
          ,
          <article-title>A Comparative Study of Summarization Algorithms Applied to Legal Case Judgments</article-title>
          ,
          <source>in: Proc. European Conference on Information Retrieval (ECIR)</source>
          ,
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>P.</given-names>
            <surname>Bhattacharya</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Paul</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Ghosh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Ghosh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Wyner</surname>
          </string-name>
          ,
          <article-title>Identification of rhetorical roles of sentences in indian legal judgments</article-title>
          ,
          <source>in: Proc. International Conference on Legal Knowledge and Information Systems (JURIX)</source>
          ,
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>T.</given-names>
            <surname>Leburu-Dingalo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N. P.</given-names>
            <surname>Motlogelwa</surname>
          </string-name>
          , E. Thuma,
          <string-name>
            <given-names>M.</given-names>
            <surname>Modungo</surname>
          </string-name>
          ,
          <article-title>Ub at fire 2020 precedent and statute retrieval</article-title>
          ,
          <source>in: Proceedings of FIRE 2020 - Forum for Information Retrieval Evaluation</source>
          ,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>L.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Liu</surname>
          </string-name>
          , Z. Han,
          <article-title>Query revaluation method for legal information retrieval</article-title>
          ,
          <source>in: Proceedings of FIRE 2020 - Forum for Information Retrieval Evaluation</source>
          ,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Kong</surname>
          </string-name>
          ,
          <article-title>Language model-based approaches for legal assistance</article-title>
          ,
          <source>in: Proceedings of FIRE 2020 - Forum for Information Retrieval Evaluation</source>
          ,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>T.</given-names>
            <surname>Fink</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Recski</surname>
          </string-name>
          ,
          <string-name>
            <surname>A. Hanbury,</surname>
          </string-name>
          <article-title>Fire2020 aila track: Legal domain search with minimal</article-title>
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>