<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Document Level Embeddings for Identifying Similar Legal Cases and Laws (AILA 2020 shared Task)</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Intisar Almuslim</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Diana Inkpen</string-name>
          <email>diana.inkpen@uottawa.ca</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>School of Electrical Engineering and Computer Science</institution>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>University of Ottawa</institution>
          ,
          <addr-line>ON</addr-line>
          ,
          <country country="CA">Canada</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2020</year>
      </pub-date>
      <fpage>16</fpage>
      <lpage>20</lpage>
      <abstract>
        <p>In this age of legal big data, there are massive amounts of legal documents available online. The Information Retrieval systems play an important role in accurately retrieving the related information based on a given query. Therefore, there is a need for automated systems which can identify the set of the most relevant precedents (prior cases) and suitable laws (statues) for any situation. This automatic linking of the relevant documents helps lawyers to properly deal with related cases. This work seeks to develop such an automatic system to identify relevant prior cases and statutes for a given query. It has been submitted as a participation in the shared task track named Artificial Intelligence for Legal Assistance (AILA 2020). In this work, we propose a text similarity approach to retrieve previous cases and statues that are similar to a given case from a set of legal documents. For this legal document retrieval task, we present our approach that finds the similarity between a given query and prior cases/statutes documents by applying three variations of word representation based on Glove, Doc2Vec, and TFIDF methods. After feature extraction and documents vectorization, the similarity between each query document and all the prior cases/statute documents are determined using cosine similarity scores. Then, a ranked list of candidate documents is retrieved for each query. Our experiments demonstrate that TFIDF method achieved reasonable results when compared to Doc2Vec and Glove methods which usually need large training datasets.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Legal Information Retrieval</kwd>
        <kwd>Document similarity</kwd>
        <kwd>Word representation</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        In those countries with the Common Law systems like the United States, Canada, and India,
there are two main sources of law: the established laws (or statues) and the precedents (or
prior cases) [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. In the judgement of the Common Law system, the primary consideration is to
recognize the most related legal principles and the similar cases of a given situation. The method
of identifying such representative and associated documents, however, is a time-consuming
process. As a consequence, automating this procedure to be used for legal purposes is quite
important.
      </p>
      <p>
        In this work, we describe our methods that were submitted to the Artificial Intelligence for
Legal Assistance (AILA 2020) track [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] of the Forum for Information Retrieval Evaluation (FIRE
2020). This track defined two main tasks such as Task1: Precedent &amp; Statute Retrieval and Task2:
Rhetorical Role Labelling for Legal Judgements. Our team participated only in Task1 that contains
two subtasks. Subtask 1A: Identifying relevant prior cases for a given situation and Subtask 1B:
Identifying the most relevant statutes for a given situation. For both subtasks, a dataset of Indian
legal documents have been provided to the participants, with Indian statutes and prior cases
decided by Indian courts of Law. In order to find the similarity between the terms in the query
and the legal documents, we calculated the relevance between the query and the document by
utilizing three diferent word representations: GloVe, Doc2Vec, and TF-IDF.
      </p>
      <p>The rest of this paper is organized as follows: Section 2 presents related works in the area of
legal document retrieval; task description and dataset details are provided in Section 3; Section
4 explains the proposed methods; Section 5 relates to the evaluation results. Finally, Section 6
concludes the work with some directions of future improvements.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related Work</title>
      <p>
        With the increased availability of legal text in digital form, the focus on developing automated
legal information retrieval models and applications have received attention from the research
community. In recent years, there have been some competitions on the topic of legal information
retrieval. The FIRE 2017 IRLeD Track [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] defined two tasks; Catch Phrase Extraction and
Precedence Retrieval. The FIRE 2019 AILA Track [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] also focused on two tasks: to identify a
set of related prior cases as well as relevant statutes for a given situation. In the same manner,
COLIEE-2019 (Competition on Legal Information Extraction and Entailment) [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] has been held
as an associated event of the International Conference on Artificial Intelligence and Law (ICAIL).
The main objective of this competition is to develop techniques of retrieval, extraction and
entailment in the legal field.
      </p>
      <p>
        For achieving these tasks of retrieving the required legal documents, the submitted works
have used various traditional methods of Information Retrieval, as well as diferent Machine
Learning (ML) and Deep Learning (DL) methods. Among many techniques that have been
reported in the existing literature, the most applied ones are TF-IDF [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ], BM25 [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] or ensembles
of the two [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]. Also, embedding methods like Word2Vec, Sent2Vec, and Doc2Vec were widely
applied by many of the related studies [
        <xref ref-type="bibr" rid="ref10 ref8 ref9">8, 9, 10</xref>
        ]. Other researchers try to use ML methods
[
        <xref ref-type="bibr" rid="ref11 ref12">11, 12</xref>
        ] to capture semantic similarity. With the recent increase of DL technology and Natural
Language Processing (NLP), many researchers applied Neural Network (NN) models to legal
information retrieval [
        <xref ref-type="bibr" rid="ref13 ref9">9, 13</xref>
        ] . In addition to these methods, some studies focused on solving the
challenges by employing techniques such as keyphrase extraction [14] and text summarization
[15, 16].
      </p>
    </sec>
    <sec id="sec-3">
      <title>3. Dataset Details and Tasks Description</title>
      <p>In this work, we have carried out experiments using the dataset released by the organizers
of AILA 2020 shared task1. This dataset includes a collection of Indian legal documents, i.e.,
statutes in India and prior cases decided by Indian courts of Law. In addition, a set of 50
queries (Q1 to Q50) that contain the descriptions of the legal situations have been given to the
participants.</p>
      <p>The objective of the Task1A is to identify the relevant prior cases for the given query. To
perform it, a set of 3257 labeled prior cases (C1 to C3257) documents have been provided. The
participants of this task have to retrieve the most similar/ relevant case documents with respect
to the situation in the given query. For performing Task1B, 197 statues (S1 to S197) were released
that contain titles and textual descriptions and the task is to identify the most relevant statues
for each query.</p>
    </sec>
    <sec id="sec-4">
      <title>4. Methods Description</title>
      <p>In our approach, we have used diferent document embedding models with the cosine similarity
measure [17] to find the similarity between documents. We submitted 3 runs for each subtask
and used three diferent variations of document representations in these runs. For both subtasks,
the considered methods are (i) GloVe, (ii) Doc2Vec and (iii) TF-IDF in the three runs, respectively.
We have implemented our methodology in Python and the steps used in our approach are
presented in Figure 1.</p>
      <p>As input, the collection of prior cases, statutes and query documents are provided in the
form of text files. For all runs, the first step is the same which is the pre-processing of the text.
For every query and candidate document, we kept only words with letters and all the other
characters are removed as well as stop-words. All documents were tokenized, and every token
was converted to lowercase and lemmatized using the WordNet Lemmatizer from NLTK 2. After
the pre-processing phase, the steps performed in all the three methods are explained in detail in
the following subsections.</p>
      <sec id="sec-4-1">
        <title>4.1. GloVe-based Method</title>
        <p>In the first experiment, we have applied the pre-trained Global Vectors for Word Representation
(GloVe) 3 to represent the documents of prior cases/statutes and queries as vectors. GloVe
is provided by the Stanford NLP team [18] as various models from 25, 50, 100, 200 to 300
dimensions base on 2-, 6-, 42-, and 840-billion tokens. We used the one with 50 dimensions
and 6 billion tokens. The words with respect to a particular document are vectorized using
2http://www.nltk.org
3https://nlp.stanford.edu/projects/glove
GloVe. For those words that do not have pre-trained embeddings, we calculated and assigned to
them the average vector of all the word vectors. The average of all the word vectors of each
document is determined and that average represents the vector for that document. As a result,
the vector representations of all the prior case/statute documents and the query documents are
obtained.</p>
        <p>After that, the obtained vectors of queries and prior cases/statutes are compared using cosine
similarity. As illustrated, for subtask 1A, cosine similarity scores between each query and prior
case documents are calculated, sorted, and ranked. Similarly, for subtask 1B, cosine similarities
between each query and statute documents are found. The similarity scores are then sorted and
ranked. The final result gives a list of ranked documents for each query document.</p>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. Doc2Vec-based Method</title>
        <p>In this experiment, we first trained a Doc2Vec [ 19] model using the Gensim 4 library (Doc2Vec
and TaggedDocument from gensim.models.doc2vec) on 3257 case documents for subtask 1A, and
on 197 statute documents for subtask 1B. Then, this model is used for documents vectorization.
The Doc2Vec method is based on the Word2Vec method that was originally proposed by [20].
While the Word2Vec algorithm builds distributed semantic representations of unique words,
the Doc2Vec method is generally applicable to the text segments of any length – from sentences
to whole documents.</p>
        <p>For this method, each tokenized document is labelled with a tag to uniquely identify
documents. The hyperparameters applied are as follows: vector dimension: 100, number of epochs:
50. The trained model is then used to obtain the vector representation of each case document
for relevant prior cases task or of each statute document for relevant statutes task. Then, the
obtained vectors of queries and cases/statutes are compared using cosine similarity as the
description illustrated above, and the candidates are returned in decreasing order of similarity
score with the query.</p>
      </sec>
      <sec id="sec-4-3">
        <title>4.3. TFIDF-based Method</title>
        <p>In the last experiment, we used the TF-IDF [17] feature weighting measure. In this method, the
vector for each document is constructed using TF-IDF weights (or scores) which give high value
for a given term if that term appears often in that particular document and not frequently in
other documents. As mentioned earlier, after the vectorization step, the cosine similarity scores
between each query document and all the prior cases/statutes documents are determined and
then ranked.</p>
        <p>For each current query, the prior cases and the statutes are ranked based on the similarity
scores (the candidate document with highest similarity score is retrieved first). The TF-IDF
scores are obtained for the features by using scikit-learn library (Tfidf Vectorizer from
sklearn.feature_extraction.text). The similarity between each current query and the prior cases/statutes
are obtained using cosine_similarity from sklearn.metrics.pairwise.</p>
        <p>4https://radimrehurek.com/gensim/models/doc2vec.html</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Results and Discusion</title>
      <p>
        This section describes the results obtained in the two tasks by our proposed approaches. To
evaluate the performance of the submitted works, diferent evaluation metrics like MAP (Mean
Average Precision), P@10 (Precision@10), BPREF (Binary PREFerence-based measure) and
recip_rank (Reciprocal Rank) have been used [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. The primary measure was the MAP score,
as reported by the organizers, because this metric involves both Precision and Recall. The
performance metrics of all our models on the testing data are shown in Table 1 and Table 2.
      </p>
      <p>According to the results, Tf-IDF vectorization seems to perform better than the other
vectorization methods, namely GloVe and Doc2Vec, for the tasks of identifying the relevant prior
cases/statutes. Among all 26 runs submitted for Task 1A and 29 runs for Task 1B, our Tf-IDF
model got a rank of 17 and 11 on Task 1A and Task 2B, respectively based on the MAP metric.</p>
      <p>The embedding methods GloVe, and Doc2Vec have not achieved good results. The main reason
is probably because these methods require a huge amount of data to be trained on. In addition,
pre-trained general-domain models, like GloVe, do not perform well in very specialized domains
such as the legal domain. From the released results of all participants, the best performing
methods achieved a MAP score of 0.1573 and 0.3851 on Task1A and Task1B, respectively. These
results confirm that the tasks of finding relevant precedent cases and statutes are undeniably
challenging tasks.</p>
    </sec>
    <sec id="sec-6">
      <title>6. Conclusion</title>
      <p>This paper presents the system description of our participation in the Artificial Intelligence
for Legal Assistance (AILA 2020) track, which is an associated event of Forum for Information
Retrieval Evaluation (FIRE 2020). The track provided two main tasks. Our team only participated
in Task1. The shared task defined Task1 with two subtasks as follows: Task 1A,identifying
relevant prior cases for a given query and Task 1B, identifying the most relevant statutes for
a given query. We submitted three runs for each subtask. In both subtasks, our models are
based on Glove, Doc2Vec and TFIDF. For retrieving the relevant cases and statutes, initially the
documents were vectorized using the three methods. After vectorizing, the documents were all
ranked by their similarity to the query using cosine similarity.</p>
      <p>It can be seen from the final evaluation results that the TF-IDF-based method achieved
satisfactory results when compared to our other submitted methods. The models’ performance
can be improved by extracting diferent features or by applying diferent similarity measures.
Further improvement might also be achieved by using state-of-the-art deep learning models
pre-trained on legal text such as LEGAL-BERT [21].
[14] A. Mandal, K. Ghosh, A. Pal, S. Ghosh, Automatic catchphrase identification from legal
court case documents, in: Proceedings of the 2017 ACM on Conference on Information
and Knowledge Management, 2017, pp. 2187–2190.
[15] Y. Shao, Z. Ye, Thuir@ aila 2019: Information retrieval approaches for identifying relevant
precedents and statutes., in: FIRE (Working Notes), 2019, pp. 46–51.
[16] V. Tran, M. L. Nguyen, K. Satoh, Building legal case retrieval systems with lexical matching
and summarization using a pre-trained phrase scoring model, in: Proceedings of the
Seventeenth International Conference on Artificial Intelligence and Law, 2019, pp. 275–282.
[17] C. D. Manning, H. Schütze, P. Raghavan, Introduction to information retrieval, Cambridge
university press, 2008.
[18] J. Pennington, R. Socher, C. D. Manning, Glove: Global vectors for word representation, in:
Proceedings of the 2014 conference on empirical methods in natural language processing
(EMNLP), 2014, pp. 1532–1543.
[19] Q. Le, T. Mikolov, Distributed representations of sentences and documents, in: International
conference on machine learning, 2014, pp. 1188–1196.
[20] T. Mikolov, K. Chen, G. Corrado, J. Dean, Eficient estimation of word representations in
vector space, arXiv preprint arXiv:1301.3781 (2013).
[21] I. Chalkidis, M. Fergadiotis, P. Malakasiotis, N. Aletras, I. Androutsopoulos, Legal-bert:
The muppets straight out of law school, arXiv preprint arXiv:2010.02559 (2020).</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>P.</given-names>
            <surname>Bhattacharya</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Ghosh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Ghosh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Pal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Mehta</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Bhattacharya</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Majumder</surname>
          </string-name>
          ,
          <article-title>Fire 2019 aila track: Artificial intelligence for legal assistance</article-title>
          ,
          <source>in: Proceedings of the 11th Forum for Information Retrieval Evaluation</source>
          ,
          <year>2019</year>
          , pp.
          <fpage>4</fpage>
          -
          <lpage>6</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>P.</given-names>
            <surname>Bhattacharya</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Mehta</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Ghosh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Ghosh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Pal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Bhattacharya</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Majumder</surname>
          </string-name>
          ,
          <article-title>Overview of the FIRE 2020 AILA track: Artificial Intelligence for Legal Assistance</article-title>
          ,
          <source>in: Proceedings of FIRE 2020 - Forum for Information Retrieval Evaluation</source>
          ,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>A.</given-names>
            <surname>Mandal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Ghosh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Bhattacharya</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Pal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Ghosh</surname>
          </string-name>
          ,
          <article-title>Overview of the fire 2017 irled track: Information retrieval from legal documents</article-title>
          .,
          <source>in: FIRE (Working Notes)</source>
          ,
          <year>2017</year>
          , pp.
          <fpage>63</fpage>
          -
          <lpage>68</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>J.</given-names>
            <surname>Rabelo</surname>
          </string-name>
          , M.-
          <string-name>
            <given-names>Y.</given-names>
            <surname>Kim</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Goebel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Yoshioka</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Kano</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Satoh</surname>
          </string-name>
          ,
          <article-title>A summary of the coliee 2019 competition</article-title>
          , in
          <source>: JSAI International Symposium on Artificial Intelligence</source>
          , Springer,
          <year>2019</year>
          , pp.
          <fpage>34</fpage>
          -
          <lpage>49</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>R.</given-names>
            <surname>Rameshkannan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Rajalakshmi</surname>
          </string-name>
          , Dlrg@ aila
          <year>2019</year>
          :
          <article-title>context-aware legal assistance system</article-title>
          ,
          <source>Proceedings of FIRE</source>
          (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Zhao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Ning</surname>
          </string-name>
          , L. Liu,
          <string-name>
            <given-names>C.</given-names>
            <surname>Huang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Kong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Han</surname>
          </string-name>
          ,
          <string-name>
            <surname>Z</surname>
          </string-name>
          . Han,
          <article-title>Fire2019@ aila: Legal information retrieval using improved bm25</article-title>
          .,
          <source>in: FIRE (Working Notes)</source>
          ,
          <year>2019</year>
          , pp.
          <fpage>40</fpage>
          -
          <lpage>45</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>R.</given-names>
            <surname>More</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Patil</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Palaskar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Pawde</surname>
          </string-name>
          ,
          <article-title>Removing named entities to find precedent legal cases</article-title>
          .,
          <source>in: FIRE (Working Notes)</source>
          ,
          <year>2019</year>
          , pp.
          <fpage>13</fpage>
          -
          <lpage>18</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>S.</given-names>
            <surname>Kayalvizhi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Thenmozhi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Aravindan</surname>
          </string-name>
          ,
          <article-title>Legal assistance using word embeddings</article-title>
          .,
          <source>in: FIRE (Working Notes)</source>
          ,
          <year>2019</year>
          , pp.
          <fpage>36</fpage>
          -
          <lpage>39</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>S.</given-names>
            <surname>Mandal</surname>
          </string-name>
          , S.
          <string-name>
            <surname>D. Das</surname>
          </string-name>
          ,
          <article-title>Unsupervised identification of relevant cases &amp; statutes using word embeddings</article-title>
          .,
          <source>in: FIRE (Working Notes)</source>
          ,
          <year>2019</year>
          , pp.
          <fpage>31</fpage>
          -
          <lpage>35</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>S.</given-names>
            <surname>Renjit</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. M.</given-names>
            <surname>Idicula</surname>
          </string-name>
          , Cusat nlp@
          <fpage>aila</fpage>
          -fire2019:
          <article-title>Similarity in legal texts using document level embeddings</article-title>
          .,
          <source>in: FIRE (Working Notes)</source>
          ,
          <year>2019</year>
          , pp.
          <fpage>25</fpage>
          -
          <lpage>30</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>T. F.</given-names>
            <surname>Paulino-Passos</surname>
          </string-name>
          ,
          <string-name>
            <surname>G.</surname>
          </string-name>
          ,
          <article-title>Retrieving legal cases with vector representations of text</article-title>
          ,
          <source>Proceedings of the 6th Competition on Legal Information Extraction/Entailment. COLIEE</source>
          (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>T. A.</given-names>
            <surname>El</surname>
          </string-name>
          <string-name>
            <surname>Hamdani</surname>
          </string-name>
          ,
          <string-name>
            <surname>R.</surname>
          </string-name>
          ,
          <article-title>Coliee case law competition task 1: The legal case retrieval task</article-title>
          ,
          <source>Proceedings of the 6th Competition on Legal Information Extraction/Entailment. COLIEE</source>
          (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>J.</given-names>
            <surname>Rossi</surname>
          </string-name>
          , E. Kanoulas,
          <article-title>Legal information retrieval with generalized language models</article-title>
          ,
          <source>Proceedings of the 6th Competition on Legal Information Extraction/Entailment. COLIEE</source>
          (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>