<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Rhetorical Labeling for Legal Judgements using fastText</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Tebo Leburu-Dingalo</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Edwin Thuma</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Gontlafetse Mosweunyane</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Nkwebi Peace Motlogelwa</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Department of Computer Science, University of Botswana</institution>
        </aff>
      </contrib-group>
      <abstract>
        <p>This paper describes our participating systems in the FIRE AILA 2021 shared task on predicting rhetorical roles for sentences in a legal judgement document. In particular we propose three multi-class classifiers to predict for each of the sentences a rhetorical role from the following: facts, arguments, ratio of the decision, precedent, statutes, ruling of lower court and ruling of present court. Each of the classifiers uses a supervised fastText model. As input tokens the first classifier uses unigrams, the second one used bigrams and the last one uses trigrams. Our system that uses trigrams attains an F-Score of 0.340 followed closely by the bigram system at 0.338 while the baseline has a score of 0.317.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Rhetorical Role</kwd>
        <kwd>Facts</kwd>
        <kwd>Arguments</kwd>
        <kwd>fastText</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        ratio of the decision and ruling by the present court . The task was started as part of FIRE
2020 AILA Track [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. For the task, a training dataset consisting of 50 documents containing 9,
308 sentences in total with rhetorical labels assigned by law experts was used, while the test
dataset had an additional set of 10 case documents. The dataset was provided by [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. The 21
runs submitted by the 9 teams employed diferent methods for rhetorical role labelling. The best
performing system in terms of F-Score and Recall was by team ju_nlp [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] who experimented
with the transformer architecture ROBERTA (state-of-the-art deep learning model) and BiLSTM
with diferent epochs of the model training for the diferent runs. Scores attained for F-score
and Recall were 0.468 and 0.501 respectively. Team heu_gjm [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] deployed TF-IDF features and
deep semantic features using BERT, with diferent classifiers namely Logistic Regression, Linear
Kernel SVM and AdaBoost. The BERT model with Logistic Regression gave the best precision for
the task at 0.541. Team double_liu [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] used bag-of-words based features with SVM and Adaboost
as classifiers. The team also used the BERT model, which outperformed all systems submitted in
terms of accuracy at 0.619. Results from the task show that even with the use of complex deep
learning methods rhetorical labelling remains a dificult problem to solve as none of the methods
proposed achieves optimal performance. In this work we attempt to address the rhetorical role
labelling problem through the use of a fastText classifier. FastText is a linear classifier which has
been shown to perform on par with deep learning algorithms in text classification while training
at faster speeds and utilizing less processing power [
        <xref ref-type="bibr" rid="ref8 ref9">8, 9</xref>
        ]. In addition our choice of the fastText
model is motivated by its capability to support out of dictionary words which can be useful
when working with domain specific corpora. Furthermore, the model allows the use of phrases
as input tokens to preserve word order, a practice that has proven efective for classification
problems [
        <xref ref-type="bibr" rid="ref10 ref11 ref12">10, 11, 12</xref>
        ]. Thus alongside exploring the efectiveness of the fastText classifier in the
detection of rhetorical roles we will further investigate the efectiveness of using bigrams and
trigrams in improving classification accuracy.
      </p>
    </sec>
    <sec id="sec-2">
      <title>2. Methodology</title>
      <p>Rhetorical Labelling (RL) entails segmenting a document into several coherent sentences and
assigning rhetorical roles to these sentences. A rhetorical role describes a semantic function
that a sentence plays in a document. The task calls for the labelling of sentences into seven
roles as follows: Facts referring to the chronology of events that led to filling the case, ruling by
lower court, Arguments of contending parties, relevant cited statute, relevant precedent cited,
ratio of the decision referring to rationale/reasoning given for the final judgement and ruling by
present court referring to the final decision given by the court. For our study, the task will be
approached as a classification problem where each role is considered a class, and each instance
of a sentence in a document is classified into only one of the classes. In our experiments we
deploy a supervised fastText text classifier trained on the provided Task 1 dataset.</p>
      <p>
        Fasttext 1 is an open source toolkit developed for efective learning of text representations
and text classification [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. FastText incorporates the context of words in its embeddings as
surrounding words are taken into account when learning a word representation. Furthermore
fastText represents each word as a bag of character n-grams in addition to the word itself which
is useful for corpora with rare non-dictionary words. Text representations are obtained by
averaging word representations. The representations are then fed into a linear classifier and
classes determined by deploying a loss function that computes probability distribution over
predefined class labels. By default fastText accepts unigrams as input tokens, however this
can be varied to for instance bigrams and trigrams. The loss function generally used is the
softmax which is can be changed to hierarchical softmax for larger number of classes to speed
up training.
      </p>
    </sec>
    <sec id="sec-3">
      <title>3. Experimental Setup</title>
      <sec id="sec-3-1">
        <title>3.1. Dataset</title>
        <p>The training dataset consists of 70 documents with variable number of sentences of diferent
lengths. The test data consists of 10 documents also with varying number of sentences. Each of
the sentences in the training dataset is annotated with one of the seven classes. It was noted
that the data was unbalanced with an unequal distribution among the classes. Measures to be
used for evaluation are Precision, Recall and F1 Score.</p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Platform</title>
        <p>The Python Programming Language and its libraries is used for all experiments. The Fasttext
Open Source Library is used for classification.</p>
      </sec>
      <sec id="sec-3-3">
        <title>3.3. Pre-Processing</title>
        <p>Training data sentences are converted to lower case, contractions fixed and punctuations
removed. The NLTK library is used to remove stop words. The Porter Stemmer is used to stem
the words. To conform to fasttext input file requirements, each sentence is rearranged and a
prefix “ __label__” afixed to the start of each class label. The final format for each sentence is
shown in the example below:
__label__Facts none of her children survived her</p>
        <p>For training, data is converted into input text files for training and validation using a ratio of
70/30.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Runs Description</title>
      <p>In our approach we consider the influence of word order in improving performance. We
therefore train a classifier with similar parameters while varying the length of word tokens. The
submitted runs are for the diferent models of the classifier obtained for diferent input tokens.
Each test sentence was pre-processed to lower case, fix contractions, remove stop-words and
also stem the words. The Porter Stemmer is used to stem the words.
For our baseline the fastText classifier model is trained for 25 epochs at a learning rate of 0.5
with WordNgrams set to unigrams.</p>
      <p>In an efort to improve performance, in our second run we use bigrams as our input tokens while
the model’s learning rate and epochs remain at 25 and 0.5 respectively. A slight improvement is
noticed over the baseline in terms of both training accuracy and precision.
4.3. UB_BW RUN 3
In our third run the model’s parameters are retained as per the two previous runs, however
the input tokens are set to trigrams. A negligible improvement is noted in terms of training
accuracy and precision over the second run. Training data results based on Precision, and
performance accuracy (on the training set) are shown in the Table 2 and Table 1.</p>
    </sec>
    <sec id="sec-5">
      <title>5. Results and Analysis</title>
      <p>The performance of our runs relative to other teams systems on the test data is shown in Table 2.
For our baseline system we used unigrams as input tokens while for the second and third
systems bigrams and trigrams were used respectively. It can be observed from the results that
the system that the trigrams based system UB_BW RUN 3 performed much better than the
baseline that used unigrams UB_BW RUN 1 across all measures. However a negligible diference
is noticed between the trigrams and bigrams system UB_BW RUN 2. A category wise analysis
of the results extracted from the results as shown in Table 1 shows all systems performed poorly
in terms of predicting labels for the classes Ruling by Lower Court and Statute. It can also
be noted that for other classes the baseline system and the bigram system outperformed the
trigram system. However the trigram system outperforms the other systems in terms of F-score
for all classes</p>
    </sec>
    <sec id="sec-6">
      <title>6. Discussion and Conclusion</title>
      <p>In this paper we explored the efectiveness of using phrases with the fastText classifier to assign
rhetorical labels to sentences in a court case document. While our systems did not give good
performance overall we believe that with enhancements and more training data the fastText
classifier has potential to benefit the rhetorical labelling task. We also observe performance
with the introduction of bigrams and trigrams in the model which indicates that phrases can
have a positive influence in a classification task. Going forward we aim to further investigate
the influence of phrases in improving text classification by performing empirical evaluation
with various models.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>P.</given-names>
            <surname>Bhattacharya</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Mehta</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Ghosh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Ghosh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Pal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Bhattacharya</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Majumder</surname>
          </string-name>
          ,
          <article-title>Overview of the FIRE 2020 AILA track: Artificial intelligence for legal assistance</article-title>
          , in: P. Mehta,
          <string-name>
            <given-names>T.</given-names>
            <surname>Mandl</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Majumder</surname>
          </string-name>
          , M. Mitra (Eds.), Working Notes of FIRE 2020 -
          <article-title>Forum for Information Retrieval Evaluation, Hyderabad</article-title>
          , India,
          <source>December 16-20</source>
          ,
          <year>2020</year>
          , volume
          <volume>2826</volume>
          <source>of CEUR Workshop Proceedings, CEUR-WS.org</source>
          ,
          <year>2020</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>11</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>V.</given-names>
            <surname>Parikh</surname>
          </string-name>
          , U. Bhattacharya,
          <string-name>
            <given-names>P.</given-names>
            <surname>Mehta</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Bandyopadhyay</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Bhattacharya</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Ghosh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Ghosh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Pal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Bhattacharya</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Majumder</surname>
          </string-name>
          ,
          <article-title>Overview of the third shared task on artificial intelligence for legal assistance at fire 2021</article-title>
          , in: FIRE (Working Notes),
          <year>2021</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>V.</given-names>
            <surname>Parikh</surname>
          </string-name>
          , U. Bhattacharya,
          <string-name>
            <given-names>P.</given-names>
            <surname>Mehta</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Bandyopadhyay</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Bhattacharya</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Ghosh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Ghosh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Pal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Bhattacharya</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Majumder</surname>
          </string-name>
          ,
          <article-title>Fire 2021 aila track: Artificial intelligence for legal assistance</article-title>
          ,
          <source>in: Proceedings of the 13th Forum for Information Retrieval Evaluation</source>
          ,
          <year>2021</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>P.</given-names>
            <surname>Bhattacharya</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Paul</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Ghosh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Ghosh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Wyner</surname>
          </string-name>
          ,
          <article-title>Identification of rhetorical roles of sentences in indian legal judgments</article-title>
          , in: M.
          <string-name>
            <surname>Araszkiewicz</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          <string-name>
            <surname>Rodríguez-Doncel</surname>
          </string-name>
          (Eds.),
          <source>Legal Knowledge and Information Systems - JURIX</source>
          <year>2019</year>
          :
          <article-title>The Thirty-</article-title>
          second Annual Conference, Madrid, Spain,
          <source>December 11-13</source>
          ,
          <year>2019</year>
          , volume
          <volume>322</volume>
          <source>of Frontiers in Artificial Intelligence and Applications</source>
          , IOS Press,
          <year>2019</year>
          , pp.
          <fpage>3</fpage>
          -
          <lpage>12</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>S. B.</given-names>
            <surname>Majumder</surname>
          </string-name>
          ,
          <string-name>
            <surname>D. Das</surname>
          </string-name>
          ,
          <article-title>Rhetorical role labelling for legal judgements using ROBERTA</article-title>
          , in: P. Mehta,
          <string-name>
            <given-names>T.</given-names>
            <surname>Mandl</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Majumder</surname>
          </string-name>
          , M. Mitra (Eds.), Working Notes of FIRE 2020 -
          <article-title>Forum for Information Retrieval Evaluation, Hyderabad</article-title>
          , India,
          <source>December 16-20</source>
          ,
          <year>2020</year>
          , volume
          <volume>2826</volume>
          <source>of CEUR Workshop Proceedings, CEUR-WS.org</source>
          ,
          <year>2020</year>
          , pp.
          <fpage>22</fpage>
          -
          <lpage>25</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>J.</given-names>
            <surname>Gao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Ning</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Han</surname>
          </string-name>
          ,
          <string-name>
            <surname>L</surname>
          </string-name>
          . Kong,
          <string-name>
            <given-names>H.</given-names>
            <surname>Qi</surname>
          </string-name>
          ,
          <article-title>Legal text classification model based on text statistical features and deep semantic features</article-title>
          , in: P. Mehta,
          <string-name>
            <given-names>T.</given-names>
            <surname>Mandl</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Majumder</surname>
          </string-name>
          , M. Mitra (Eds.), Working Notes of FIRE 2020 -
          <article-title>Forum for Information Retrieval Evaluation, Hyderabad</article-title>
          , India,
          <source>December 16-20</source>
          ,
          <year>2020</year>
          , volume
          <volume>2826</volume>
          <source>of CEUR Workshop Proceedings, CEUR-WS.org</source>
          ,
          <year>2020</year>
          , pp.
          <fpage>35</fpage>
          -
          <lpage>41</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>L.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Liu</surname>
          </string-name>
          , Z. Han,
          <article-title>Query revaluation method for legal information retrieval</article-title>
          , in: P. Mehta,
          <string-name>
            <given-names>T.</given-names>
            <surname>Mandl</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Majumder</surname>
          </string-name>
          , M. Mitra (Eds.), Working Notes of FIRE 2020 -
          <article-title>Forum for Information Retrieval Evaluation, Hyderabad</article-title>
          , India,
          <source>December 16-20</source>
          ,
          <year>2020</year>
          , volume
          <volume>2826</volume>
          <source>of CEUR Workshop Proceedings, CEUR-WS.org</source>
          ,
          <year>2020</year>
          , pp.
          <fpage>18</fpage>
          -
          <lpage>21</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>A.</given-names>
            <surname>Joulin</surname>
          </string-name>
          , E. Grave,
          <string-name>
            <given-names>P.</given-names>
            <surname>Bojanowski</surname>
          </string-name>
          , T. Mikolov,
          <article-title>Bag of tricks for eficient text classification</article-title>
          ,
          <source>CoRR abs/1607</source>
          .01759 (
          <year>2016</year>
          ). arXiv:
          <volume>1607</volume>
          .
          <fpage>01759</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>V.</given-names>
            <surname>Zolotov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Kung</surname>
          </string-name>
          ,
          <article-title>Analysis and optimization of fasttext linear text classifier</article-title>
          ,
          <source>CoRR abs/1702</source>
          .05531 (
          <year>2017</year>
          ). arXiv:
          <volume>1702</volume>
          .
          <fpage>05531</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>R.</given-names>
            <surname>Johnson</surname>
          </string-name>
          , T. Zhang,
          <article-title>Efective use of word order for text categorization with convolutional neural networks</article-title>
          ,
          <source>CoRR abs/1412</source>
          .1058 (
          <year>2014</year>
          ). arXiv:
          <volume>1412</volume>
          .
          <fpage>1058</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>C.</given-names>
            <surname>Chang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Masterson</surname>
          </string-name>
          ,
          <article-title>Using word order in political text classification with long short-term memory models</article-title>
          ,
          <source>Political Analysis</source>
          <volume>28</volume>
          (
          <year>2020</year>
          )
          <fpage>395</fpage>
          -
          <lpage>411</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>S.</given-names>
            <surname>Jameel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Lam</surname>
          </string-name>
          , L. Bing,
          <article-title>Supervised topic models with word order structure for document classification and retrieval learning</article-title>
          ,
          <source>Inf. Retr. J</source>
          .
          <volume>18</volume>
          (
          <year>2015</year>
          )
          <fpage>283</fpage>
          -
          <lpage>330</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>