<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>A Machine Learning Approach to Identifying Sections in Legal Briefs</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Scott Vanderbeck and Joseph Bockhorst</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Chad Oldfather</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Dept. of Elec. Eng. and Computer Science, University of Wisconsin - Milwaukee</institution>
          ,
          <addr-line>P.O. Box 784 , 2200 E. Kenwood Blvd., Milwaukee, WI 53201-0784</addr-line>
          ,
          <country country="US">USA</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Marquette University Law School</institution>
          ,
          <addr-line>P.O. Box 1881, Milwaukee, WI 53201-1881</addr-line>
          ,
          <country country="US">USA</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2010</year>
      </pub-date>
      <abstract>
        <p>With an abundance of legal documents now available in electronic format, legal scholars and practitioners are in need of systems able to search and quantify semantic details of these documents. A key challenge facing designers of such systems, however, is that the majority of these documents are natural language streams lacking formal structure or other explicit semantic information. In this research, we describe a two-stage supervised learning approach for automatically identifying section boundaries and types in appellee briefs. Our approach uses learned classifiers in a two-stage process to categorize white-space separated blocks of text. First, we use a binary classifier to predict whether or not a text block is a section header. Next, we classify those blocks predicted to be section headers in the first stage into one of 19 section types. A cross-validation experiment shows our approach has over 90% accuracy on both tasks, and is significantly more accurate than baseline methods.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Introduction</title>
      <p>Now that most of the briefs, opinions and other legal
documents produced by court systems are routinely encoded
electronically and widely available in online databases, there
is interest throughout the legal community for computational
tools that enable more effective use of these resources.
Document retrieval from keyword or Boolean searches are key
tasks that have long been a focus of natural language
processing (NLP) algorithms for the legal domain. However,
the simple whole document word-count representations and
document similarity measures that are typically employed
for retrieval limits their relevance to a relatively narrow set
of tasks. Practicing attorneys and legal academics are
finding that the existing suite of tools fall short of meeting their
growing and complex information needs.</p>
      <p>
        Consider, for example, Empirical Legal Studies (ELS), a
quickly growing area of legal scholarship that aims to
apply quantitative, social-science research methods to
questions of law. ELS research studies are increasingly likely
to have a component that involves computational processing
of large collections of legal documents. One example, are
studies of the role of ideological factors that assign an
ideology value to legal briefs (e.g., conservative or liberal
        <xref ref-type="bibr" rid="ref2">(Evans
et al. 2006)</xref>
        ). One problem that may arise in settings like
this that employ a general similarity measure not tailored to
the task at hand is that documents are more likely to group
by topics, for instance the type of law, than by, say, ideology.
      </p>
      <p>One general technique that has the potential to improve
performance on a wide range of ELS and retrieval tasks is to
vary the influence of different sections of a document. For
example, studies on ideology, may reduce the influence of
content in the “Statement of Facts” section while
increasing the influence of the “Argument” section. However,
although most briefs have similar types of sections, there are
no formal standards for easily extracting them.
Computational techniques are needed. Toward that end, we describe
here a machine learning approach to automatically
identifying sections in legal briefs.</p>
      <sec id="sec-1-1">
        <title>Problem Domain</title>
        <p>Our focus here is on briefs written for appellate court cases
heard by the United States Courts of Appeals. The appeals
process begins when one party to a lawsuit, called the
appellant, asserts that a trial court’s action was defective in one or
more ways by filing an appellant brief. The other party (the
appellee) responds with an appellee brief, arguing why the
trial courts action should stand. In turn, the appeals court
provides its ruling in a written opinion. While there is good
reason to investigate methods for identifying structure in all
three kinds of documents, for simplicity we restrict our focus
here to appellee briefs. We conduct our experiment using a
set of 30 cases heard by the First Circuit in 2004.</p>
        <p>In the federal courts, the Federal Rules of Appellate
Procedure require that appellant briefs include certain
sections, and that appellees include some corresponding
sections while being free to omit others. There is, however, no
standard as to section order or how breaks between sections
are to be indicated. Moreover, parties often fail to adhere to
the requirements of the rules, with the result being that
authors exercise considerable discretion in how they structure
and format the documents.</p>
      </sec>
      <sec id="sec-1-2">
        <title>Related Work</title>
        <p>
          Many genres of text are associated with particular
conventional structures. Automatically determining all of these
types of structures for a large discourse is a difficult and
unsolved problem
          <xref ref-type="bibr" rid="ref7">(Jurafsky &amp; Martin 2000)</xref>
          . Much of the
previous NLP work in the legal domain concerns
Information Retrieval (IR) and the computation of simple features
such as word frequency
          <xref ref-type="bibr" rid="ref4">(Grover et al. 2003)</xref>
          .
        </p>
        <p>
          Additional work has been done in the legal domain with
the focus on summarizing documents. Grover et al.
developed a method for automatically summarizing legal
documents from the British legal system. Their method was
based on a statistical classifier that categorized sentences in
the order that they may be seen as a candidate text excerpt
in a summary
          <xref ref-type="bibr" rid="ref4">(Grover et al. 2003)</xref>
          .
        </p>
        <p>
          <xref ref-type="bibr" rid="ref3">Farzindar and Lapalme (2004)</xref>
          also described a method
for summarizing legal documents. As part of their analysis,
they performed thematic segmentation on the documents.
Finding that more classic method for segmentation
          <xref ref-type="bibr" rid="ref1 ref5">(Hearst
1994; Choi 2000)</xref>
          did not provide satisfactory results, they
developed a segmentation process based on specific
knowledge of their legal documents. For their study groups of
adjacent paragraphs were grouped into blocks of text based
on the presence of section titles, relative position within the
document and linguistic markers.
        </p>
        <p>
          The classic algorithm for topic segmentation is
TextTiling where like sentences and topics are grouped together
          <xref ref-type="bibr" rid="ref6">(Hearst 1997)</xref>
          . More general methods for topic
segmentation of a document are generally based on the cohesiveness
of adjacent sentences. It is possible to build lexical chains
that represent the lexical cohesiveness of adjacent sentences
in a document based on important content terms,
semantically related references, and resolved anaphors
          <xref ref-type="bibr" rid="ref8">(Moens &amp;
De Busser 2001)</xref>
          . Lexical chains and cohesiveness can then
be used to infer the thematic structure of a document.
        </p>
        <p>In contrast to approaches such as these that are based
on inferring the relatedness of sentences in section bodies,
our approach focuses identifying and catagorizing section
headers. These general approaches are complementary as it
would be relatively straightforward to construct a combined
method that considers both headers and bodies.</p>
      </sec>
      <sec id="sec-1-3">
        <title>Overview</title>
        <p>Our analysis begins with a pre-processing step that converts
documents to sequences of text blocks, roughly at the
paragraph level (see below for details). We next construct
feature vector representations for all blocks. Labeled training
sets and supervised learning methods are used to induce two
kinds of classifiers: one for distinguishing section header
blocks from non-header blocks, and one for classifying the
section type of headers. Figure 1 shows a flowchart of the
processing for classifying a block of text in the test set. Note
that although the type of non-header blocks is not predicted
directly, after classifying of all blocks in a document the
predicted section for a non-header block is given by the type of
the nearest preceding section header.</p>
      </sec>
    </sec>
    <sec id="sec-2">
      <title>Models and Methods</title>
      <sec id="sec-2-1">
        <title>Dataset</title>
        <p>Appellee briefs in our dataset are available as HTML files.
The HTML is not well formed or standardized and provides
little insight into the structure of the briefs. The HTML
elements do not contain attributes, block level elements, id’s,
Task I
Task II</p>
        <p>Block of Text 
classes, etc. that may indicate section breaks or section
types. Further, document formatting is inconsistent and
nonstandardized. For example, one author may use italics for
section headings, another bold, while yet another uses inline
text. Formatting sometimes even varies from section to
section within the same document. Thus, we ignore formatting
such as italics or bold, and focus our analysis on the word
and character sequenece.</p>
        <p>Preprocessing was performed on the documents to divide
the documents into blocks of text. A block of text is
essentially a continuous sequence of text from the original
document with a line break immediately before and after. We
extract blocks by converting each HTML document to an XML
document that recognizes all of the line breaks and white
spaces from the original HTML. Examples of document
elements that correspond to blocks extracted from the XML
include paragraphs, section headings, section sub-headings,
footnotes, and table-of-contents entries.</p>
        <p>The XML files were manually reviewed and annotated by
the author (SV). Each block is assigned two class labels:
1. is header A binary value indicating whether or not a
block is a section heading.
2. section type A discrete value that for section headers
only indicates section type. As we only predict the type
of header blocks, the value of “None” is assigned to
nonheaders. Table 1 shows the section types we identified in
our dataset.</p>
      </sec>
      <sec id="sec-2-2">
        <title>Feature Vector Representation</title>
        <p>Along with the two class labels, we represent each block
of text with a 25 element vector of features values.
Ta(a)
Feature Name
leadingAsterisk
leadingNumeral
endsInPeriod
endsInNumeral
stringLength
percentCaps
ellipses
contains(“argument”)
contains(“authori”)
contains(“case”)
contains(“conclusion”)
contains(“contents”)
contains(“corporate”)
contains(“disclosure”)
contains(“fact”)
contains(“issue”)
contains(“jurisdiction”)
contains(“of”)
contains(“prayer”)
contains(“present”)
contains(“review”)
contains(“standard”)
contains(“statement”)
contains(“summary”)
contains(“table”)</p>
        <p>(b)
leadingAsterisk:
endsInPeriod:
stringLength:
percentCaps:
leadingNumeral:
endsInNumeral:
ellipses:</p>
        <p>FALSE
FALSE
21
1
TRUE
FALSE
FALSE</p>
        <p>Domain
binary
binary</p>
      </sec>
      <sec id="sec-2-3">
        <title>Description</title>
        <p>True if the block begins with an asterisk (*)
True if the block begins with an Arabic or Roman numeral
(optionally preceded by an asterisk).</p>
        <p>True if the block ends with a period (.)
True if the block ends with an Arabic or Roman numeral.
Number of characters in the block.</p>
        <p>The % of alpha characters that are capitalized.</p>
        <p>True if the block contains an ellipses (i.e. “...”).</p>
        <p>Each of these features is an indicator for
a specifc string. The feature contains(s) is
true if the block contains a word that begins
with the string s and false otherwise.</p>
        <p>contains(“of”):
contains(“table”:
contains(“contents”:
(all other string match features):</p>
        <sec id="sec-2-3-1">
          <title>TRUE TRUE TRUE FALSE</title>
          <p>is header:
section type:</p>
        </sec>
        <sec id="sec-2-3-2">
          <title>TRUE Table of Contents ble 2(a) lists the features we use, Table 2(b) shows the feature and class values for the block of text “II. TABLE OF</title>
          <p>CONTENTS”.</p>
          <p>The features chosen were engineered through visual
inspection of section headings, intuition, and trial and error.
Other attributes were considered such as the length and
percentage of capital letters of the previous and next blocks of
text, however, these did not improve model performance.
The group of features named contains(s) are string
matching features, which are true if the block of text contains
exactly one word that begins with the string s. We construct a
string match feature from all words that occur five or more
times in the 252 header blocks.</p>
        </sec>
      </sec>
      <sec id="sec-2-4">
        <title>Learning</title>
        <p>The task of identifying section headers and the type of
section is divided into two steps (Figure 1). The first step
classifies a block of text as either a section heading or not a
section heading. For this task, supervised machine learning
algorithms are used to learn a binary classifier. The second
task takes each block of text classified in the first step as a
heading and uses a second classifier to predict the specific
type of section. Again supervised machine learning is used
to learn a classifier, this time with 19 classes. For both tasks,
multiple types of classifiers including naive Bayes, logistic
regression, decision trees, support vector machines and
neural networks were considered.</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Evaluation</title>
      <p>With the abundance of legal documents available, it is
important that they be structured in ways usable by computers
(Wynera 2010). We hypothesize the task of structuring our
legal documents into relevant sections can be accomplished
with a supervised machine learning classifier that first
identifies section headers, and then assigns a section type to the
header.</p>
      <p>To test this hypothesis we have conducted an experiment
on 30 appellee briefs from cases heard by the US 1st
Circuit in 2004. No effort was made to restrict the cases to
a particular area of the law, and indeed a variety of
different types of cases is represented in this set. The
legal briefs were obtained as HTML files through WestLaw
(www.westlaw.com). In the 30 documents, a total of 252
section headers were identified. Note that subsection
headers are not included as part of this task as there is very little
commonality in authors use of subsections. Additionally,
subsections are generally specific to the legal case being
addressed, and not the overall document. Of the 252 total
section headers, 116 unique strings were identified (not
accounting for any difference in formatting or upper / lower
case). Manual inspection of the 116 variations revealed that
the headers cluster into the 19 different section types listed
in Table 2(b). A 20th section type “None” was added to be
used as the class label for blocks of text that do not represent
section headers.</p>
      <p>We conducted a leave-one-case-out cross-validation
experiment. That is, in each experiment all blocks from one
of our documents was held out of the training set and used
as test data to estimate our models’ ability to generalize to
unseen documents.</p>
      <p>For the first task, all blocks of text in the training set are
used. For the second task, only training set blocks of text
labeled as section headings are used for training. This
decision was made because we only wish to use the second
classifier to label the section type of true section headers. Also,
this approach sidesteps the inconsistency that arises when
a block of text is identified as a heading in the first stage,
but as section type “None” in the second stage. We may
revisit this decision in future work as a “None” prediction in
stage two could potentially be used to catch false positives
from the first stage. With the current dataset, however, it
was found that the number of correctly identified headings
being labeled as “None” vs. the correction of false positives
was not worth the tradeoff. Therefore, we take the approach
described above.</p>
      <p>We evaluate models on the first task by the percentage
of headings or non-headings correctly classified as well as
precision and recall rates where:
precision =</p>
      <p>#true positives
#true positives + #f alse positives
and
recall =</p>
      <p>#true positives
#true positives + #f alse negatives
Note blocks of text that are a section header represent our
positive class. Precision and recall are both of particular
importance for our first task. Examining our dataset, 95.4% of
blocks of text are non-headings. The extreme case of
classifying all blocks of text as non-headings would then result
in very high overall accuracy and 100% recall rate for
nonheadings, at the expense of poor precision.</p>
      <p>We compare our machine learning approach to a regular
expression baseline. The regular expression used for this
baseline approach may be summarized as the concatenation
of the following list of parts:</p>
      <sec id="sec-3-1">
        <title>1. The beginning of the string</title>
      </sec>
      <sec id="sec-3-2">
        <title>2. An optional asterisk</title>
        <p>3. An optional Roman Numeral or Natural Number followed
by an optional period and space
4. A list of zero or more all capitalized words</p>
      </sec>
      <sec id="sec-3-3">
        <title>5. The end of the string</title>
        <p>Blocks that contain a match to the regular expression are
predicted to be headers. This regular expression should
correctly identify many section headings as many are entirely
capitalized, while excluding false positives such as table of
contents entries that are generally followed by a page or
section number of some form.</p>
        <p>Our second task is then evaluated in two ways. The first
is the overall percentage of predicted headings that are
assigned the correct section heading type. The second metric
is an adjusted metric that does not penalize the second task
for errors made in the first task. If the input to the second
classification task was a non-heading to begin with, this
classifier would inherently fail as it is attempting to determine
the section heading type when no such type actually exists.
Therefore, we account for this disparity in our results and
also present the number of section heading types predicted
correctly divided by actual headings correctly classified by
the first task.</p>
        <p>A baseline approach is only considered for the first task
of identifying whether or not a block of text is a section
heading. A baseline approach for the secondary task of
assigning a label of one of our 20 classes could be developed
through a complicated regular expression or a form of
sequential logic, but was not considered in this project. Our
most frequent section heading type, “Argument”, accounts
for 12% of cases. Therefore, that level of accuracy could be
achieved by simply always predicting “Argument”.</p>
        <p>Last, a combined metric is presented where we merged
the results from both steps of classification to determine
the overall percentage of section headings that are correctly
identified and assigned the correct type.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Results</title>
      <sec id="sec-4-1">
        <title>Task 1 - Identifying Section Headings</title>
        <p>A total of 5,442 blocks of text were identified in our dataset.
Table 3 shows a comparison of the baseline method with our
supervised machine learning based approach for the task of
identifying if a block of text is a section heading or not. With
the exception of naive Bayes (which performed worse), all
other classifiers performed similarly.</p>
      </sec>
      <sec id="sec-4-2">
        <title>Total Blocks of Text:</title>
      </sec>
      <sec id="sec-4-3">
        <title>Correctly Classified:</title>
      </sec>
      <sec id="sec-4-4">
        <title>Percentage Correct:</title>
      </sec>
      <sec id="sec-4-5">
        <title>Baseline</title>
        <p>5442
5288
97.2%</p>
      </sec>
      <sec id="sec-4-6">
        <title>Learning</title>
      </sec>
      <sec id="sec-4-7">
        <title>Based</title>
        <p>5442
5409
99.4%</p>
        <p>As expected the baseline approach performed very well
with 97.2% accuracy. This, represents a small gain over
calling all blocks non-headings (95.4%). As we
hypothesized, the learning based classifier performed much better
with 99.4% accuracy. As seen in the confusion matrix in
Table 4, the logistic regression classifier had a similar
number of false positives and false negatives. Precision and
recall statistics are presented in Table 5. As seen in the table,
there is a significant difference in the recall rates of
headings (92.1% vs. 61.5%) which is of great importance to the
ultimate goal.</p>
        <p>Examining incorrectly classified blocks, the most
frequent was “Standard of Review” and accounted for 24% of
all errors. Examination of this reveals that the “Standard of
Review” is often included as a subsection of the “Argument”
section of the brief by many authors, while others choose to
make a standalone section. For example, the block of text “1.
STANDARD OF REVIEW” was incorrectly classified as a
heading in one instance. In this case the author did not use
a numbering scheme for the primary section (“Argument” in
this case), but numbered the sub-sections on the document
confusing our model. Similar errors occurred for the section
type “Statement of Facts” and accounted for 12% of all
errors.’ With additional post processing of the classification, it
may be possible to account for these types of errors further
increasing model performance.</p>
      </sec>
      <sec id="sec-4-8">
        <title>Task 2 - Predicting Section Type</title>
        <p>Table 6 summarizes the result of the secondary classifier that
assigns section types to any block of text classified as a
heading by the first task. The first task identified 245 blocks of
text as headings. Of these, only 18 were assigned an
incorrect section heading type for an overall accuracy of 92.7%.
However, 13 of these 18 were not actually classes to begin
with so the secondary classifier could not have assigned a
correct class label. Adjusting for this, 232 blocks of text
were correctly identified as headings and of these only 5
were given an incorrect label for an adjusted 97.8%
accuracy.</p>
      </sec>
      <sec id="sec-4-9">
        <title>Total Headings</title>
      </sec>
      <sec id="sec-4-10">
        <title>Identified</title>
      </sec>
      <sec id="sec-4-11">
        <title>Actual Headings</title>
      </sec>
      <sec id="sec-4-12">
        <title>Identified</title>
      </sec>
      <sec id="sec-4-13">
        <title>Count</title>
        <p>245
232</p>
      </sec>
      <sec id="sec-4-14">
        <title>Correctly</title>
      </sec>
      <sec id="sec-4-15">
        <title>Labeled</title>
        <p>227
227</p>
      </sec>
      <sec id="sec-4-16">
        <title>Percent</title>
      </sec>
      <sec id="sec-4-17">
        <title>Correct</title>
        <p>92.7%
97.8%
We presented a supervised machine learning approach for
structuring legal documents into relevant sections. Our
approach is based on two steps. The first step identifies blocks
of text that are section headings. In the second step, blocks
of text classified as section headings are then input into a
second step to predict section type.</p>
        <p>We evaluated our approach with a cross-validation
experiment. The first task of identifying section headers using a
binary logistic regression classifier was shown to perform
with 99.4% accuracy. The secondary task is then used with
92.7% accuracy to determine the type of section one is
looking at. The NLP approach provides a 2.2% improvement
in accuracy over the baseline regular expression based
approach, and more importantly provides a significantly higher
recall rate in identifying section headings vs. non section
headings.</p>
        <p>While it may be possible to create a non-learning based
approach (more complex than the baseline approach
presented) to perform the given subtask, it has been shown that
a machine learning and NLP approach are very well suited
for this problem. This paper only researched appellee briefs,
but there is ample reason to believe that this approach would
provide similar results for appellant briefs, the judges
written opinion, and other similar documents.</p>
        <p>The significance of our learned models having
significantly higher recall rates than baseline models becomes of
even greater importance when one considers that approaches
would be available to correct or account for false positives
(i.e. non-headings classified as headings), however, it would
be far more difficult, if even possible, to correct for false
negatives (i.e. actual headings classified as non-headings).</p>
        <p>While not formally discussed in this paper, it is possible
to implement secondary logic to correct for some of the
classification errors we encountered. For instance, our most
frequent error in the first task was the “Standard of Review.
Logic could be implemented as a post processing step that
says if a block of text is called a section heading and
classified with the section heading type “Standard of Review, but
is preceded by the section type “Argument, remove this as a
section heading. In our dataset this correction would correct
5 of 7 mistakes made labeling “Standard of Review“ and
improve accuracy for the first task to 99.5% and 94.6% for the
second task.</p>
        <p>In addition, allowing the secondary classifier to identify
sections that it assigns the class label “None could correct
some false positives incorrectly classified as section
headings by the first task. In our dataset, 4 such corrections could
have been made further improving accuracy. However, if
implementing this change one must consider the
implications of giving an actual section break heading the section
type “None versus the improvement from corrections.</p>
        <p>We considered 20 different potential class labels for each
section. For specific tasks it may be found that this number
can be reduced to even as few as two (i.e. relevant or
nonrelevant) sections. This could be done as part of the
classification or as part of a post process mapping the classifications
output by the classifier to a smaller groups of classes for the
ultimate task. This may potentially further improve overall
performance.</p>
        <p>In our approach, the secondary task was treated as
individual classifications. It may be possible to treat the
secondary classification problem as a Hidden Markov Model or
Continuous Random Field. Doing so may improve
performance as when an author does include a section in his/her
legal briefs, they are generally in a consistent order.</p>
        <p>Last, the majority of misclassifications in both tasks
appears to be the result of sparse data and infrequently used
section headings. While learning curves were not created,
it is suspected that additional data could provide the
classifier with information about many these sections and improve
overall model performance.</p>
        <p>With the current model, and the potential for further future
improvements, section related information can reliably be
identified with supervised machine learning based methods
in poorly structured legal documents.</p>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          <string-name>
            <surname>Choi</surname>
            ,
            <given-names>F. Y. Y.</given-names>
          </string-name>
          <year>2000</year>
          .
          <article-title>Advances in Domain-Independent Linear Text Segmentation</article-title>
          .
          <source>In Proc. Conference of the North American Chapter of the Association for Computational Linguistics (NAACL)</source>
          ,
          <fpage>26</fpage>
          -
          <lpage>33</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          <string-name>
            <surname>Evans</surname>
            ,
            <given-names>M. C.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>McIntosh</surname>
            ,
            <given-names>W. V.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Lin</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ; and
          <string-name>
            <surname>Cates</surname>
            ,
            <given-names>C. L.</given-names>
          </string-name>
          <year>2006</year>
          .
          <article-title>Recounting the Courts? Applying Automated Content Analysis to Enhance Empirical Legal Research</article-title>
          . SSRN eLibrary.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          <string-name>
            <surname>Farzindar</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Lapalme</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          <year>2004</year>
          .
          <article-title>Legal text summarization by exploration of the thematic structures and argumentative roles</article-title>
          .
          <source>In In Text Summarization Branches Out Conference held in conjunction with ACL</source>
          <year>2004</year>
          ,
          <volume>27</volume>
          -
          <fpage>38</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          <string-name>
            <surname>Grover</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Hachey</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Hughson</surname>
            ,
            <given-names>I.;</given-names>
          </string-name>
          and
          <string-name>
            <surname>Korycinski</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <year>2003</year>
          .
          <article-title>Automatic summarisation of legal documents</article-title>
          .
          <source>In Proceedings of the 9th international conference on Artificial intelligence and law</source>
          , ICAIL '
          <volume>03</volume>
          ,
          <fpage>243</fpage>
          -
          <lpage>251</lpage>
          . New York, NY, USA: ACM.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          <string-name>
            <surname>Hearst</surname>
            ,
            <given-names>M. A.</given-names>
          </string-name>
          <year>1994</year>
          .
          <article-title>Multi-paragraph segmentation of expository text</article-title>
          .
          <source>In Proceedings of the 32nd annual meeting on Association for Computational Linguistics</source>
          , ACL '
          <volume>94</volume>
          ,
          <fpage>9</fpage>
          -
          <lpage>16</lpage>
          . Stroudsburg, PA, USA: Association for Computational Linguistics.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          <string-name>
            <surname>Hearst</surname>
            ,
            <given-names>M. A.</given-names>
          </string-name>
          <year>1997</year>
          .
          <article-title>TextTiling: Segmenting text into multi-paragraph subtopic passages</article-title>
          .
          <source>Computational Linguistics</source>
          <volume>23</volume>
          (
          <issue>1</issue>
          ):
          <fpage>33</fpage>
          -
          <lpage>64</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          <string-name>
            <surname>Jurafsky</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Martin</surname>
            ,
            <given-names>J. H.</given-names>
          </string-name>
          <year>2000</year>
          .
          <article-title>Speech and Language Processing: An Introduction to Natural Language Processing</article-title>
          ,
          <source>Computational Linguistics and Speech Recognition (Prentice Hall Series in Artificial Intelligence)</source>
          .
          <source>Prentice Hall</source>
          ,
          <volume>1</volume>
          <fpage>edition</fpage>
          . neue
          <source>Auflage kommt im Frhjahr</source>
          <year>2008</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          <string-name>
            <surname>Moens</surname>
          </string-name>
          ,
          <string-name>
            <surname>M.-F.</surname>
            , and De Busser,
            <given-names>R.</given-names>
          </string-name>
          <year>2001</year>
          .
          <article-title>Generic topic segmentation of document texts</article-title>
          .
          <source>In Proceedings of the 24th annual international ACM SIGIR conference on Research and development in information retrieval</source>
          ,
          <source>SIGIR '01</source>
          ,
          <fpage>418</fpage>
          -
          <lpage>419</lpage>
          . New York, NY, USA: ACM.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>