<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>What Are the Facts? Automated Extraction of Court-Established Facts from Criminal-Court Opinions</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Klára Bendová</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Tomáš Knap</string-name>
          <email>tomas.knap@prf.cuni.cz</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Jan Černý</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Vojtěch Pour</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Jaromir Savelka</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
          <xref ref-type="aff" rid="aff3">3</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Ivana Kvapilíková</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Jakub Drápal</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>6 June</institution>
          ,
          <addr-line>2025, Chicago</addr-line>
          ,
          <country country="US">USA</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Faculty of Mathematics and Physics, Charles University</institution>
          ,
          <addr-line>Prague, Czechia</addr-line>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Proceedings of the Seventh International Workshop on Automated Semantic Analysis of Information in Legal Text</institution>
          ,
          <addr-line>ASAIL 2025</addr-line>
        </aff>
        <aff id="aff3">
          <label>3</label>
          <institution>School of Computer Science, Carnegie Mellon University</institution>
          ,
          <addr-line>Pittsburgh PA</addr-line>
          ,
          <country country="US">USA</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2026</year>
      </pub-date>
      <abstract>
        <p>Criminal justice administrative data contain only a limited amount of information about the committed ofense. There is, however, an unused source of extensive information in continental European courts' decisions: Descriptions of criminal behaviors in verdicts by which ofenders are found guilty. In this paper, we study the feasibility of extraction of these descriptions from publicly available court decisions from Slovakia. We use two diferent approaches for retrieval: regular expressions and large language models (LLM). Our baseline was a simple method employing regular expressions to identify typical words occurring before the beginning of the description and after. The advanced regular expressions approach further focused on letter-spacing and its normalization (insertion of spaces between individual letters), typical for delineation of the description. The LLM approach involved prompting the Gemini Flash 2.0 model to extract the descriptions using a predefined set of instructions. Although the baseline identified descriptions in only 40.5% of the verdicts of our test set, both methods significantly outperformed it, achieving 97% with advanced regular expressions and 98.75% with LLM and a combination of both 99.5%. Evaluation by law students showed that both advanced methods matched human annotations in about 90% of cases, compared to just 34.5% for the baseline. LLMs fully matched human-labeled descriptions in 91.75% of instances and a combination of advanced regular expression with LLM brings 92% match.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;argument mining</kwd>
        <kwd>court decisions</kwd>
        <kwd>criminal behavior</kwd>
        <kwd>NLP</kwd>
        <kwd>LLM</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>Empirical quantitative legal research is often hampered by insuficient detail in data, particularly in
criminal law, where information about criminal behavior is frequently incomplete or lacking.
Administrative datasets typically record only the relevant section of the penal code, ofering a general definition
of the ofense. However, they usually provide little information beyond this classification, making it
dificult for researchers to discern the specific behaviors involved and to understand how variations in
behavior shape state responses.</p>
      <p>More detailed information about criminal behavior is recorded in textual form. Criminal verdicts in
most continental European countries contain a description of the criminal behavior: An authoritative
description of what behavior an ofender is found guilty of and for which a sentence is imposed. These
descriptions thus provide crucial and otherwise unavailable pieces of information about behavior.
Criminal verdicts are available online in several European and non-European countries (e. g., Slovakia,
Estonia, Moldova and China), while in others they are available to researchers (e. g., Finland).</p>
      <p>In this paper, we show the dificulties of extracting these descriptions of criminal behaviors from
court verdicts. After discussing the related work and the data we employed, we present two diferent
extraction methods and their specifics. Then, we present the outcomes of individual methods (and their
combinations) and their reliability. We conclude by presenting how we worked with dificult cases and
the future use of LLMs in perfecting this task.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related Work</title>
      <p>
        Extracting various semantic and/or functional elements from court opinions has been established as
a key task in legal text processing. This is because these loosely structured and sometimes noisy
documents contain enormous amounts of useful knowledge that can potentially be utilized in many
diferent applications. Prior research can be distinguished into two categories. First, the task may be
defined as labeling small textual units, often sentences, according to some predefined type system (e.
g., rhetorical roles such as evidence, reasoning, and conclusion). This approach has been applied to
administrative decisions from the U.S. [
        <xref ref-type="bibr" rid="ref1 ref2">1, 2</xref>
        ], multi-domain court decisions from India [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ], international
arbitration decisions [
        <xref ref-type="bibr" rid="ref13 ref4">4</xref>
        ], multi-{domain,country} adjudicatory decisions [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ], or opinions of the European
Court of Human Rights [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. Identifying a section that states an outcome of the case has also received
considerable attention separately [
        <xref ref-type="bibr" rid="ref10 ref7 ref8 ref9">7, 8, 9, 10</xref>
        ]. In this work, we focus on detecting one specific sentence
in each opinion—the one that authoritatively states the facts of the case.
      </p>
      <p>
        Alternatively, the task could be to segment the text into a small number of contiguous parts, typically
comprising multiple paragraphs. Diferent variations of this task were applied to several legal domains
from countries such as Canada [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ], the Czech Republic [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ], France [13], and the U.S. [14]. One
approach to segmentation has focused on automatically identifying the rhetorical roles of sentences
[
        <xref ref-type="bibr" rid="ref11">11, 15, 16</xref>
        ]. In [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] the authors employed linguistic markers to segment Canadian decisions into four
units: Introduction, Context, Juridical Analysis, and Conclusion. A similar scheme was proposed in [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ],
including some additional types such as Dissent, Footnotes, or Party Claims. In [17], the authors identify
typical language structures that are used in various types of Premises or Conclusions. These are then
expressed in the form of a Context Free Grammar for parsing legal arguments. In [18, 19], conditional
random fields (CRF) were applied to segment legal documents into seven labeled components, with
each label representing a corresponding rhetorical role.
      </p>
    </sec>
    <sec id="sec-3">
      <title>3. Data</title>
      <p>Slovak courts’ verdicts are, in many respects, typical of the continental Germanic legal culture. They
are clearly divided into two parts: the dispositive part, which presents the decisions that were taken,
and the reasoning. The dispositive part (i) identifies the case, the court, the verdict, and the ofender; (ii)
announces whether the ofender was found guilty or not; (iii) describes the criminal behavior for which
the ofender was found guilty (or innocent); (iv) subsumes the behavior within a legal definition of an
ofense; and (v) pronounces the consequences (sentences and reparation of damages). An example of
this part is in Appendix B. Some verdicts further contain reasoning—an explanation of why the ofender
was found guilty, why a specific sentence was imposed, and a description of the proceedings. Reasoning
is often infrequent and of low quality [20].</p>
      <p>Slovakia allows anyone to download .json files with all court decisions and to use an API to search
for specific decisions. While such convenient accessibility is very beneficial, researchers should be
concerned about possible flaws in the published data, especially missing data and the possible resulting
limited representativeness. If an administrative dataset containing secondary data about court decisions
made in a country is available, it helps researchers identify both how many court verdicts are missing
and whether there is a pattern among the missing values. Slovakia has an administrative dataset of high
quality containing secondary data describing all criminal court verdicts from 2018 to 2022 (hereafter
referred to as the "administrative dataset") [21]. There were 126,795 verdicts decided during this period,
according to this dataset.</p>
      <p>We first downloaded all court verdicts and linked the verdicts with the administrative dataset using a
unique court docket number and court name. We were able to link 77.64% of cases. Then we employed
the court docket numbers from the administrative dataset to retrieve the missing verdicts via the API,
which allowed us to link an additional 12.42% of cases. Overall, we were unable to link 9.94% of cases
that should exist; yet, they are (i) not present in the .json or API, (ii) present in .json but matching fails
due to a corrupt .json file, or (iii) present in the API but matching fails due to repeated requests and
website limitations. Successful matching, to some extent, depended on which court decided the case.
While 40 out of 54 district courts had a success rate higher than 90%, other courts had a success rate
above 60%, with the exception of one that had a success rate of 12%. These diferences present the need
to work with administrative datasets to ensure that full-text court verdicts are representative and to
determine their limitations. This provides us with 112,864 court verdicts, within which we attempted to
identify descriptions of criminal behavior.</p>
      <p>We chose two test sets of verdicts to annotate the data. The first general stratified sample contains 400
judgments, in which we controlled for years and the representation of diferent courts during sampling.
These were then annotated by trained law students (2 groups of 200 statements), with each statement
being annotated by two persons (an agreement of 97% in both groups; disagreements were resolved
by one of the authors). This dataset is intended to evaluate the overall performance of the individual
methods. Since we expected high performance on the general dataset, we also created an additional
set of 200 judgments to evaluate more challenging cases. This supplementary set includes a mix of
judgments in which the letter-spacing extraction yielded only a single candidate expression—indicating
a higher likelihood that the judgment lacks a factual sentence—and judgments in which the rule-based
extraction method failed to identify any relevant sentence at all. The data were annotated under the
same conditions (95.5% agreement; disagreements were resolved by one of the authors).</p>
    </sec>
    <sec id="sec-4">
      <title>4. Experimental Design</title>
      <sec id="sec-4-1">
        <title>4.1. Structure of Court Decision</title>
        <p>Agree. Purpose
97% overall performance
95.5% stress-test edge cases
Our analysis leveraged two fundamental formal features observable across all judgments: A consistent
structure (described above) and the systematic use of a specific typographic convention throughout the
dataset, specifically letter-spacing. The letters of a word are spaced out (e. g., L I K E T H I S).
Letterspacing is generally applied to single words or short phrases and is typically placed between paragraphs.
Its function is to introduce a new section of the judgment. As a result, letter-spacing tends to appear
with a limited set of expressions used across decisions, as we can see in Appendix B. A key advantage of
letter-spacing is its robustness during format conversion: even when judgments are somewhat messily
transformed from PDF to .json, letter-spacing is usually preserved—unlike paragraph-ending characters,
which often sufer from noisy encoding. However, letter-spacing has a domain-specific nature; while
common in legal documents, it is rarely encountered in other textual domains.</p>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. Baseline</title>
        <p>As we describe in Section 4.1, both the overall structure and the letter-spacing expression between
paragraphs appear to be consistent at first glance. This observation motivated us to extract factual
statements using regular expressions due to their simplicity, accuracy, and ease of use. In the baseline
approach, we chose not to introduce any complexity into the regex patterns. We manually listed starting
and ending phrases as fixed patterns (examples of phrases are in Table 3) and used them for extraction.
The baseline achieves a fact sentence extraction success rate of 31.69%.</p>
        <p>Ending Expression
therefore/thus</p>
      </sec>
      <sec id="sec-4-3">
        <title>4.3. Advanced Regular Expressions</title>
        <p>Building on observations of simple regular expressions, we developed more flexible patterns to eliminate
variations in the input text. These flexible patterns focus on preventing unwanted irregularities—such as
extra spaces, unexpected line breaks, and other conversion issues. Adding optional whitespace matching
between characters in key phrases significantly improved resiliency to formatting inconsistencies. 1
This improved both the success rate and the quality of the extracted fact sentences, as shown in Table 5.</p>
        <p>The regular expressions themselves were automatically extracted directly from the court verdicts. We
focused only on preserving the order of the letter-spaced expressions from the rulings in the document.
We then grouped expressions by their relative position and annotated those that function as openers
or closers of factual sentences. This yielded a set of starting ( = 40) and ending ( = 2) points to
identify fact sentences.</p>
      </sec>
      <sec id="sec-4-4">
        <title>4.4. LLMs: Standalone and Building on Previous Methods</title>
        <p>During the initial phases of our research, relying solely on a rule-based approach for extracting factual
statements proved insuficient. We have made a natural progression to using more sophisticated
approaches. The main challenge with defining a narrow scope of rules, especially in a language where
the individual words are typically more varied, is that one can hardly create a complete list of all the
possible words and their combinations. LLMs are an ideal solution for extending a narrow list of rules
to a dataset that presents uncertainties about the exact wording in each individual case.</p>
        <p>Utilizing LLMs involves two main aspects: selecting an appropriate model and providing it with the
necessary context, a process known as prompting. Our model selection was guided by two primary
criteria: its established suitability for relevant applications such as parsing and extracting from large
texts, and overall cost. Based on this evaluation, Gemini Flash 2.02 was identified as the most suitable
option. Our temperature was 0.0, which helps reduce hallucinations in extracting factual sentences
from judgments and supports better replicability of our research. Our initial experiments revealed that
generic descriptions of factual statements were insuficient for reliable extraction using LLMs. Achieving
success requires a carefully constructed prompt specifically designed for this task. We developed a
specialized prompt that combines concrete examples of factual statements, a clear definition of the
expected output, and, importantly, explicit indicators of where these statements typically begin and
end (Appendix A).</p>
        <p>The critical breakthrough in our approach occurred when we followed the phrases used by previous
methods. We incorporated specific textual markers commonly found at the beginning or end of factual
statements, as shown in Table 3. This structural guidance transformed the task from pure content
understanding to a more focused pattern recognition and extraction challenge. The model was directed
to identify text between these markers and extract only the factual components while omitting legal
evaluations. The results were returned in .json format for further processing.</p>
        <p>A single Slovak judgment contains on average 4,083 characters (approx. 1,020 tokens), to which we
prepend a 10,497-character prompt (approx. 2,624 tokens). The Gemini Flash 2.0 output averages 1 257
characters (approx. 314 tokens). At Google’s July 2025 pricing (USD 0.10 / M input tokens, 0.40 / M
output tokens), this corresponds to 0.00049 USD per judgment (approx. 0.49 USD for 1,000 decisions).</p>
        <p>Despite recent improvements in LLM accuracy, we observed occasional instances of text
hallucination, particularly with special characters or uncommon legal terminology. To address this issue, we
implemented a post-processing step using a function that aligns the model’s output with the original
text. This verification method efectively eliminated hallucinations, ensuring fidelity to the source
documents.</p>
      </sec>
      <sec id="sec-4-5">
        <title>4.5. Combination Approach</title>
        <p>Our experimental design includes a combined methodology that integrates the strengths of both
advanced regular expressions and LLM approaches. This hybrid pipeline operates sequentially:
1) Advanced regular expressions first attempt to extract factual statements from court decisions. 2)
For cases where the regular expression approach fails to identify any factual sentences, we apply the
LLM-based extraction.</p>
        <p>This combination strategy maximizes eficiency while addressing the limitations of each individual
method. The regular expressions provide fast processing for standard document structures, while the
LLM handles more complex linguistic variations and non-standard document formats.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Results</title>
      <p>To empirically assess our approach, we tested the ability of the method to extract specific factual
sentences against a set of 400 reference statements.</p>
      <p>First, we assess the method’s ability to identify a factual statement in the court decision. While
the baseline identified a description of criminal behavior in only 40.5% of verdicts, both advanced
regular expressions and LLMs performed significantly better, achieving success rates of 97% and 98.75%,
respectively. If we combine both approaches, specifically by employing LLMs only on cases where
advanced regular expressions fail to extract any factual sentences, we achieve an improved overall
accuracy of 99.5%. The reason the LLM did not achieve the extraction of 100% factual statements is
that five cases could not be processed due to exceeding the model token limits or triggering content
safety flags within the API, likely related to sensitive topics such as the narcotics mentioned in the
cases. Feeding the texts to the model in pieces and employing prompt engineering to bypass the safety
mechanism would likely further increase performance.</p>
      <p>We further focused on the quality of the extracted sentences. Extraction quality was measured
at the character level, ignoring diacritics. The results are displayed in Table 5 by the quality of the
match. When evaluating based on exact match, the performance of the LLM is undeniable. However,
if we allow for minor character-level variations of up to 5%, the results of advanced regex and LLMs
are surprisingly comparable—89.5% for the advanced regex and 91.75% for the LLM. In contrast, the
baseline approach achieved this level of approximate matching in only 34.5% of cases. By combining
both approaches—specifically applying the LLM only to cases where advanced regex fails to extract
accurate factual sentences—we achieve 92% accuracy on the test data.</p>
      <p>We then focused on problematic cases, i.e., verdicts where no factual statement was extracted or
where the extracted statement shared less than 50% character overlap with the manually annotated
version. Such cases typically result from non-standard typographic formatting of the verdict or the use
of less frequent terms in letter-spaced expressions. In some instances, the extraction failed because the
factual statement was entirely missing from the verdict, which is usually due to their procedural nature
(e. g., acquittal judgments). We prepared a dataset of 200 such verdicts exhibiting these deviations and
had it annotated by two annotators under the same conditions as the main test dataset. The same LLM
prompt was then applied to this data.</p>
      <p>As shown in Table 7, the LLM was able to perform high-quality factual statement extraction in 84%
of cases within this challenging dataset, and it failed to identify a factual statement in 12% of cases,
resulting in 0% similarity. Investigation confirmed that, in nearly all these instances, the LLM correctly
followed its strict prompt instructions. The failures occurred because the factual sentences in this
challenging subset used grammatical structures or formats not anticipated by the specific rules in the
prompt. The LLM, therefore, acted as instructed by reporting no match rather than misinterpreting
the content. This highlights the need to update the prompt, generalizing the description of a factual
sentence based on the wider variety of formats observed.</p>
      <p>To analyze hallucinations, we compared 100 LLM-generated factual statements with the corresponding
spans in the source judgments. Character-level similarity averaged 0.99; only one case (1 %) fell below
90 %.</p>
      <p>To avoid leaking hallucinated text into downstream data, we post-process every model output: we
locate the predicted span in the source document and replace the generated string with that exact
substring. The factual sentences stored in our dataset are therefore verbatim excerpts from the original
judgments; hallucinations appear only in cases where no suficiently similar span can be aligned.</p>
    </sec>
    <sec id="sec-6">
      <title>6. Discussion: Integrating all Methods</title>
      <p>We did not train a supervised model due to the high cost of large-scale annotations. Our strategy was to
ifrst explore few-shot prompting of an LLM, measure its performance against manually annotated data,
and consider supervised training only if the LLM fell short. A small experiment with training a model
on the annotated test set could be interesting; however, it would leave us without a clean held-out set
for evaluation, making it dificult to fairly compare the approaches.</p>
      <p>Our goal was to develop a scalable solution that could be applied across a wide range of court decisions
without the need for extensive labeled training data. To that end, we combined regular expressions
with few-shot prompting of an LLM. This approach allows us to inject a small number of annotated
examples directly into the prompt, guiding the model’s output in a flexible and easily adjustable way.
By avoiding model training and relying instead on prompt-level supervision and rule-based heuristics,
we were able to build a system that is applicable in low-resource legal settings and avoids the high cost
of large-scale manual annotations.</p>
      <p>The LLM-based approach introduces computational demands that must be balanced against
extraction quality. While more resource-intensive than simpler methods, the accuracy benefits justify
their application, particularly for complex or non-standard documents. For large-scale processing of
thousands of verdicts, we recommend a staged approach in which computationally expensive LLM
processing is reserved for documents where simpler methods yield low-confidence results.</p>
      <p>It is worth noting that a recurring pattern observed in cases with imperfect matches was the LLMs’
occasional dificulty in precisely delineating the boundaries of the target factual statement within the
broader text. Specifically, mismatches often arose not from extracting incorrect information; rather,
they stemmed from the model including extraneous text immediately following the intended end of the
factual statement or, in some other cases, the LLM finished the factual sentence before it was instructed
to do so.</p>
      <p>These detailed results afirm the LLMs’ precision, particularly their high success rate, but also
highlight challenges with edge cases and computational intensity, which require careful consideration
for scalability. Future enhancements include employing models with larger context windows to handle
longer texts without truncation and refining prompts for greater task specificity to further improve
accuracy and relevance, particularly for the less successful cases.</p>
      <p>Finally, transferring the methods to other jurisdictions also remains a challenge. Each new legal
system would require its own regex inventory, reflecting local idioms and citation habits. The LLM
approach proved more robust in our pilots, but our sample is too narrow to draw general conclusions
about out-of-domain performance without retraining and expert validation. We therefore refrain from
reporting cross-jurisdiction metrics and leave systematic evaluation on foreign corpora for future work.</p>
    </sec>
    <sec id="sec-7">
      <title>7. Conclusion</title>
      <p>In this paper, we presented a method for collecting published court decisions and extracting factual
sentences—coherent descriptions of criminal conduct—with strong potential for further research. In the
extraction process, we identified the consistent structure of court decisions and a legal typographic
convention known as “letter-spacing” as key features. We leveraged these observations in three
extraction approaches: an automated search for the start and end-markers of factual sentences (baseline),
an advanced regular expression-based script, and, surprisingly efectively, extraction using a LLM. The
baseline approach lacked the complexity to successfully extract factual sentences from verdicts. In
contrast, the advanced regular expressions identified descriptions in 97% of verdicts and Gemini Flash
2.0 extracted them in 98.75% of the test data. The combination of both methods extracted descriptions
in 99.5% of cases. Manual annotation revealed that 91.75% of descriptions retrieved by the LLM and
89.5% of those retrieved by regular expressions match the descriptions identified by human annotators.
The combination of advanced regular expressions and LLM achieved 92% accuracy. In future work, we
aim to expand the dataset with court decisions from additional countries, providing empirical data for
comparative research in criminal law.</p>
    </sec>
    <sec id="sec-8">
      <title>Acknowledgments</title>
      <p>This study was funded by the Czech Grant Foundation (grant number 25-16848M entitled "Just Sentences:
Analyzing and Enhancing Proportionality and Consistency Using Typical Crimes"). The authors have
no competing interests.</p>
    </sec>
    <sec id="sec-9">
      <title>Declaration on Generative AI</title>
      <p>The authors have not employed any Generative AI tools.
[13] P. Boniol, G. Panagopoulos, C. Xypolopoulos, R. El Hamdani, D. R. Amariles, M. Vazirgiannis,
Performance in the courtroom: Automated processing and visualization of appeal court decisions
in france, in: Proceedings of the Natural Legal Language Processing Workshop 2020, 2020.
[14] J. Savelka, K. D. Ashley, Segmenting us court decisions into functional and issue specific parts., in:</p>
      <p>JURIX 2018, 2018, pp. 111–120.
[15] B. Hachey, C. Grover, Extractive summarisation of legal texts, Artificial Intelligence and Law 14
(2006) 305–345.
[16] M.-F. Moens, Summarizing court decisions, Information processing &amp; management 43 (2007)
1748–1764.
[17] A. Wyner, R. Mochales-Palau, M.-F. Moens, D. Milward, Approaches to text mining arguments
from legal cases, Springer, 2010.
[18] M. Saravanan, B. Ravindran, S. Raman, Improving legal document summarization using graphical
models, Frontiers in Artificial Intelligence and Applications 152 (2006) 51.
[19] M. Saravanan, B. Ravindran, Identification of rhetorical roles for segmentation and summarization
of a legal judgment, Artificial Intelligence and Law 18 (2010) 45–76.
[20] J. Drápal, A. Bolocan-Holban, J. Ginter, M. Plesničar, K. Tomšů, M. Vidaicu, Sentence justification
in first-level courts in post-communist europe, in: Justifying Punishment, Routledge, 2024, pp.
39–62.
[21] M. Terkanič, D. Gálisová, Revízia trestných štatistických listov a výkazov na slovensku: Desať
vecí, ktoré by sme najradšej spravili inak, Česká kriminologie 6 (2021) 1–11.</p>
    </sec>
    <sec id="sec-10">
      <title>Appendix A: Prompt</title>
      <sec id="sec-10-1">
        <title>Task: Extracting factual statements from text</title>
      </sec>
      <sec id="sec-10-2">
        <title>Your task is to extract the main factual statement ONLY from the provided text.</title>
      </sec>
      <sec id="sec-10-3">
        <title>A factual statement in a criminal judgment is that part of the judgment which precisely and specifically describes the act for which the accused is convicted. It is a detailed description of what happened - when, where, how, and with what consequence the criminal ofense was committed.</title>
      </sec>
      <sec id="sec-10-4">
        <title>The factual statement must contain all the legal elements of the criminal oefnse for which the accused is convicted, both the objective aspect (action, consequence, causal link) and the subjective aspect (intent or negligence).</title>
      </sec>
      <sec id="sec-10-5">
        <title>It is an important part of the operative part of the judgment and must be formulated in such</title>
        <p>a way that it is completely clear for which specific act the accused is convicted, and that
this act is unmistakable from any other. This is key for the principle of "ne bis in idem" (not
twice in the same matter), which prevents someone from being tried twice for the same
act.</p>
      </sec>
      <sec id="sec-10-6">
        <title>How else can we recognize the factual statement?</title>
      </sec>
      <sec id="sec-10-7">
        <title>It is located after the phrase "sa uznáva vinnou, že..." (is found guilty that...) - which is a</title>
        <p>typical introduction to the factual statement in Slovak criminal judgments.</p>
      </sec>
      <sec id="sec-10-8">
        <title>It contains a detailed and specific description of the committed act:</title>
      </sec>
      <sec id="sec-10-9">
        <title>Who: "The accused P. Z. ... as a parent obliged to continuously care for the upbringing..."</title>
      </sec>
      <sec id="sec-10-10">
        <title>What: "inconsistently approached the fulfillment of her duties... and created conditions for</title>
        <p>minors... for the emergence of undesirable habits in the form of long-term truancy"
When: "in the school year 2017/2018... in the period from 01. 09. 2017 to 30. 06. 2018"</p>
      </sec>
      <sec id="sec-10-11">
        <title>How: detailed description of the conduct including specific numbers (93 and 194 unexcused hours)</title>
      </sec>
      <sec id="sec-10-12">
        <title>Consequence: "through negligence, exposed persons younger than eighteen years to the danger of neglect by allowing them to lead an idle life"</title>
      </sec>
      <sec id="sec-10-13">
        <title>It contains all the legal elements of the constituent facts of the criminal oefnse of</title>
        <p>endangering the moral upbringing of youth according to § 211 par. 1 letter b) of the Criminal</p>
      </sec>
      <sec id="sec-10-14">
        <title>Code.</title>
      </sec>
      <sec id="sec-10-15">
        <title>It ends with a comma, followed by the legal qualification: "thereby c ommitting the ofense</title>
        <p>of endangering the moral upbringing of youth according to § 211 par. 1 letter b) of the</p>
      </sec>
      <sec id="sec-10-16">
        <title>Criminal Code."</title>
      </sec>
      <sec id="sec-10-17">
        <title>It is located in the operative part of the judgment (before the reasoning).</title>
      </sec>
      <sec id="sec-10-18">
        <title>It unequivocally and unmistakably describes the act for which the accused is convicted.</title>
        <p>sa u z n á v a z a v i n n é h o, ž e/u z n á v a s a z a v i n n é h o/sú vinní, že/u z n á v a s a v i n n
ý m, ž e/sa uznáva vinným, že/skutkovom základe, že (all can be either spaced out or not)</p>
      </sec>
      <sec id="sec-10-19">
        <title>THE FACTUAL STATEMENT ALWAYS IN 100% OF CASES ENDS WITH THIS AND SIMILAR</title>
      </sec>
      <sec id="sec-10-20">
        <title>TERM</title>
        <p>t e d a (t h u s / t h e r e f o r e)</p>
      </sec>
      <sec id="sec-10-21">
        <title>If the text does NOT meet these conditions, it means it does not have a factual statement</title>
      </sec>
      <sec id="sec-10-22">
        <title>If you do not find a factual statement in the text, use this structure:</title>
        <p>{
"skutkova_veta": null,
"no_factual_statement_reason": "Explanation why the text does
not contain a factual statement."
}</p>
      </sec>
      <sec id="sec-10-23">
        <title>Example:</title>
      </sec>
      <sec id="sec-10-24">
        <title>Provided text: "Obvinený Ján Novák dňa 12.5.2022 o 15:30 hod. na ulici Dlhá v Bratislave</title>
        <p>fyzicky napadol poškodeného Petra Svobodu, čím sa dopustil trestného činu ublíženia na
zdraví podľa § 156 trestného zákona." (Accused Ján Novák on 12.5.2022 at 15:30 hrs. on</p>
      </sec>
      <sec id="sec-10-25">
        <title>Dlhá street in Bratislava physically assaulted the victim Peter Svoboda, thereby committing the criminal ofense of bodily harm according to § 156 of the Criminal Code.)</title>
      </sec>
      <sec id="sec-10-26">
        <title>Expected AI output:</title>
        <p>{
}</p>
        <p>"skutkova_veta": "Ján Novák dňa 12.5.2022 o 15:30 hod. na
ulici Dlhá v Bratislave fyzicky napadol Petra Svobodu.",
"no_factual_statement_reason": null
{
"skutkova_veta": null,
"no_factual_statement_reason": "The text only contains
procedural information about the termination of proceedings, it
does not contain a description of the act."
}</p>
      </sec>
      <sec id="sec-10-27">
        <title>HOW TO WORK WITH TEXT THAT HAS A FACTUAL STATEMENT? Examples of factual statements: 1. 2.</title>
        <p>on 29.12.2019 at around 17.05 h in Bratislava in OC Centrál in the DM Drogerie store
gradually took from
the shelves 1 pc Denim VPH original 100 ml valued at 12.58 Eur, 3 pcs Gillette VPH Artic ice
100 ml valued at 48.93 Erur, 1 pc Denim VPH original 100 ml valued at 12.58 Eur, 4 pcs</p>
      </sec>
      <sec id="sec-10-28">
        <title>Gillette M Fusion power syst. valued at 89.94 Eur, 2 pcs Gillette M Fusion 1+1NH valued at</title>
        <p>35.97 Eur, 2 pcs Gillette M Fusion proshield syst. 1+4 NH valued at 49.98 Eur, 1 pc Denim</p>
      </sec>
      <sec id="sec-10-29">
        <title>VPH black 100 ml valued at 56.61Eur, 3 pcs Gillette Mach 3 syst. 1+1 NH valued at 49.95,</title>
      </sec>
      <sec id="sec-10-30">
        <title>Eur, immediately placed the mentioned goods into a plastic bag and subsequently passed through the checkout zone without paying, thereby causing damage to the company DM drogerie markt, s.r.o., Bratislava, Na pántoch 18, IČO: 31 393 781, in the amount of 356.54 Eur,</title>
        <p>1. at an unspecified time at the end of August 2010 in the village of Y. the accused K.
entered without consent
and knowledge of the owner into the yard of family house no. XXX and from an
unlocked shed took a motor
mower brand Jičín, red color, which he pushed out in front of the gate, where all
three accused together loaded it
into a motor vehicle brand Q. U. and drove away, thereby causing damage to the
owner of the mower Q. D., born XX.XX.XXXX
in the amount of 179.20 €</p>
      </sec>
      <sec id="sec-10-31">
        <title>2. at an unspecified time, in the period from 21:30 hrs. on 31.08.2010 to 07:00 hrs. on 01.09.2010, in</title>
      </sec>
      <sec id="sec-10-32">
        <title>S. Y., city district Y. all three accused came to the grocery store X. N., with a brought crowbar removed two padlocks on the iron cage located at the entrance to the store, from where they took 14 pcs</title>
        <p>propane-butane cylinders with gas lfiling weighing 10 kg, loaded them into a motor
vehicle brand Ford</p>
      </sec>
      <sec id="sec-10-33">
        <title>Transit and drove away, thereby causing damage to the owner of the stolen</title>
        <p>cylinders, company C. M. B. H. O. K. Z. G. Z..G..,
VAT ID: XX XXX XXX damage in the amount of 370.44 € and to the owner of the
damaged locks and stolen
gas, cooperative X. N. U., G. S., VAT ID: XX XXX XXX damage in the amount of 202.80
€
3. at an unspecified time, in the period from 17:00 hrs. on 31.08.2010 to 10:00 hrs. on
02.09.2010 in
the village of G. S. all three accused came to the grocery store, with a brought
crowbar removed the padlocks
on the iron cage located at the entrance to the store, from where they took 10 pcs
propane-butane
cylinders with gas lfiling weighing 10 kg, loade d them into a Ford Transit motor
vehicle and drove away, thereby
causing damage to the owner of the stolen cylinders, company C. Z..G.., VAT ID: XX
XXX XXX, damage in the amount of 264.60
€ and to the owner of the damaged locks and stolen gas, I.. M. S., born XX.XX.XXXX
damage in the amount of
142.80 €
dated 29.05.2010, legally eefctive on 29.05.2010, he was found guilty of committing
the oefnse "Theft"
under § 212 par. 2 letter a), par. 3 letter a) of the Criminal Code,
3.
that
on 29. 3. 2013 at 10.45 hrs. drove a personal motor vehicle brand Renault Scénic, reg. no.
LM-040 CF, in</p>
      </sec>
      <sec id="sec-10-34">
        <title>Liptovský Mikuláš along Demänovská cesta in the direction from OD Kaufland towards</title>
      </sec>
      <sec id="sec-10-35">
        <title>Palúčanská street and near the</title>
      </sec>
      <sec id="sec-10-36">
        <title>Elementary School Demänovská cesta was stopped by a patrol of the Regional Directorate of the Police Force Žilina, rapid response unit</title>
      </sec>
      <sec id="sec-10-37">
        <title>PZ Žilina and performed this activity despite the fact that by the decision on the ofense of</title>
        <p>the OR PZ, District
Traifc Inspectorate Liptovský Mikuláš under no. ORPZ-LM-ODI2-P-364/2011 dated 26. 8.
2011, which
became legally efective on 26. 8. 2011, he was imposed a ban on driving motor vehicles for
a period of
36 months from the legal efectivity of the decision,
in the period for the month of March 2018, December 2018, January 2019 until
21.08.2019 inclusive as the father of minor
son G. Č., born XXXX, he fails to fulfill his maintenance obligation in Y. and in other
places where he stays,
although this obligation arises from the Family Act and was determined for him by the
judgment of the District Court Skalica file ref.
8P/155/2017 dated 28.02.2018, legally efective on 07.03.2018, by which he was
entrusted to the custody of the mother
D. O., residing at Y., M. XX and the father was ordered to contribute to the maintenance
of the minor son G. Č. in the amount of 150,
€ by the 15th day of each month in advance into the hands of the mother, thereby causing
for the specified period a debt on
maintenance in the amount of 1,200,
• € into the hands of the mother D. O., residing at Y., M. XX,</p>
      </sec>
      <sec id="sec-10-38">
        <title>Accused P. Z. born XX. XX. XXXX P. X. permanently residing at L.Á. S. XX, X. X. is found guilty</title>
        <p>that as a parent obliged to continuously care for the upbringing and comprehensive
development of the child, she inconsistently approached the fulfillment of he r duties
according to the Family Act No. 36/2005 Coll. as amended and created conditions for the
minors C. Z., born XX. XX. XXXX, permanently residing at L. S. XX, X. X. and Š. Z., born XX.
XX. XXXX, permanently residing at L. S. XX, X. X., for the emergence of undesirable habits in
the form of long-term truancy, when C. Z. in the school year 2017/2018 missed
unexcusedly 93 hours from the teaching process at Š. Z. Š., Ď. XX, X. X., in the period from
the danger of neglect by allowing them to lead an idle life,
Expected AI output:
{
}</p>
        <p>"skutkova_veta": "ako rodič povinný sústavne sa starať o
výchovu a všestranný rozvoj dieťaťa nedôsledne pristupovala k
plneniu si povinností podľa Zákona o rodine č. 36/2005 Z. z. v
platnom znení a vytvorila maloletým C. Z., narodenému XX. XX.
XXXX, trvalo bytom L. S. XX, X. X. a Š. Z., narodenému XX. XX.
XXXX, trvalo bytom L. S. XX, X. X., podmienky pre vznik
nežiaducich návykov vo forme dlhodobého záškoláctva, keď C. Z.
v školskom roku 2017/2018 z vyučovacieho procesu na Š. Z. Š.,
Ď. XX, X. X., v období od 01. 09. 2017 do 30. 06. 2018 vymeškal
neospravedlnene 93 hodín a Š. Z. v školskom roku 2017/2018 z
vyučovacieho procesu na Š. Z. Š., Ď.D. XX, X. X., v období od
01. 09. 2017 do 30. 06. 2018 vymeškal neospravedlnene 194
hodín, vydala z nedbanlivosti osoby mladšie ako osemnásť rokov
nebezpečenstvu spustnutia tým, že im umožnila viesť záhaľčivý
život",</p>
        <p>"no_factual_statement_reason": null
Instructions:
1. Identify the main factual statement in the provided text
2. Extract the factual statement exactly as it is stated in the text
3. DO NOT invent any new facts or information that are not explicitly stated in the
provided text.
4. NEVER include the text before and after the key passage sa u z n á v a z a v i n n é h o,
ž e/u z n á v a s a z a v i n n é h o/sú vinní, že/u z n á v a s a v i n n ý m, ž e/sa uznáva
vinným, že/skutkovom základe, že (all can be either spaced out or not)/teda...</p>
      </sec>
      <sec id="sec-10-39">
        <title>5. Return the response in JSON format with the following structure:</title>
        <p>"skutkova_veta": "Extracted factual statement from the
provided text",</p>
        <p>"no_factual_statement_reason": null</p>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>V. R.</given-names>
            <surname>Walker</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Pillaipakkamnatt</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. M.</given-names>
            <surname>Davidson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Linares</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. J.</given-names>
            <surname>Pesce</surname>
          </string-name>
          ,
          <article-title>Automatic classification of rhetorical roles for sentences: Comparing rule-based scripts with machine learning</article-title>
          .,
          <source>in: ASAIL@ ICAIL</source>
          <year>2019</year>
          ,
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>L.</given-names>
            <surname>Zhong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Zhong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Zhao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <surname>K. D. Ashley</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Grabmair</surname>
          </string-name>
          ,
          <article-title>Automatic summarization of legal decisions using iterative masking of predictive sentences</article-title>
          ,
          <source>in: ICAIL</source>
          <year>2019</year>
          ,
          <year>2019</year>
          , pp.
          <fpage>163</fpage>
          -
          <lpage>172</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>P.</given-names>
            <surname>Bhattacharya</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Paul</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Ghosh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Ghosh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Wyner</surname>
          </string-name>
          ,
          <article-title>Identification of rhetorical roles of sentences in indian legal judgments</article-title>
          ,
          <source>in: JURIX</source>
          <year>2019</year>
          , volume
          <volume>322</volume>
          , IOS Press,
          <year>2019</year>
          , p.
          <fpage>3</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>K.</given-names>
            <surname>Branting</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Weiss</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Brown</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Pfeifer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Chakraborty</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Ferro</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Pfaf</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Yeh</surname>
          </string-name>
          ,
          <article-title>Semisupervised methods for explainable legal prediction</article-title>
          ,
          <source>in: ICAIL</source>
          <year>2019</year>
          ,
          <year>2019</year>
          , pp.
          <fpage>22</fpage>
          -
          <lpage>31</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>J.</given-names>
            <surname>Šavelka</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Westermann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Benyekhlef</surname>
          </string-name>
          ,
          <article-title>Cross-domain generalization and knowledge transfer in transformers trained on legal data</article-title>
          ,
          <source>in: ASAIL@ JURIX</source>
          <year>2020</year>
          ,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>I.</given-names>
            <surname>Habernal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Faber</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Recchia</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Bretthauer</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Gurevych</surname>
          </string-name>
          ,
          <string-name>
            <surname>I.</surname>
          </string-name>
          <article-title>Spiecker genannt Döhmann, C. Burchard, Mining legal arguments in court decisions</article-title>
          ,
          <source>Artificial Intelligence and Law</source>
          <volume>32</volume>
          (
          <year>2024</year>
          )
          <fpage>1</fpage>
          -
          <lpage>38</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>H.</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Šavelka</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K. D.</given-names>
            <surname>Ashley</surname>
          </string-name>
          ,
          <article-title>Using argument mining for legal text summarization</article-title>
          ,
          <source>in: JURIX</source>
          <year>2020</year>
          , volume
          <volume>334</volume>
          , IOS Press,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>A.</given-names>
            <surname>Petrova</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Armour</surname>
          </string-name>
          , T. Lukasiewicz,
          <article-title>Extracting outcomes from appellate decisions in us state courts</article-title>
          ,
          <source>in: JURIX</source>
          <year>2020</year>
          ,
          <year>2020</year>
          , p.
          <fpage>133</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>M.</given-names>
            <surname>Grabmair</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K. D.</given-names>
            <surname>Ashley</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Sureshkumar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Nyberg</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V. R.</given-names>
            <surname>Walker</surname>
          </string-name>
          ,
          <article-title>Introducing luima: an experiment in legal conceptual retrieval of vaccine injury decisions using a uima type system and tools</article-title>
          ,
          <source>in: Proceedings of the 15th international conference on artificial intelligence and law</source>
          ,
          <year>2015</year>
          , pp.
          <fpage>69</fpage>
          -
          <lpage>78</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>A.</given-names>
            <surname>Bansal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Bu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Mishra</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Ashley</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Grabmair</surname>
          </string-name>
          ,
          <article-title>Document ranking with citation information and oversampling sentence classification in the luima framework</article-title>
          ,
          <source>in: Legal Knowledge and Information Systems</source>
          , IOS Press,
          <year>2016</year>
          , pp.
          <fpage>33</fpage>
          -
          <lpage>42</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>A.</given-names>
            <surname>Farzindar</surname>
          </string-name>
          ,
          <string-name>
            <surname>G.</surname>
          </string-name>
          <article-title>Lapalme, Letsum, an automatic text summarization system in law field</article-title>
          ,
          <source>in: Proceedings of JURIX</source>
          <year>2004</year>
          ,
          <year>2004</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>J.</given-names>
            <surname>Harašta</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Šavelka</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Kasl</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Míšek</surname>
          </string-name>
          , et al.,
          <article-title>Automatic segmentation of czech court decisions into multi-paragraph parts, Jusletter IT (</article-title>
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          <article-title>4. at an unspecified time in the period from the beginning of December 2010 to 13</article-title>
          .02.2011 in S. Y. on
          <string-name>
            <surname>I.. R.</surname>
          </string-name>
          <article-title>street the accused K. broke the entrance door lock area on garage no. 1429, entered inside and inside the garage broke the locked trunk door of the vehicle parked there, brand</article-title>
          <string-name>
            <given-names>X. D.</given-names>
            ,
            <surname>Lic</surname>
          </string-name>
          . No. S.
          <article-title>-XXXAJ, from which he took a four-wheeled tractor mower brand Rider F XX, serial no. XXXXX and from the garage took 8 pcs winter tires, a men's mountain bike, blue color, brand Author and wooden sleds, thereby causing damage to the owner of the garage and stolen items</article-title>
          , N.. K. R.,
          <source>born XX</source>
          .
          <article-title>XX.XXXX damage in the amount of 1,921.01 €, whereas the accused K. acted this way despite the fact that by the penal order of the District Court M. no</article-title>
          .
          <source>XT/XX/XXXX 01. 09</source>
          .
          <year>2017</year>
          to
          <volume>30</volume>
          . 06.
          <year>2018</year>
          ,
          <article-title>and Š. Z. in the school year 2017/2018 missed unexcusedly 194 hours from the teaching process at Š</article-title>
          . Z. Š.,
          <string-name>
            <surname>Ď.D. XX</surname>
            ,
            <given-names>X. X.</given-names>
          </string-name>
          ,
          <source>in the period from 01</source>
          . 09.
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          2017 to 30. 06.
          <year>2018</year>
          ,
          <article-title>through negligence, exposed persons younger than eighteen years to</article-title>
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>