<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>Corresponding author.
†These authors contributed equally.
$ ekamateri@iee.ihu.gr (E. Kamateri); renukswamy.chikkamath@hm.edu (R. Chikkamath); msa@ihu.gr (M. Salampasis);
linda.andersson@artificialresearcher.com (L. Andersson); markus.endres@hm.edu (M. Endres)</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Enhancing patent retrieval using automated patent summarization⋆</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Eleni Kamateri</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Renukswamy Chikkamath</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Michail Salampasis</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Linda Andersson</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Markus Endres</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Artificial Researcher - IT GmbH</institution>
          ,
          <country country="SE">Sweden</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Department of Information and Electronic Engineering, International Hellenic University</institution>
          ,
          <addr-line>Sindos, 57400 Thessaloniki</addr-line>
          ,
          <country country="GR">Greece</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Hochschule München</institution>
          ,
          <addr-line>Loth str. 34, 80335 München</addr-line>
          ,
          <country country="DE">Germany</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2025</year>
      </pub-date>
      <volume>000</volume>
      <fpage>0</fpage>
      <lpage>0003</lpage>
      <abstract>
        <p>Efective query formulation is a key challenge in long-document Information Retrieval (IR). This challenge is particularly acute in domain-specific contexts like patent retrieval, where documents are lengthy, linguistically complex, and encompass multiple interrelated technical topics. In this work, we present the application of recent extractive and abstractive summarization methods for generating concise, purpose-specific summaries of patent documents. We further assess the utility of these automatically generated summaries as surrogate queries across three benchmark patent datasets and compare their retrieval performance against conventional approaches that use entire patent sections.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;patent retrieval</kwd>
        <kwd>query formulation</kwd>
        <kwd>patent summarization</kwd>
        <kwd>big bird</kwd>
        <kwd>summary segment</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Drafting a representative abstract that accurately summarizes the core concepts of an invention is
an important step in the patenting process. A well-crafted abstract conveys the core concepts of an
invention, therefore it enhances both the readability and discoverability of a patent throughout its
lifecycle. For instance, integrating summaries into search snippets could reduce examiner search time,
or help an inventor to quickly grasp prior art. A good summary may also be a great assistance for a
patent professional to evaluate the technical or legal scope of a patent. Furthermore, a good summary
retaining technical details and key claims could be used for downstream task such as patent prior-art
and classification [
        <xref ref-type="bibr" rid="ref1 ref2">1, 2</xref>
        ].
      </p>
      <p>
        However, many human-authored patent abstracts do not summarize the invention efectively. This
shortcoming may arise from various factors, such as the urgency to submit the application, regulatory
constraints on abstract length, limited attention by inventors, and, last but not least, the intentional
vagueness often employed to avoid narrowing the scope of legal protection and reduce discoverability
in prior-art searches [
        <xref ref-type="bibr" rid="ref3 ref4">3, 4</xref>
        ]. Consequently, relying directly on the patent abstract for producing a search
query for patent retrieval tasks is often inefective.
      </p>
      <p>
        As a result, patent abstracts are often supplemented by human-selected keywords. Alternatively,
content is extracted from other sections of the patent, such as the description or claims, to produce
queries that will enhance retrieval performance [
        <xref ref-type="bibr" rid="ref5 ref6 ref7">5, 6, 7</xref>
        ]. To address this need, various methods have been
developed to automatically generate search queries from patent applications employing simple intuitive
heuristics (e.g., the first X words), statistical techniques, language modeling and other methods. Some
of these methods enhance abstracts by directly incorporating content from the description and claims
sections [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. Other methods aim to identify the most discriminative terms across diferent sections by
comparing term statistics within a given patent to those of a broader corpus, often leveraging language
modelling estimation techniques [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]. Query enrichment may also involve query terms extracted from
patent citations or classification codes [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ].
      </p>
      <p>
        When using LLMs, there are two basic approaches for handling large documents -such as patents- for
text retrieval and other text-related tasks. The first is document chunking to overcome token limits of
LLMs. The second is document summarization which aims to preserve the semantic flow and concepts
of a long document leading to a reduced text representation for eficient processing. AI-generated
summaries are increasingly adopted across domains for their conciseness and informativeness. In the
patent domain, however, their use has been primarily explored in classification tasks [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ], where recent
studies show that classifiers trained on generated summaries outperform those trained on original
abstracts. Beyond classification, there is a growing need for purpose-specific summaries tailored to other
downstream tasks. These summaries can take diferent forms depending on what the task requires (e.g.,
abstract, extended summary, or first claim), summarization approach (e.g., abstractive or extractive),
input source (e.g., description, claims, or brief description) and output length (e.g., 50 to 300 words).
One promising application is patent retrieval task, where high-quality and rightly sized summaries
can serve as search queries that contain meaningful context but not so large that they overwhelm the
model.
      </p>
      <p>Building on these considerations, this study presents a pipeline for automated patent summarization.
Furthermore, it examines the efectiveness of these automatically generated summaries as search
queries for prior-art search, a critical task in the global patent operation. We employ state-of-the-art
language models and semantically rich patent documents segments to generate both extractive and
abstractive summaries, with the main goal to improve retrieval performance. We assess multiple input
source combinations and summarization approaches across three patent datasets, determining which
configurations produce the most informative summaries for retrieval tasks. Our results show that
AI-generated summaries, when used as queries, consistently outperform other traditional strategies
that rely on full patent sections.</p>
      <p>The remainder of the paper is organized as follows: Section 2 and Section 3 detail the semantically
important segments found in patent description and present summarization techniques. Section 4
outlines the methodology adopted in our study. In Section 5 we present the experimental results, while
in Section 6 we discuss the findings and conclude the paper.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Patent description for search</title>
      <p>
        Among patent sections in a patent document, the description is consistently identified as the most
informative and valuable source for query generation [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. As the longest component, it ofers a detailed
elaboration of the proposed invention, often extending to several thousand words [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ]. However, the
absence of a standardized structure or mandated format for organizing patent description complicates
automatic processing. This challenge is partially mitigated by the widespread use of conventional
headings, such as Background, Summary of the Invention, Brief Description of the Drawings, and
Detailed Description of the Invention. These headings are commonly adopted by patent applicants to
organize the description into semantically coherent segments.
      </p>
      <p>
        Notably, the combination of the Background and Summary of the Invention, which are collectively
referred to as the “background summary”, has proven to be the most efective source for extracting
high-value query terms from U.S. patent documents [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ]. The practical importance of these segments
led to increased interest in their standardization. In response, the USPTO has expanded its public API
to provide direct access to key description sub-sections, like the Background and Brief Description. The
Brief Description, labeled as “brief” in the USPTO API, spans from the beginning of the description to
the end of the Summary of the Invention segment, efectively capturing what is traditionally known as
the “background summary”.
      </p>
      <p>
        Similarly, the Harvard USPTO Patent Dataset (HUPD) [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ] uses USPTO documents’ semi-structure and
simple heuristics to extract more meaningful segments like the Summary of the Invention. This segment
ofers a more comprehensive and informative description of the invention substantially augmenting the
abstract’s content while improving classification performance [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ].
      </p>
      <p>Although these segments enhance patent retrieval and classification, they appear inconsistently across
patent documents. For instance, these or other corresponding headings often do not exist in European
patents. This structural inconsistency underscores further the motivation for automated methods capable
of generating summary-like content (resembling the USPTO’s Summary of the Invention), particularly in
documents that lack clearly defined segments in the description. To that end, summarization techniques
play a key role in bridging this gap by producing coherent, informative summaries.</p>
    </sec>
    <sec id="sec-3">
      <title>3. Patent summarization</title>
      <p>Extractive and abstractive summarization represent two distinct approaches to generate summaries from
text. Extractive summarization uses statistical or neural models to select the most relevant sentences
directly from the source text, preserving the original text. In contrast, abstractive summarization uses
transformer models to generate a condensed version of the content by rephrasing or synthesizing
information using natural language generation techniques, producing more coherent and readable
summaries. Finally, hybrid approaches may be used to combine accuracy (selecting existing text using
extractive methods) as well as fluency by enhancing the previously extracted text with abstractive
methods.</p>
      <p>
        Typically, extractive models work by generating sentence-level embeddings, clustering these
embeddings, and selecting the sentences nearest to the cluster centroids as the most representative.
State-of-the-art models like BERT [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ] and its variant SBERT [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ] are widely used for this purpose.
These models efectively extract the most informative sentences without altering them, maintaining the
original phrasing and structure of the source document.
      </p>
      <p>
        Abstractive summarization models, on the other hand, follow an encoder-decoder architecture: they
encode the input text, generate a summary through a decoding process, and produce a fluent, often
restructured, output [
        <xref ref-type="bibr" rid="ref17 ref18">17, 18</xref>
        ]. Notable models in this category include PEGASUS [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ], T5 [
        <xref ref-type="bibr" rid="ref20">20</xref>
        ], and
BART [
        <xref ref-type="bibr" rid="ref21">21</xref>
        ], which have demonstrated strong performance on long-document summarization tasks.
Although GPT-based models [
        <xref ref-type="bibr" rid="ref22">22</xref>
        ] also demonstrate strong performance in abstractive summarization,
their closed-source nature and token-based pricing present practical limitations for large-scale use.
      </p>
      <p>
        In the patent domain, recent research has primarily focused on generating summaries from the
description and/or claims sections using Large Language Models (LLMs) [
        <xref ref-type="bibr" rid="ref13 ref23">13, 23</xref>
        ]. We refer to these
outputs as automated summaries, to distinguish them from the human-authored "Summary of the
Invention" segments found within the description (hereafter referred to as summary segments). A
number of previous studies have fine-tuned pre-trained language models on patent datasets. For
instance, one early work [
        <xref ref-type="bibr" rid="ref24">24</xref>
        ] trained Seq2Seq [
        <xref ref-type="bibr" rid="ref25">25</xref>
        ], PointGenerator [
        <xref ref-type="bibr" rid="ref26">26</xref>
        ], and SentRewriting [
        <xref ref-type="bibr" rid="ref27">27</xref>
        ] models
on the BIGPATENT dataset, using the description as input and the abstract as the target output.
BigBird-Pegasus, a long-sequence transformer model, was later fine-tuned on BIGPATENT for improved
summarization of patent texts [
        <xref ref-type="bibr" rid="ref28">28</xref>
        ]. Similarly, the work in [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ] adapted two versions of the T5 model
for patent data, using either the claims or the description sections to generate abstracts. Another
study reported in [
        <xref ref-type="bibr" rid="ref29">29</xref>
        ] investigated which patent sections are most informative for generating the first
independent claim (also referred to as the first claim), using PEGASUS and PointGenerator models,
concluding that the summary segment is the most suitable input source for this task. These eforts
highlight the importance of fine-tuning and domain adaptation, as general-purpose transformer models
often struggle with the technical and legal precision required in patent domain. Moreover, although
recent LLMs show considerable promise to process long patent descriptions (such as GPT-4 [
        <xref ref-type="bibr" rid="ref30">30</xref>
        ] and
Llama-3.14 [
        <xref ref-type="bibr" rid="ref31">31</xref>
        ]), their capabilities have not been extensively investigated in the context of
patentrelated tasks [
        <xref ref-type="bibr" rid="ref32">32</xref>
        ]. By examining both the structural characteristics of patent documents and the current
state of research on patent summarization, we have identified several open issues that merit further
investigation to improve the quality and utility of generated summaries, as follows:
1. Existing models are typically not trained on the full patent text. Due to token length constraints,
typically limited to 512 or 1024 tokens in standard LLMs, such as T5, PEGASUS, and BART, and
extended up to 4096 tokens or more in models like BigBird-Pegasus, the input text often needs
to be truncated, segmented or adapted. This can hinder the model’s ability to fully contextual
capture understanding.
2. Evaluation commonly relies on existing abstracts as ground-truth summaries, despite their
frequent shortcomings in terms of clarity, completeness, and informativeness.
3. A strong semantic alignment exists between the first claim and the summary segment. However,
this relationship remains underexplored in current summarization approaches.
      </p>
      <p>These limitations highlight the need for summarization strategies that are specifically tailored to
the structure and practical use cases of patent documents. In particular, for retrieval tasks, such as
prior-art search, where the generated summary serves as an efective retrieval query, it is essential to
adopt summarization approaches that leverage high-value sections or combination of them to generate
summaries that attain high retrieval performance. To operationalize this approach, our study follows a
pipeline that first extracts key patent segments, then trains summarization models, and finally evaluates
their efectiveness of the automated summaries in prior-art retrieval across multiple benchmark datasets.</p>
    </sec>
    <sec id="sec-4">
      <title>4. Methodology</title>
      <p>This study hypothesizes that generated summaries can improve the eficiency of prior-art search. To
validate this hypothesis, we designed a five-stage experimental workflow, beginning with summary
generation and ending with an evaluation of their efectiveness as retrieval queries. The performance
of these summaries as queries is compared against sections, abstract, claims, and description, which are
commonly used by patent professionals to formulate queries. In the following sections, we describe the
data collections, we detail each stage of the methodology, and explain the evaluation process used to
assess both the intrinsic quality of the generated summaries and their impact on retrieval performance.</p>
      <sec id="sec-4-1">
        <title>4.1. Data collections</title>
        <p>
          We utilize four patent datasets, each serving a distinct role. The first dataset, HUPD, allows the
extraction of salient sections from patent documents. The second dataset, BIGPATENT, is employed
for the intrinsic evaluation of the generated summaries, i.e. to measure the quality of the automated
summaries compared to reference summaries. Finally, the next two datasets, CLEF-IP 2013 and USPTO,
are used for extrinsic evaluation applying summaries for prior-art search.
4.1.1. HUPD [
          <xref ref-type="bibr" rid="ref13">13</xref>
          ]
The HUPD is a large-scale, structured corpus of English-language utility patent applications filed to
the USPTO between 2004 and 2018. Each JSON-formatted entry contains rich metadata, including
bibliographic details, classification codes, inventor information, and full text fields such as abstract,
claims, background, summary, and description.
4.1.2. BIGPATENT [
          <xref ref-type="bibr" rid="ref24">24</xref>
          ]
The BIGPATENT dataset is a large-scale patent summarization benchmark comprising approximately 1.3
million U.S. patent documents. It pairs the description section with its corresponding abstract, serving
as a ground truth for training and fine-tuning summarization models. For our intrinsic evaluation, we
randomly sampled 1,000 patents from the available 67,072 of the test set.
4.1.3. CLEF-IP [
          <xref ref-type="bibr" rid="ref33">33</xref>
          ]
The CLEF-IP collection consists of patent documents sourced from the EPO and WIPO. The English topic
set from the CLEF-IP 2013 campaign originally comprises 50 topics [
          <xref ref-type="bibr" rid="ref34">34</xref>
          ]. However, because the topics
are based on patent claims rather than individual documents, there is no strict one-to-one mapping
between topics and documents. In total, these topics correspond to 37 unique documents. Due to
missing relevant documents for some topics in the indexed dataset, we further reduced the set to 24
English-language patents for our experiments. Each topic patent is associated with between 2 and 8
manually identified relevant documents, based on expert-curated citation links, making this dataset a
reliable benchmark for evaluating prior-art retrieval performance.
4.1.4. USPTO-Explainable AI for Patent Professionals [
          <xref ref-type="bibr" rid="ref35">35</xref>
          ]
This USPTO dataset was released as part of a Kaggle competition aimed at advancing explainable AI in
the patent domain. Each topic patent is associated with a set of 50 most similar patents, identified using
content similarity measures rather than citation-based relevance. From this dataset, we selected 3,343
topic patents in which semantically coherent segments were automatically detected. Unlike CLEF-IP,
which relies on citation-based ground truth, this benchmark provides an opportunity to test retrieval
performance under automated similarity-based relevance.
        </p>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. Patent part extraction</title>
        <p>The first step in our pipeline focuses on identifying and extracting sections of each patent to serve as
query sections or as input sources for summarization. Specifically, we use the description and claims
sections, and when identifiable, include the brief description, summary and first claim.</p>
        <p>To detect these segments, we utilized the HUPD dataset, which is currently the only resource
providing annotated labels (i.e., tags) for the background and summary segments within the description
section of US patent documents. Based on these annotations, we constructed a dictionary of relevant
summary headings, which we then used as a reference to identify candidate headings in unannotated
patents. For each heading labeled as a summary heading, the subsequent content was marked as a
summary segment. Once the summary segment was identified, the brief description was also derived by
selecting the text spanning from the beginning of the description section up to the end of the summary.
Finally, the first claim was extracted using heuristic rules, specifically, by identifying the first claim that
is not dependent on any previous claim.</p>
      </sec>
      <sec id="sec-4-3">
        <title>4.3. Patent summarization</title>
        <p>
          In this phase, we employ three summarization models, the BERT [
          <xref ref-type="bibr" rid="ref15">15</xref>
          ], SBERT [
          <xref ref-type="bibr" rid="ref16">16</xref>
          ] and BigBird-Pegasus
[
          <xref ref-type="bibr" rid="ref28">28</xref>
          ] (from now on referred to as BigBird). BERT and SBERT are utilized for extractive summarization,
focusing on identifying the most relevant sentences from the input text. BigBird, which has been
pre-trained on the BIGPATENT dataset, serves as our primary abstractive summarization model, to
handle long-form patent text efectively.
        </p>
        <p>For BERT and SBERT, we used models’ default configurations. Specifically, for SBERT, we employed
the “paraphrase-MiniLM-L6-v2” model. For BigBird, we used a model fine-tuned on the BIGPATENT
dataset. Two configuration variants of the BigBird model were explored: the default configuration,
which generates relatively short summaries (typically between 50 and 100 words, depending on the
input), and a modified version, where the model’s generation parameters were adjusted (i.e., the length
penalty and minimum/maximum length settings) to produce longer summaries ranging from 250 to 300
words.</p>
        <p>In this study, the BigBird model is further fine-tuned to generate summaries using the brief description
and first claim as inputs, which are two sections identified as particularly informative within patent
documents. The target output for this fine-tuning process is the summary segment. This exploration
aims to set the foundation for future research on fine-tuning summarization models to replicate other
valuable parts of patent text, such as the extended, author-crafted summary segments found within the
description section.</p>
        <p>To achieve this, a new dataset, which is a subpart of the HUPD dataset, is specially created. Specifically,
we extracted from the HUPD dataset 402,921 patents that have a distinct summary segment with a
length between 150 and 250 words. Then, we extracted the brief description and first claim and selected
those patents whose brief description and first claim together had a length between 700 and 800 words.
This selection criterion allows to skip any adjustment steps of the input text during the training, such as
truncation or chunking, which may negatively afect the model’s interpretation. All these steps led to a
dataset comprising 48,322 patents, which was finally used to fine-tune the BigBird model. An overview
of the summarization and training parameters is shown in Table 1.</p>
      </sec>
      <sec id="sec-4-4">
        <title>4.4. Patent retrieval</title>
        <p>
          For the patent retrieval task, we use the FAISS vector database to store and retrieve semantic vectors.
Both queries and patent documents are embedded using the GTE-large-en-v1.5 model [
          <xref ref-type="bibr" rid="ref36">36</xref>
          ], hereafter
referred to as GTE, which has approximately 409 million parameters, a 1,024-dimensional embedding
size, and supports input lengths up to 8,192 tokens. This enables the generation of rich embeddings that
efectively represent full patents. The model achieved state-of-the-art performance on the Massive Text
Embedding Benchmark (MTEB) within its size category, making it well-suited for our application. To
reduce hardware complexity, we limit input length to 3,000 tokens, which is sucfiient to capture both
independent and dependent claims and is used as corpus embeddings. While alternative embedding
models and strategies, such as using diferent sections (e.g., abstracts, descriptions) or specific segments
(e.g., brief descriptions, summaries) as corpus representations ofer promising avenues for exploration,
we leave these investigations for future work.
        </p>
        <p>Since our primary goal is to assess the impact of generated summaries on prior-art retrieval, we
restrict the vector index to a subset of 200,000 patent documents drawn from our two prior-art datasets.
Retrieval performance is then compared against simpler methods that use entire patent sections or
extracted segments as queries. Given that our summarization pipeline is based on semantic models, and
our aim is to isolate the contribution of the generated summaries from the retrieval technique itself, we
focus exclusively on embedding-based retrieval method rather than keyword-based approaches. The
integration of additional retrieval methods is left for future work.</p>
      </sec>
      <sec id="sec-4-5">
        <title>4.5. Intrinsic and extrinsic evaluation</title>
        <p>To evaluate the efectiveness of the generated summaries, we perform two types of evaluations: intrinsic
and extrinsic.</p>
        <p>
          For the intrinsic evaluation, the automatically generated summaries are compared against reference
summaries, either the original patent abstract or the annotated summary segment (which has been
proved to provide an extended and improved summary). This evaluation aims to determine how
accurately the generated summaries captures the key content of the patents. For this purpose, we
use ROUGE scores [
          <xref ref-type="bibr" rid="ref37">37</xref>
          ] to assess textual overlap and compute semantic similarity, calculated as the
cosine similarity between embeddings produced by Google’s BERT-for-Patent model [
          <xref ref-type="bibr" rid="ref10">10</xref>
          ], comparing
the generated and reference summaries.
        </p>
        <p>For the extrinsic evaluation, we measure the impact of using the generated summaries as queries in a
prior-art search task. Specifically, we compare their retrieval performance against traditional query
strategies using standard IR metrics, including Mean Average Precision (MAP), Precision (P), and Recall
(R).</p>
        <p>Summary Segment</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Results</title>
      <sec id="sec-5-1">
        <title>5.1. Evaluation of summary generation in the BIGPATENT dataset</title>
      </sec>
      <sec id="sec-5-2">
        <title>5.2. Evaluation of prior-art retrieval in the CLEF-IP 2013 dataset</title>
        <p>
          In the CLEF-IP 2013, we followed the TREC-based guidelines provided by CLEF-IP [
          <xref ref-type="bibr" rid="ref33">33</xref>
          ], using TRecTools
[
          <xref ref-type="bibr" rid="ref38">38</xref>
          ] to calculate Precision and Recall at various cut-of levels (@5, @10 and @30), as well as MAP@100.
        </p>
        <p>Table 3 reports the retrieval results of conventional query strategies, where entire patent sections
commonly used by professionals are employed verbatim as queries. We observe that queries formulated
using the claims text achieved the best retrieval results. This outcome is largely attributed to the fact
that the claims text was also used for generating the corpus embeddings (using GTE as the embedding
model), thereby ensuring a higher degree of semantic alignment between the query and the indexed
documents. While extracting patent segments for query formulation could ofer valuable insights, it
was not feasible to implement this approach efectively, since the description sub-sections were not
consistently identifiable across all 24 topics.</p>
        <p>Table 4 presents the retrieval results when using automated summaries as query inputs. These
summaries were generated using various summarization methods and diferent patent sections as input
sources. The results demonstrate that queries formulated with these automated summaries consistently
outperform those based on standard patent sections. Overall, the default summaries generated by the
BigBird model although outperform the standard patent sections, they were found to be insuficient in
capturing the breadth of important patent content compared to longer summaries generated by the
adjusted BigBird model, which achieve best retrieval performance.</p>
        <p>Then comes the SBERT-based summaries, particularly when the description text was used as input,
or the default BigBird when using the claims text. Although the BERT- and SBERT-based summaries
achieved good retrieval scores, especially when produced from the description, they often retained
much of the original text without adequately condensing it, which is crucial for overcoming the token
limitations of LLMs in text retrieval tasks.</p>
      </sec>
      <sec id="sec-5-3">
        <title>5.3. Evaluation of prior-art in the USPTO dataset</title>
        <p>In the USPTO dataset, we followed the USPTO Kaggle competition guidelines, computing MAP@50 as
the primary metric. For Precision and Recall, we applied the same cut-of levels used in the CLEF-IP
2013 evaluation (@5, @10 and @30) to ensure consistency and comparability across datasets.</p>
        <p>Table 5 presents retrieval results based on traditional patent sections commonly used as queries, such
as the abstract, claims, and description. It also includes results from using high-value segments, such as
the summary segment, brief description, and first claim, individually and in combination. Similarly,
both queries and the corpus were represented using the GTE embedding model.</p>
        <p>Interestingly, the queries formulated by high-value segments (e.g., brief description or summary
segment/brief description and first claim), consistently outperform those based on conventional patent
sections. This underscores the importance of targeted content selection in enhancing retrieval
efectiveness.</p>
        <p>Table 6, on the other hand, reports retrieval performance when automated summaries are used
as queries. Then, we observe that depending on the summarization method and input, automated
summaries can significantly outperform their respective original patent sections.</p>
        <p>Avg. Words
109
982
6,962
26.31%
27.72%
23.89%
14.58%
15.83%
12.50%
35.80%
36.40%
28.12%</p>
        <p>In particular, the adjusted BigBird model, which generates longer summaries of approximately 250-300
words compared to the default version, outperforms the default BigBird model in retrieval performance.
Furthermore, it achieves results that are comparable to, or slightly better than, those obtained using the
simpler query formulation techniques outlined in Table 5. Notably, this approach demonstrates strong
eficiency, as it achieves similar retrieval performance using generated summaries that are substantially
more concise than the original patent sections.</p>
        <p>Regarding extractive models, queries generated using SBERT summaries based on the description
text achieved the highest retrieval scores across all metrics. In contrast, queries generated from the
claims or brief description text performed worse than those using the original texts. Interestingly,
despite their strong retrieval performance, SBERT-generated summaries, averaging 807 words. This
Claims
Description</p>
        <p>BERT
SBERT
BigBird*
BigBird**
BERT
SBERT
BigBird*
BigBird**</p>
        <p>BigBirdFT</p>
        <p>Note: *: Pre-trained BigBird (default), **: Adjusted pre-trained BigBird, FT: Fine-tuned BigBird
29.38%
26.60%
28.93%
32.12%
28.56%
28.93%
26.85%
30.73%
25.03%
46.92%
43.91%
49.94%
53.07%
49.63%
51.33%
48.31%
48.13%
48.10%
contrast highlights the importance of aligning summary generation with its intended purpose, whether
to enhance readability and support human assessment, or to optimize performance in downstream
tasks such as prior-art retrieval and classification.
25.00%
23.33%
24.17%
27.50%
15.00%
15.00%
15.83%
18.33%
15.83%
15.42%
14.17%
16.67%
13.75%
51.18%
60.60%
62.54%
63.94%
60.34%
63.73%
64.18%
64.36%
65.84%
49.57%
62.60%
57.74%
59.41%
49.48%
60.83%
51.71%
44.77%
52.82%
55.16%
56.42%
53.52%
56.62%
56.91%
57.71%
58.78%
43.57%
55.39%
51.25%
52.57%
43.71%
54.09%
45.41%
6.81%
6.53%
7.08%
7.78%
6.94%
7.08%
6.81%
6.94%
6.53%
32.84%
37.88%
40.06%
41.63%
39.52%
41.34%
41.89%
42.91%
43.68%
31.87%
40.82%
38.06%
39.01%
31.98%
39.88%
33.68%</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>6. Discussion and conclusion</title>
      <p>Overall, the findings of this study are promising, demonstrating that patent retrieval benefits from
using targeted patent segments (when detectable) and automated summaries as queries, compared
to relying solely on traditional sections typically employed by patent professionals, such as abstract,
descriptions or claims. Across both prior-art datasets, CLEF-IP and USPTO, automated summaries
consistently outperformed conventional query inputs. Among the summarization methods and input
configurations evaluated, the adjusted BigBird model using claims as input and the SBERT model
applied to the description section emerged as the most efective abstractive and extractive approaches,
respectively, yielding the highest retrieval performance across both datasets.</p>
      <p>Moreover, our initial experimental results support the hypothesis that summarization models can be
further adapted to produce comprehensive and contextually relevant summaries, although confirming
this required extensive validation. This approach presents a promising direction for future advancements
in patent summarization.</p>
      <p>
        This work was initially motivated by our participation in the European Patent Ofice’s (EPO) CodeFest
2024 competition [
        <xref ref-type="bibr" rid="ref39">39</xref>
        ], where it was selected as one of the top six finalists. Building on this foundation,
we aim to further advance our research on patent segmentation and summarization techniques by
evaluating their impact on patent retrieval performance across additional prior art test collections.
Additionally, exploring alternative corpus representations and integrating additional retrieval methods
will be key areas of focus in future work.
      </p>
    </sec>
    <sec id="sec-7">
      <title>Acknowledgments</title>
      <p>This research work was supported by the Hellenic Foundation for Research and Innovation (HFRI)
under the HFRI PhD Fellowship grant (Fellowship Number: 10695).</p>
    </sec>
    <sec id="sec-8">
      <title>Declaration on Generative AI</title>
      <p>During the preparation of this work, the authors used GPT-4 in order to: Grammar and spelling check
and paraphrase and reword. After using these tools/services, the authors reviewed and edited the
content as needed and take full responsibility for the publication’s content.
The code used for running the experiments of this article will become available in the a GitHub
repository.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>M.</given-names>
            <surname>Lupu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Mayer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Tait</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. J.</given-names>
            <surname>Trippe</surname>
          </string-name>
          (Eds.),
          <source>Current Challenges in Patent Information Retrieval</source>
          , Springer,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>Y.-H.</given-names>
            <surname>Tseng</surname>
          </string-name>
          ,
          <string-name>
            <surname>C.-J. Lin</surname>
            ,
            <given-names>Y.-I. Lin</given-names>
          </string-name>
          ,
          <article-title>Text mining techniques for patent analysis</article-title>
          ,
          <source>Information Processing &amp; Management</source>
          <volume>43</volume>
          (
          <year>2007</year>
          )
          <fpage>1216</fpage>
          -
          <lpage>1247</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>S.</given-names>
            <surname>Bashir</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Rauber</surname>
          </string-name>
          ,
          <article-title>Improving retrievability of patents in prior-art search</article-title>
          ,
          <source>in: Proceedings of the European Conference on Information Retrieval (ECIR)</source>
          ,
          <year>2010</year>
          , pp.
          <fpage>457</fpage>
          -
          <lpage>470</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>A.</given-names>
            <surname>Ali</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Tufail</surname>
          </string-name>
          , L.
          <string-name>
            <surname>C. De Silva</surname>
            ,
            <given-names>P. E.</given-names>
          </string-name>
          <string-name>
            <surname>Abas</surname>
          </string-name>
          ,
          <article-title>Innovating patent retrieval: A comprehensive review of techniques, trends, and challenges in prior art searches</article-title>
          ,
          <source>Applied System Innovation</source>
          <volume>7</volume>
          (
          <year>2024</year>
          )
          <fpage>91</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>S.</given-names>
            <surname>Cetintas</surname>
          </string-name>
          , L. Si,
          <article-title>Efective query generation and postprocessing strategies for prior art patent search</article-title>
          ,
          <source>Journal of the American Society for Information Science and Technology</source>
          <volume>63</volume>
          (
          <year>2012</year>
          )
          <fpage>512</fpage>
          -
          <lpage>527</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>V.</given-names>
            <surname>Stamatis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Salampasis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Diamantaras</surname>
          </string-name>
          ,
          <article-title>A novel re-ranking architecture for patent search</article-title>
          ,
          <source>World Patent Information</source>
          <volume>78</volume>
          (
          <year>2024</year>
          )
          <fpage>102282</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>S.</given-names>
            <surname>Althammer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Hofstätter</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Hanbury</surname>
          </string-name>
          ,
          <article-title>Cross-domain retrieval in the legal and patent domains: a reproducibility study</article-title>
          ,
          <source>in: Proceedings of the European Conference on Information Retrieval (ECIR)</source>
          ,
          <year>2021</year>
          , pp.
          <fpage>3</fpage>
          -
          <lpage>17</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>P.</given-names>
            <surname>Mahdabi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Keikha</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Gerani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Landoni</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Crestani</surname>
          </string-name>
          ,
          <article-title>Building queries for prior-art search</article-title>
          ,
          <source>in: Proceedings of the Information Retrieval Facility Conference (IRFC)</source>
          ,
          <source>Multidisciplinary Information Retrieval</source>
          ,
          <year>2011</year>
          , pp.
          <fpage>3</fpage>
          -
          <lpage>15</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>P.</given-names>
            <surname>Mahdabi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Andersson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Hanbury</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Crestani</surname>
          </string-name>
          ,
          <article-title>Report on the clef-ip 2011 experiments: Exploring patent summarization</article-title>
          , in: CLEF Notebook Papers/Labs/Workshop,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>R.</given-names>
            <surname>Srebrovic</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Yonamine</surname>
          </string-name>
          ,
          <article-title>Leveraging the BERT Algorithm for Patents with TensorFlow and BigQuery</article-title>
          , White paper,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <surname>EPO</surname>
          </string-name>
          ,
          <article-title>Chapter iv: The european search report</article-title>
          , https://www.epo.org/en/legal/guide-epc/
          <year>2023</year>
          /ga_ c4_
          <article-title>2_4</article-title>
          .html,
          <year>2023</year>
          . [Online].
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>X.</given-names>
            <surname>Xue</surname>
          </string-name>
          , W. B.
          <string-name>
            <surname>Croft</surname>
          </string-name>
          ,
          <article-title>Transforming patents into prior-art queries</article-title>
          ,
          <source>in: Proceedings of the International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR)</source>
          ,
          <year>2009</year>
          , pp.
          <fpage>808</fpage>
          -
          <lpage>809</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>M.</given-names>
            <surname>Suzgun</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Melas-Kyriazi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Sarkar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. D.</given-names>
            <surname>Kominers</surname>
          </string-name>
          ,
          <string-name>
            <surname>S. Shieber,</surname>
          </string-name>
          <article-title>The harvard uspto patent dataset: A large-scale, well-structured, and multi-purpose corpus of patent applications</article-title>
          ,
          <source>Advances in Neural Information Processing Systems</source>
          <volume>36</volume>
          (
          <year>2024</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>S. C.</given-names>
            <surname>Pujari</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Mantiuk</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Giereth</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Strötgen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Friedrich</surname>
          </string-name>
          ,
          <article-title>Evaluating neural multi-field document representations for patent classification</article-title>
          ,
          <source>in: Proc. International Workshop on Bibliometric Enhanced Information Retrieval (BIR) co-located with ECIR</source>
          ,
          <year>2022</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>D.</given-names>
            <surname>Miller</surname>
          </string-name>
          ,
          <article-title>Leveraging bert for extractive text summarization on lectures</article-title>
          , arXiv preprint arXiv:
          <year>1906</year>
          .04165,
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>N.</given-names>
            <surname>Reimers</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Gurevych</surname>
          </string-name>
          ,
          <article-title>Sentence-bert: Sentence embeddings using siamese bert-networks</article-title>
          ,
          <source>in: Proc. EMNLP-IJCNLP</source>
          ,
          <year>2019</year>
          , pp.
          <fpage>3982</fpage>
          -
          <lpage>3992</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>S.</given-names>
            <surname>Abdul</surname>
          </string-name>
          <string-name>
            <surname>Samad</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Sushma</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G. Bharathi</given-names>
            <surname>Mohan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Samuji</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Repakula</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. R.</given-names>
            <surname>Kothamasu</surname>
          </string-name>
          ,
          <article-title>Advancing abstractive summarization: Evaluating gpt-2, bart</article-title>
          , t5
          <article-title>-small, and pegasus models with baseline in rouge and bleu metrics</article-title>
          ,
          <source>in: Proc. ICICDS</source>
          ,
          <year>2024</year>
          , pp.
          <fpage>119</fpage>
          -
          <lpage>131</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>J.</given-names>
            <surname>Ding</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Kolapudi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Pobbathi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Nguyen</surname>
          </string-name>
          ,
          <article-title>Quality evaluation of summarization models for patent documents</article-title>
          ,
          <source>in: Proc. QRS</source>
          ,
          <year>2023</year>
          , pp.
          <fpage>250</fpage>
          -
          <lpage>259</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>J.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Saleh</surname>
          </string-name>
          , P. Liu, Pegasus:
          <article-title>Pre-training with extracted gap-sentences for abstractive summarization</article-title>
          ,
          <source>in: Proc. ICML</source>
          ,
          <year>2020</year>
          , pp.
          <fpage>11328</fpage>
          -
          <lpage>11339</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>C.</given-names>
            <surname>Rafel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Shazeer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Roberts</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Narang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Matena</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. J.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <article-title>Exploring the limits of transfer learning with a unified text-to-text transformer</article-title>
          ,
          <source>Journal of Machine Learning Research</source>
          <volume>21</volume>
          (
          <year>2020</year>
          )
          <fpage>1</fpage>
          -
          <lpage>67</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>M.</given-names>
            <surname>Lewis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Goyal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Ghazvininejad</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Mohamed</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Levy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Stoyanov</surname>
          </string-name>
          , L. Zettlemoyer, Bart:
          <article-title>Denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension</article-title>
          ,
          <source>in: Proc. ACL</source>
          ,
          <year>2019</year>
          , pp.
          <fpage>7871</fpage>
          -
          <lpage>7880</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>N. F.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Lin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Hewitt</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Paranjape</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Bevilacqua</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Petroni</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Liang</surname>
          </string-name>
          ,
          <article-title>Lost in the middle: How language models use long contexts</article-title>
          ,
          <source>Transactions of the Association for Computational Linguistics</source>
          <volume>12</volume>
          (
          <year>2024</year>
          )
          <fpage>157</fpage>
          -
          <lpage>173</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <given-names>S.</given-names>
            <surname>Moreno</surname>
          </string-name>
          ,
          <article-title>Transformers-Based Abstractive Summarization for the Generation of Patent Claims</article-title>
          ,
          <source>Ph.D. thesis, Politecnico di Torino</source>
          ,
          <year>2023</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [24]
          <string-name>
            <given-names>E.</given-names>
            <surname>Sharma</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <article-title>Bigpatent: A large-scale dataset for abstractive and coherent summarization</article-title>
          ,
          <source>in: Proc. ACL</source>
          ,
          <year>2019</year>
          , pp.
          <fpage>2204</fpage>
          -
          <lpage>2213</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          [25]
          <string-name>
            <given-names>I.</given-names>
            <surname>Sutskever</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Vinyals</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q. V.</given-names>
            <surname>Le</surname>
          </string-name>
          ,
          <article-title>Sequence to sequence learning with neural networks</article-title>
          ,
          <source>in: Advances in Neural Information Processing Systems</source>
          , volume
          <volume>27</volume>
          ,
          <year>2014</year>
          , pp.
          <fpage>3104</fpage>
          -
          <lpage>3112</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          [26]
          <string-name>
            <given-names>A.</given-names>
            <surname>See</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. J.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C. D.</given-names>
            <surname>Manning</surname>
          </string-name>
          ,
          <article-title>Get to the point: Summarization with pointer-generator networks</article-title>
          ,
          <source>in: Proc. ACL</source>
          ,
          <year>2017</year>
          , pp.
          <fpage>1073</fpage>
          -
          <lpage>1083</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          [27]
          <string-name>
            <surname>Y.-C. Chen</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Bansal</surname>
          </string-name>
          ,
          <article-title>Fast abstractive summarization with reinforce-selected sentence rewriting</article-title>
          ,
          <source>in: Proc. ACL</source>
          ,
          <year>2018</year>
          , pp.
          <fpage>675</fpage>
          -
          <lpage>686</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          [28]
          <string-name>
            <given-names>M.</given-names>
            <surname>Zaheer</surname>
          </string-name>
          , G. Guruganesh,
          <string-name>
            <given-names>K. A.</given-names>
            <surname>Dubey</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Ainslie</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Alberti</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Ontanon</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Pham</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Ravula</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Yang</surname>
          </string-name>
          , et al.,
          <article-title>Bigbird: Transformers for longer sequences</article-title>
          ,
          <source>in: Advances in Neural Information Processing Systems</source>
          , volume
          <volume>33</volume>
          ,
          <year>2020</year>
          , pp.
          <fpage>17283</fpage>
          -
          <lpage>17297</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          [29]
          <string-name>
            <given-names>O.</given-names>
            <surname>Tonguz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Qin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Gu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H. H.</given-names>
            <surname>Moon</surname>
          </string-name>
          ,
          <article-title>Automating claim construction in patent applications: The cmumine dataset</article-title>
          ,
          <source>in: Proc. NLLP</source>
          ,
          <year>2021</year>
          , pp.
          <fpage>205</fpage>
          -
          <lpage>209</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          [30]
          <string-name>
            <given-names>J.</given-names>
            <surname>Achiam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Adler</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Agarwal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Ahmad</surname>
          </string-name>
          ,
          <string-name>
            <given-names>I.</given-names>
            <surname>Akkaya</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F. L.</given-names>
            <surname>Aleman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Almeida</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Altenschmidt</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Altman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Anadkat</surname>
          </string-name>
          , et al.,
          <source>Gpt-4 technical report, arXiv preprint arXiv:2303.08774</source>
          ,
          <year>2024</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref31">
        <mixed-citation>
          [31]
          <string-name>
            <given-names>A.</given-names>
            <surname>Grattafiori</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Dubey</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Jauhri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Pandey</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Kadian</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Al-Dahle</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Letman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Mathur</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Schelten</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Vaughan</surname>
          </string-name>
          , et al.,
          <source>The llama 3 herd of models, arXiv preprint arXiv:2407.21783</source>
          ,
          <year>2024</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref32">
        <mixed-citation>
          [32]
          <string-name>
            <given-names>L.</given-names>
            <surname>Jiang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. M.</given-names>
            <surname>Goetz</surname>
          </string-name>
          ,
          <article-title>Natural language processing in the patent domain: A survey</article-title>
          ,
          <source>Artificial Intelligence Review</source>
          <volume>58</volume>
          (
          <year>2025</year>
          )
          <fpage>214</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref33">
        <mixed-citation>
          [33]
          <string-name>
            <given-names>F.</given-names>
            <surname>Piroi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Lupu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Hanbury</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Zenz</surname>
          </string-name>
          , Clef-ip
          <year>2011</year>
          :
          <article-title>Retrieval in the intellectual property domain</article-title>
          , in: CLEF Notebook Papers/Labs/Workshop,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref34">
        <mixed-citation>
          [34]
          <string-name>
            <given-names>F.</given-names>
            <surname>Piroi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Lupu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Hanbury</surname>
          </string-name>
          ,
          <article-title>Passage retrieval starting from patent claims: A clef-ip 2013 task overview</article-title>
          ,
          <source>in: Working Notes for CLEF 2013 Conference</source>
          ,
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref35">
        <mixed-citation>
          [35]
          <string-name>
            <given-names>S.</given-names>
            <surname>Beliveau</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Cenkci</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Alcantara</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Frost</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Garret</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Reddy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Dane</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Cukierski</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Howard</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Demkin</surname>
          </string-name>
          ,
          <article-title>Uspto-explainable ai for patent professionals</article-title>
          ,
          <source>Kaggle</source>
          ,
          <year>2024</year>
          . Available: https://kaggle. com/competitions/uspto-explainable-ai.
        </mixed-citation>
      </ref>
      <ref id="ref36">
        <mixed-citation>
          [36]
          <string-name>
            <given-names>X.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Long</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Xie</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Dai</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Tang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Lin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Xie</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Huang</surname>
          </string-name>
          , et al.,
          <article-title>mgte: Generalized long-context text representation and reranking models for multilingual text retrieval</article-title>
          ,
          <source>in: Proc. EMNLP Industry Track</source>
          ,
          <year>2024</year>
          , pp.
          <fpage>1393</fpage>
          -
          <lpage>1412</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref37">
        <mixed-citation>
          [37]
          <string-name>
            <surname>C.-Y. Lin</surname>
          </string-name>
          ,
          <article-title>Rouge: A package for automatic evaluation of summaries</article-title>
          , in: Text Summarization Branches Out,
          <year>2004</year>
          , pp.
          <fpage>74</fpage>
          -
          <lpage>81</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref38">
        <mixed-citation>
          [38]
          <string-name>
            <given-names>J.</given-names>
            <surname>Palotti</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Scells</surname>
          </string-name>
          , G. Zuccon,
          <string-name>
            <surname>Trectools:</surname>
          </string-name>
          <article-title>An open-source python library for information retrieval practitioners involved in trec-like campaigns</article-title>
          ,
          <source>in: Proc. SIGIR</source>
          ,
          <year>2019</year>
          , pp.
          <fpage>1325</fpage>
          -
          <lpage>1328</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref39">
        <mixed-citation>
          [39]
          <string-name>
            <surname>EPO</surname>
          </string-name>
          ,
          <source>Codefest 2024 on generative ai</source>
          ,
          <year>2024</year>
          . Available: https://www.epo.org/en/news-events/ in-focus/codefest/codefest-2024
          <string-name>
            <surname>-</surname>
          </string-name>
          generative-ai.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>