<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Chunk-Link: Context-aware chunk completion⋆</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Kenichirou Narita</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Satoshi Munakata</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Fujitsu Ltd.</institution>
          ,
          <addr-line>4-1-1 Kamikodanaka, Nakahara-ku, Kawasaki-shi, Kanagawa 211-8588</addr-line>
          ,
          <country country="JP">Japan</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>In context engineering, Retrieval Augmented Generation (RAG) is an essential technology for improving the reliability of generative AI. However, conventional Top-K vector search methods often face challenges, particularly when enumerated content (such as bulleted lists or other list-based information) spans multiple chunks. This often leads to incomplete retrieval and inaccurate answers due to missing contextual fragments. To address this, we propose a novel method, ”Chunk-Link.” It identifies enumeration relationships across chunks to comprehensively extract all necessary contextual fragments. Our evaluation shows that Chunk-Link significantly improves chunk extraction accuracy. This, in turn, leads to a higher overall answer quality and enables more precise information retrieval and answer generation.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Large Language Model</kwd>
        <kwd>Retrieval-Augmented Generation</kwd>
        <kwd>Context Engineering</kwd>
        <kwd>Multi-Chunk Information Handling</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
    </sec>
    <sec id="sec-2">
      <title>2. Related work</title>
      <p>Enhancing RAG performance, especially when information is distributed across multiple chunks, has
become a key research area in recent years. This chapter focuses on the challenge of ”Multi-Chunk
Information Handling” and reviews relevant prior research.</p>
      <sec id="sec-2-1">
        <title>2.1. Text Chunking Strategy</title>
        <p>
          Common chunking strategies include fixed-length chunking, recursive chunking, and semantic
chunking. Semantic chunking, in particular, analyzes the content of text and generates chunks based on
semantic boundaries and contextual continuity. However, it may require adjustments to account for
domain-specific knowledge and structure, and it can be challenging to balance the granularity of chunk
division, potentially failing to fully capture the structure and semantic coherence of a document [
          <xref ref-type="bibr" rid="ref1">1</xref>
          ].
        </p>
      </sec>
      <sec id="sec-2-2">
        <title>2.2. Multi-Chunk Information Handling</title>
        <p>
          To address cases where complex questions cannot be answered with a single chunk, Kwon et al.
proposed an approach that decomposes questions into multiple subqueries and searches for and integrates
chunks corresponding to each subquery [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ]. Approaches that dynamically change chunk sizes have
also been proposed. Techniques such as Parent Document Retrieval perform searches using small
chunks and provide larger context blocks to the LLM based on the results to prevent context loss.
Zhong et al. dynamically determine the optimal granularity (chunk size or level of grouping) of the
knowledge database based on the input query and provide more appropriate context to the LLM [
          <xref ref-type="bibr" rid="ref3">3</xref>
          ].
Hei et al. propose a two-stage search process to extract chunks that are not directly related to the query.
After extracting chunks highly related to the query, they combine documents related to the query with
the extracted chunks to further search for ”dynamically related documents” that may appear unrelated
but are essential for the answer [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ].
        </p>
        <p>These studies primarily focus on the relationship between input queries and chunks. However, there
is still insuficient focus on the relationship between chunks themselves, particularly on explicitly
distinguishing the ”enumeration relationship” between chunks and comprehensively extracting necessary
chunks based on such relationships.</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Problems and Challenges</title>
      <p>In documents such as guidelines and manuals, product lists and operating procedures are commonly
summarized in a list format. When querying an LLM about such list-based information, traditional
vector search often omits parts of the enumerated content from the generated response. Fig.1 illustrates a
common scenario where enumerated content spans multiple chunks (here, each chunk corresponds to
one page). The text relevant to the enumeration, as shown in the bounding box, is distributed across
several pages. When extracting chunks from such a document using the following query, conventional
vector search cannot correctly extract the necessary chunks:</p>
      <p>What specific expansion eforts will NIST undertake to enable breakthroughs in measurement, standards,
and process capabilities for the fabrication of next-generation semiconductors?</p>
      <p>This failure occurs because conventional search primarily focuses on the direct semantic relationship
between queries and individual chunks, and thus cannot detect the complex structural relationships
between chunks that constitute a complete enumeration. In order to provide adequate answers when
list-based information is required and spans multiple chunks, it is essential to proactively identify
these inter-chunk relationships. This allows for chunk retrieval that considers not only query-chunk
relevance but also the logical connections between chunks.</p>
    </sec>
    <sec id="sec-4">
      <title>4. Chunk-Link</title>
      <p>To address this issue, we propose ”Chunk-Link.” This method focuses on inter-chunk enumeration
relationships and integrates them into conventional chunk search. Chunk-Link maintains these
enumeration relationships as explicit inter-chunk relations, and by complementing conventional query-chunk
vector search, it expands the chunk search scope (Fig.2).</p>
      <sec id="sec-4-1">
        <title>4.1. Contextual Enumeration Analysis</title>
        <p>In this proposal, we focus on the context of the text when extracting enumerations. In documents such
as books, reports, and presentations, enumeration information is written after some kind of
introduction. We define sentences that promote reader understanding and clarify the existence of enumerations
presented later as ”enumeration definition sentences.” Enumeration definition sentences are not
limited to sentences in the main text; chapter and section titles can also serve this role. The following are
some examples.</p>
        <p>• The following
• Next
• The concrete solutions are three: A, B, and C.
• 2-1. Advantages of using manufacturer parts
( a n n o u n c e s t h a t s p e c i f i c b e n e f i t s w i l l be l i s t e d o r e x p l a i n e d i n
p a r a l l e l a f t e r w a r d s )</p>
        <p>Such enumerative definitions suggest that there are enumerative sentences that follow in the context,
and LLM can extract enumerative information related to them from enumerative definitions. In this
proposal, we extract enumerative relationships between chunks through the following steps.
1. Prepare the target chunk (any one of the chunks used in RAG) and the subsequent chunks
(chunks following the target chunk in the same document).
2. Extract the enumeration definition sentences from the target chunk(using LLM). (The prompt
used is provided in Appendix A).
3. Extract the enumeration sentences related to the enumeration definition sentences from the
subsequent chunks(using LLM).</p>
        <p>By focusing on enumeration preamble sentences within the context and extracting enumeration
relationships, a knowledge graph representing inter-chunk enumeration relationships can be designed.</p>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. Complementing RAG with Chunk-Link</title>
        <p>Chunk-Link operates as a complementary function for LLM-based RAG technology. In the knowledge
graph generated by Chunk-Link, entities represent chunks, and the edges between them denote
enumeration relationships. By correlating the chunks obtained during RAG’s retrieval phase with this
knowledge graph, it becomes possible to supplement the chunks retrieved by conventional RAG with
those having enumeration relationships.</p>
        <p>This ensures that LLMs receive all necessary and suficient chunks related to enumerations.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Experiment</title>
      <p>We evaluate the contribution to LLM’s answer accuracy for enumeration questions. The LLMs used for
extracting enumeration relationships with Chunk-Link are gpt-4o-mini and gemma-3-4b-it. OpenAI
API was used for generating the embedding vectors for conventional RAG. The LLM model used to
generate the final answer from the supplemented chunks and query was gpt-4o-mini.</p>
      <p>The experiment is conducted for each setting (Chunk-Link enabled/disabled) on the same dataset for
three independent runs, and their average values are reported. This ensures the stability and reliability
of the results. The statistical significance of the diferences in these average values is also evaluated,
and their reliability is discussed.</p>
      <sec id="sec-5-1">
        <title>5.1. Dataset</title>
        <p>
          To evaluate the answer accuracy of LLMs focused on enumeration problems, the authors created a
dataset that compiles enumeration problems. This dataset was created from text data extracted from
29 PDF files in Japanese and English, and consists of 186 queries. The documents were selected from the
RAG evaluation dataset allganize_rag_evaluation_dataset_ja[
          <xref ref-type="bibr" rid="ref5">5</xref>
          ] and public documents (such as
guidelines and reports published by the National Institute of Standards and Technology (NIST) and the UK
Government). An example of the dataset is shown in Table 1.
        </p>
        <p>When the created dataset was fed to an LLM (gpt-4o-mini), the accuracy using RAG was 0.504, and
when the chunks were intentionally specified as only correct chunks, the accuracy was 0.856. These
results indicate that this dataset can achieve high accuracy if the correct chunks can be extracted, but
the queries cannot maintain that accuracy using conventional vector search.</p>
      </sec>
      <sec id="sec-5-2">
        <title>5.2. Experimental Setup</title>
        <p>Compare cases where RAG is supplemented with Chunk-Link and cases where it is not. The
experimental parameters are shown in Table 2.</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>6. Results</title>
      <sec id="sec-6-1">
        <title>Values</title>
      </sec>
      <sec id="sec-6-2">
        <title>Comments</title>
        <p>The evaluation results are shown in Table 3. The efectiveness of Chunk-Link was evaluated from two
perspectives: chunk extraction accuracy and query response accuracy. The chunk search accuracy was
evaluated using recall, precision, and F1. When using Chunk-Link, both models showed an increase
of approximately 15% in recall and a decrease of approximately 5% in precision, indicating that the
enumeration relationships between chunks were extracted with high accuracy. As recall improved,
the response accuracy also increased in both models.</p>
        <p>The statistical significance of the measurement results was evaluated. Since the Recall diferences
of the chunks did not follow a normal distribution, the Wilcoxon signed-rank test was adopted for
the analysis. A significance level of α=0.05 was set. As a result, the Recall diferences were below
the significance level, indicating statistical significance. On the other hand, although the statistical
significance of Accuracy was also evaluated using McNemar’s test, the P-value was 0.190, and no
statistically significant diference was observed.</p>
        <p>Next, we analyze the impact of Chunk-Link on the level of detail in model responses through specific
success examples. We introduce how the example shown in Fig.1 was improved using Chunk-Link.
The response results with and without Chunk-Link are shown in Table 4. As can be seen from this
table, the predicted answers without Chunk-Link are limited to general explanations, while those with
Chunk-Link are confirmed to include highly detailed and specific information. In particular, the latter
answers list specific expansion eforts such as ”physical measurements for next-generation
microelectronics,” which is the result of Chunk-Link referencing more extensive and relevant information chunks.
This level of detail is further supported by the increase in ‘chunk num‘ from 5 to 6 and the significant
improvement in ‘recall‘ from 0.5 to 1.0. This enables more comprehensive and accurate information
provision in response to user inquiries.</p>
      </sec>
    </sec>
    <sec id="sec-7">
      <title>7. Limitations</title>
      <sec id="sec-7-1">
        <title>Chunk-Link abled</title>
        <p>NIST will expand
ongoing metrology
research programs to
enable breakthroughs
in measurement,
standards, and
process capabilities for
the fabrication of
next*generation
semiconductors.</p>
      </sec>
      <sec id="sec-7-2">
        <title>Dis- Chunk-Link Enabled</title>
        <p>NIST will expand ongoing metrology research programs to
enable breakthroughs in measurement, standards, and process
capabilities for the fabrication of next*generation
microelectronics. * Specific expansion eforts will include physical metrology
for next*generation microelectronics, computational metrology for
computationally intensive and data*driven aspects of
semiconductor metrology, virtualization and automation for semiconductor
manufacturing, measurement services such as standard reference
materials, calibration services, and standard reference data, and
measurement and documentary standards in areas of industry
priority including cybersecurity measures, metrics, and solutions.
6
1.0
0.333
Our evaluation relies on a small dataset of enumeration problems from specific document types (e.g.,
guidelines, manuals), limiting the reliability of answer accuracy and the statistical significance of
validation results. Furthermore, as Chunk-Link complements conventional vector search, its fundamental
retrieval accuracy is inherently dependent on query-vector similarity, potentially imposing a ceiling
on final answer accuracy.</p>
        <p>While Chunk-Link improves Recall, Precision decreases due to irrelevant enumeration relationships,
hindering overall answer accuracy. Filtering extracted enumeration relationships is essential.</p>
      </sec>
    </sec>
    <sec id="sec-8">
      <title>8. Conclusion</title>
      <p>In this paper, we propose ”Chunk-Link” to address the issue that conventional chunk search from
embedded vectors struggles to extract inter-chunk relationships. Chunk-Link extracts enumeration
relationships between chunks and complements vector search. We created a dataset to measure the answer
accuracy for enumeration questions and demonstrated that by complementing RAG with Chunk-Link,
the recall of chunks improves even when enumerated content is divided across chunks, enabling more
comprehensive and detailed answers.</p>
      <p>For future work, we will generalize to other structural relationships, such as contextual, referential,
and causal relationships, to expand Chunk-Link’s applicability and versatility.</p>
    </sec>
    <sec id="sec-9">
      <title>A. Prompts for LLM</title>
      <p>This appendix provides the prompts used for extracting enumerated definitions from the context for
our research. The prompts listed are excerpts of key parts, with output formats and examples omitted.</p>
      <p>Prompt to extract enumeration definition sentences
You are a text analysis assistant. Following the specifications below, accurately
and comprehensively extract all "Preamble Sentences" from the input text.
### 1. Definition and Role of "Preamble Sentence"
**A "Preamble Sentence"** refers to a description that meets any of the following
conditions and **explicitly announces that multiple specific pieces of information
or elements will be enumerated immediately after it, either in a list format
(e.g., bullet points or numbered lists) or in a clearly parallel structure.**
1. **Introductory/Announcing Type Description:**</p>
      <p>A short expression that **explicitly states within the sentence or at its end**
that multiple elements will follow immediately after it, either as **bullet
points (with leading symbols or numbers)**, or in a **clearly identifiable
parallel structure (e.g., "A, B, and C" or "first A, then B, finally C,").**
* **Characteristics:**</p>
      <p>This refers to a sentence that, within the text or at its end, is followed
by multiple elements (either as bullet points or enumerated in subsequent sentences).
* **Key Phrase Examples (detected in conjunction with clear subsequent enumeration):**
* "The following ~"
* "Next ~"
* "...are as follows."
* "In ..., three points are important: A, B, and C."
* Note: Expressions like "is" alone or "regarding ~" that merely indicate
a continuation of explanation are generally not considered key phrases.
2. **Chapter/Section Title:**</p>
      <p>A title placed at the beginning of a specific chapter or section in a document
structure, which **summarizes the specific content to be covered within that
entire chapter/section, and functions as an introduction or announcement for the
subsequent specific descriptions or subsections.**
* "2-1. Advantages of using manufacturer parts"
* "Chapter 3: Customer Experience Improvement Strategies"
### 2. Extraction Procedure
1. Parse the input text and segment it into sentences as much as possible, including
titles, body text, and bullet points. Then, assign a sequential number starting from
1 to each segmented sentence.
2. For each sentence/chapter/section title, determine if it is a "Preamble Sentence"
based strictly on the definition and extraction conditions. Special emphasis should
be placed on "whether multiple elements are clearly enumerated in bullet points or
a parallel structure immediately after it."
3. Format the output according to the output specification and output it in JSON format.
Following these rules, extract all preamble sentences from the document comprehensively.
### Input Text</p>
    </sec>
    <sec id="sec-10">
      <title>Acknowledgments</title>
      <p>This project would not have been possible without the immense cooperation and support of all project members
and colleagues, to whom I extend my heartfelt gratitude. We would especially like to thank Ms.Moriyama for
her valuable assistance in creating the dataset.</p>
    </sec>
    <sec id="sec-11">
      <title>Declaration on Generative AI</title>
      <p>During the preparation of this work, the author(s) used Gemini and Copilot in order to: Grammar and spelling
check. After using these tool(s)/service(s), the author(s) reviewed and edited the content as needed and take(s)
full responsibility for the publication’s content.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <surname>Multimodal</surname>
          </string-name>
          ,
          <article-title>Semantic chunking for rag: Better context, better results</article-title>
          , Multimodal.dev (
          <year>2024</year>
          ). https://www. multimodal.dev/post/semantic
          <article-title>-chunking-for-rag.</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>M.</given-names>
            <surname>Kwon</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Bang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Hwang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Jang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <article-title>A dynamic-selection-based, retrieval-augmented generation framework: Enhancing multi-document question-answering for commercial applications</article-title>
          ,
          <source>Electronics</source>
          <volume>14</volume>
          (
          <year>2025</year>
          )
          <article-title>659</article-title>
          . doi:
          <volume>10</volume>
          .3390/electronics14040659.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Zhong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Cui</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Qin</surname>
          </string-name>
          ,
          <article-title>Mix-of-granularity: Optimize the chunking granularity for retrieval-augmented generation</article-title>
          ,
          <source>in: Proceedings of the 31st International Conference on Computational Linguistics, Association for Computational Linguistics</source>
          ,
          <year>2025</year>
          , pp.
          <fpage>5756</fpage>
          -
          <lpage>5774</lpage>
          . URL: https://aclanthology.org/
          <year>2025</year>
          .coling-main.
          <volume>384</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Hei</surname>
          </string-name>
          , W. Liu,
          <string-name>
            <given-names>W.</given-names>
            <surname>Ou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Qiao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Jiao</surname>
          </string-name>
          , G. Song,
          <string-name>
            <given-names>T.</given-names>
            <surname>Tian</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Lin</surname>
          </string-name>
          ,
          <article-title>Dr-rag: Applying dynamic document relevance to retrieval-augmented generation for question-answering</article-title>
          ,
          <source>arXiv preprint arXiv:2406.07348</source>
          (
          <year>2024</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <surname>Allganize</surname>
            ,
            <given-names>RAG</given-names>
          </string-name>
          <string-name>
            <surname>Evaluation Dataset</surname>
            <given-names>JA</given-names>
          </string-name>
          ,
          <year>2024</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>