<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>RAG with Knowledge Structures</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Tolga Tel</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Goethe University</institution>
          ,
          <addr-line>Robert-Mayer-Straße 10, Frankfurt, 60325</addr-line>
          ,
          <country country="DE">Germany</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>This project explores methods to enhance the accuracy and reliability of Large Language Models (LLMs), which are prone to factual inaccuracies and hallucinations. The focus lies on improving accuracy by integrating Case-Based Reasoning and to address operational eficiency with pre- and post-processing techniques. Additionally new techniques for retrieving, adapting and retaining information will be researched. PCWrEooUrckResehdoinpgs ISSNc1e6u1r-3w-0s0.o7r3g</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Process-oriented CBR</kwd>
        <kwd>Knowledge Structures</kwd>
        <kwd>Guided Generation</kwd>
        <kwd>Retrieval Augmented Generation</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Language serves as a cornerstone for human communication and self-expression, and it also plays a
crucial role in how humans interact with machines. The increasing need for machines to efectively handle
complex language tasks, such as translation, summarization, information retrieval, and conversational
interactions, has driven the demand for generalized language models. Recent advances in language
modeling, primarily driven by the development of transformer architectures, increased computational
power, and the availability of massive training datasets, have led to significant breakthroughs. These
advancements have revolutionized the field by enabling the creation of Large Language Models (LLMs)
that can achieve near-human performance on a wide range of tasks. LLMs have emerged as
state-ofthe-art AI systems capable of processing and generating human-like text, demonstrating impressive
abilities in coherent communication and generalization across various tasks [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ].
      </p>
      <p>
        This PhD research focuses on enhancing the performance and eficiency of LLMs through improved
information retrieval and processing techniques. While LLMs have shown remarkable capabilities in
various language tasks, they often struggle with knowledge-intensive tasks and can generate incorrect
information (hallucinations) [
        <xref ref-type="bibr" rid="ref2 ref3">2, 3</xref>
        ]. Retrieval-Augmented Generation (RAG) addresses this by
integrating external knowledge bases. RAG retrieves relevant document chunks to help LLMs generate more
accurate responses, reducing factual errors [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ].
      </p>
      <p>
        A basic RAG approach involves: 1) Indexing: extracting, segmenting, embedding, and storing data
chunks in a vector database ("Retrieve-Read" [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]). 2) Retrieval: embedding the user query and finding
similar chunks in the database. 3) Generation: combining the query and retrieved context into a prompt
for the LLM.
      </p>
      <p>
        This naive approach faces challenges like imprecise retrieval and the potential for hallucination,
irrelevant, toxic, or biased outputs. Efectively integrating retrieved information is also dificult [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ].
Advanced RAG systems focus on enhancing retrieval quality beyond the naive approach. They employ
a multifaceted strategy that encompasses pre-retrieval and post-retrieval techniques [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. Pre-retrieval
focuses on optimizing the indexing structure and the original query. This involves enhancing the quality
of the indexed content by refining data granularity, optimizing index structures, adding metadata, and
aligning data for optimal retrieval. Simultaneously, the original user query is refined to make it clearer
and more suitable for the retrieval task. In [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] a framework for query rewriting was implemented and
in [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] abstraction is used as a pre-retrieval method.
      </p>
      <p>
        This work is grounded in the principles of Case-Based Reasoning, a powerful problem-solving paradigm
that leverages past experiences to address new challenges. Case-based reasoning has been formalized
as a four-step process [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]:
Retrieval involves finding the most similar case in the case base to the new problem. This is like
searching for information, where the new problem guides the search within the case base. Reuse
involves proposing a solution for the new problem based on the solution of the retrieved case. If the
cases are identical, reuse is straightforward. However, if they difer, adaptation is necessary. Revise
evaluates the proposed solution. This can involve real-world testing or simulation. Revision aims to
confirm the solution’s applicability and may involve addressing unforeseen issues. Retain updates
the case base by adding the new (learned) case, enabling future problem-solving. This step involves
deciding whether to retain all solutions or only actual cases.
      </p>
      <p>
        Knowledge Graphs (KGs) represent knowledge as nodes and edges [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ], ofering concise expert
information for RAG [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ]. While KG embeddings face sparsity [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ], textual graphs add context. GraphRAG uses
both. KG Question-Answering (KGQA) excels at factual questions but struggles with open-ended ones
[
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]. RAG can provide context to KGQA but lacks broader semantic understanding. Graph RAG aims
to address this by searching KGs, though subgraph identification is hard [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ]. "G-Retriever" [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ] uses
LLMs and GNNs for natural language interaction with graphs. Unifying LLMs and KGs [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ] aims for
more robust AI. RAG in LLMs has limitations [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ], requiring careful implementation.
The consortium OMG has established the standard BPMN [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ] for modeling languages. BPMN ofers a
standardized graphical notation for visualizing business processes. Its primary goal is to bridge the
gap between business and technical users, providing a notation that is both understandable to business
analysts and precise enough for technical developers.
      </p>
      <p>This flowchart-like notation empowers stakeholders to design, manage, and realize business processes
eficiently. Its independence from specific implementation environments ensures flexibility and
adaptability.</p>
      <p>
        Previous work with BPMN-models is found in [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ], where the use of Case-Based Reasoning (CBR)
within a Retrieval Augmented Generation (RAG) system to generate accessible explanations of business
process models is investigated. This research introduces a novel application of CBR for process-oriented
tasks.
      </p>
      <p>
        In [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ] the potential impact of Generative AI on Business Process Management (BPM) is discussed.
Generative AI can significantly impact BPM by automating routine tasks, improving customer and
employee satisfaction, and uncovering new opportunities for process improvement and redesign.
Another work regarding BPMN can be found in [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ], where natural language process description are
used to generate code. The "Tasks-Model-Extractor" is given a textual-process description as an input
and uses an LLM to extract the tasks and the control flow of the process and generates the model
representation.
      </p>
    </sec>
    <sec id="sec-2">
      <title>2. Research Plan</title>
      <p>The first approaches in this thesis can be assigned to the process-oriented CBR research. Here the idea
is to work with Business Process Models and improve RAG systems with eficient pre-processing.</p>
      <sec id="sec-2-1">
        <title>2.1. Research Objectives</title>
        <p>This research primarily focuses on improving the performance of Retrieval Augmented Generation
(RAG) systems, particularly in the context of Business Process Models. A core objective is to improve
the retrieval component of RAG by minimizing redundant information, accurately predicting necessary
information, and optimizing data structures for eficient retrieval. Additionally, this work aims to
advance the automated generation and evaluation of natural language descriptions from BPMN models,
including developing robust quality assessment mechanisms and error correction techniques.
Furthermore, this research explores several related objectives to improve the overall eficiency and
reliability of language models and knowledge integration. This includes investigating methods for
automatically extracting and structuring knowledge from text into knowledge graphs, and evaluating
the efectiveness of integrating these knowledge graphs into the RAG retrieval process to mitigate
hallucinations. Finally, it examines the feasibility of fully automated evaluation metrics for LLM outputs
and how CBR principles can be leveraged to enhance RAG performance by learning from past retrieval
and generation experiences.</p>
      </sec>
      <sec id="sec-2-2">
        <title>2.2. Approach / Methodology</title>
        <p>Latest Work: A paper on using CBR generated explanations for business process models by using clever
pre-processing to improve the accuracy while increasing the scalability of the input size, which has
been accepted to ICCBR 2025.</p>
        <p>LLMs Used: Diferent iterations of Llama 3 and Mistral were used in experiments leading to the paper
mentioned above. The focus lied on smaller models, which can be run locally on a conventional
computer to ensure accessibility and easy reproducibility of any result generated.</p>
        <p>Prompt Engineering: Prompt phrasing was refined through pre-testing, with the best-performing versions
used in the final experiment. These prompts are likely to be reused or adapted in future studies, with
domain-specific modifications.</p>
        <p>Construction of a Case-Base: For the experiments in the paper, the case-base was written by hand using
parts of business process models and writing a corresponding descriptive text.</p>
        <p>Retrieval of cases: In pre-testing, diferent retrieval strategies were used. While embedding-based
similarities did not satisfy the needs for our experiments, because cases were too similar and only a
small selection of keywords decided the usefulness of a case, another metric was used in the paper. A
taxonomy-based metric considering keywords to ensure retrieving a relevant case for the RAG system.</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Progress Summary</title>
      <p>ICCBR 2025: In my first and only submission, I presented a novel approach to addressing the challenges
associated with the description of large business process models. Recognizing the inherent complexity
and scale of these models, which often hinder eficient analysis, I developed a methodology that
helps RAG techniques. A core component of this methodology is the implementation of eficient
preprocessing strategies. These strategies are crucial for transforming the raw business process model data
into a format that is both useful for efective retrieval and the generation of coherent and contextually
relevant descriptions.</p>
    </sec>
    <sec id="sec-4">
      <title>4. Conclusion and Future Work</title>
      <p>Building upon the foundational work presented in my submitted paper, my research trajectory focuses
on a significant refinement and enhancement of the proposed methodology. Specifically, a key objective
is to elevate the precision and eficacy of the pre-processing stage. To achieve this, I intend to integrate
more sophisticated algorithmic approaches, moving beyond the initial strategies. This includes working
with a wider range of business process models, that are not well-modeled.</p>
      <p>Furthermore, a significant advancement will be the implementation of an automated evaluation
framework for the generated textual descriptions. This framework will move beyond subjective human
assessments and introduce quantifiable metrics to rigorously assess the quality of the generated outputs.
I plan to utilize Knowledge Graphs (KGs) as an information repository to develop novel retrieval
methods. This approach aims to significantly enhance the relevance of retrieved information. Furthermore, I
will explore the integration of the retain phase, dynamically expanding the KG with new knowledge, to
continuously improve retrieval accuracy.</p>
      <p>
        Possible research questions regarding my future work are:
1. How can we optimize the workload distribution between pre-processing algorithms and LLMs
for maximum eficiency and accuracy?
2. How can the principles of Case-based Reasoning be efectively leveraged to improve RAG
performance?
3. Can fully automated evaluation metrics reliably assess the quality of LLM outputs without the
need for human intervention?
4. What are the most efective methods for automatically extracting knowledge from textual
documents and representing it in a structured knowledge graph format?
5. Can the integration of knowledge graphs or textual graphs into the retrieval process of LLMs
efectively mitigate the risk of hallucinations?
Question 1 and 2 were already partly explored in [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ] but will reoccur multiple times in diferent settings
throughout this PhD research. Question 3 investigates the feasibility and reliability of fully automated
evaluation metrics for assessing the quality of outputs generated by LLMs. While human evaluation
remains the gold standard for assessing the quality of natural language generation, it is time-consuming,
expensive, and prone to subjectivity. This research aims to determine whether automated metrics
can provide reliable and objective assessments of LLM performance, reducing the reliance on human
evaluation.
      </p>
      <p>Question 4 focuses on the challenges of information extraction, including named entity recognition,
relation extraction, and event extraction, which are crucial for identifying and classifying key
components and relationships within the text. Efective knowledge representation plays an important role in
ensuring the accuracy, completeness, and consistency of the constructed knowledge graph.
Question 5 investigates the potential of integrating knowledge graphs or textual graphs into the retrieval
process of Large Language Models (LLMs) to efectively mitigate the risk of hallucinations.
Hallucinations, the generation of factually incorrect or nonsensical information by LLMs, pose a significant
challenge to their reliability and trustworthiness.</p>
      <p>By incorporating structured knowledge from knowledge graphs or the contextual relationships captured
in textual graphs, the retrieval process can be enhanced. This can involve using graph-based algorithms
to retrieve relevant information or improve context understanding.</p>
    </sec>
    <sec id="sec-5">
      <title>Declaration on Generative AI</title>
      <p>During the preparation of this work, the author(s) used Google Gemini, Grammarly for the purpose of:
Grammar, spelling check, minor Paraphrase and translations. After using this tool/service, the author(s)
reviewed and edited the content as needed and take(s) full responsibility for the publication’s content.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>H.</given-names>
            <surname>Naveed</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. U.</given-names>
            <surname>Khan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Qiu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Saqib</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Anwar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Usman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Akhtar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Barnes</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Mian</surname>
          </string-name>
          ,
          <article-title>A comprehensive overview of large language models</article-title>
          ,
          <source>arXiv preprint arXiv:2307.06435</source>
          (
          <year>2023</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>N.</given-names>
            <surname>Kandpal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Deng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Roberts</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Wallace</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Rafel</surname>
          </string-name>
          ,
          <article-title>Large language models struggle to learn long-tail knowledge</article-title>
          ,
          <source>in: International Conference on Machine Learning, PMLR</source>
          ,
          <year>2023</year>
          , pp.
          <fpage>15696</fpage>
          -
          <lpage>15707</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Cui</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Cai</surname>
          </string-name>
          , L. Liu,
          <string-name>
            <given-names>T.</given-names>
            <surname>Fu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Huang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Zhao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Chen</surname>
          </string-name>
          , et al.,
          <article-title>Siren's song in the ai ocean: a survey on hallucination in large language models</article-title>
          ,
          <source>arXiv preprint arXiv:2309.01219</source>
          (
          <year>2023</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Gao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Xiong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Gao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Jia</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Pan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Bi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Dai</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Sun</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <article-title>Retrieval-augmented generation for large language models: A survey</article-title>
          ,
          <source>arXiv preprint arXiv:2312.10997</source>
          (
          <year>2023</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>X.</given-names>
            <surname>Ma</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Gong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>He</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Zhao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Duan</surname>
          </string-name>
          ,
          <article-title>Query rewriting in retrieval-augmented large language models</article-title>
          , in: H.
          <string-name>
            <surname>Bouamor</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Pino</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          Bali (Eds.),
          <source>Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing</source>
          , Association for Computational Linguistics, Singapore,
          <year>2023</year>
          , pp.
          <fpage>5303</fpage>
          -
          <lpage>5315</lpage>
          . URL: https://aclanthology.org/
          <year>2023</year>
          .emnlp-main.
          <volume>322</volume>
          /. doi:
          <volume>10</volume>
          .18653/v1/
          <year>2023</year>
          .emnlp-main.
          <volume>322</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>W.</given-names>
            <surname>Peng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Jiang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Ou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Zeng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Xu</surname>
          </string-name>
          , E. Chen,
          <article-title>Large language model based long-tail query rewriting in taobao search</article-title>
          ,
          <source>in: Companion Proceedings of the ACM on Web Conference</source>
          <year>2024</year>
          ,
          <year>2024</year>
          , pp.
          <fpage>20</fpage>
          -
          <lpage>28</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>H. S.</given-names>
            <surname>Zheng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Mishra</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Chen</surname>
          </string-name>
          , H.-T. Cheng,
          <string-name>
            <given-names>E. H.</given-names>
            <surname>Chi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q. V.</given-names>
            <surname>Le</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <article-title>Take a step back: Evoking reasoning via abstraction in large language models</article-title>
          ,
          <source>arXiv preprint arXiv:2310.06117</source>
          (
          <year>2023</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>M. M.</given-names>
            <surname>Richter</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. O.</given-names>
            <surname>Weber</surname>
          </string-name>
          ,
          <source>Case-based reasoning</source>
          , Springer,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>T. T.</given-names>
            <surname>Procko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Ochoa</surname>
          </string-name>
          ,
          <article-title>Graph retrieval-augmented generation for large language models: A survey</article-title>
          , in: 2024 Conference on AI, Science, Engineering, and
          <string-name>
            <surname>Technology</surname>
          </string-name>
          (AIxSET), IEEE,
          <year>2024</year>
          , pp.
          <fpage>166</fpage>
          -
          <lpage>169</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>H.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. O.</given-names>
            <surname>Shafiq</surname>
          </string-name>
          ,
          <article-title>Triple-aware reasoning: A retrieval-augmented generation approach for enhancing question-answering tasks with knowledge graphs and large language models</article-title>
          ,
          <source>in: The 37th Canadian Conference on Artificial Intelligence</source>
          ,
          <year>2024</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Hu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Lei</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Pan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Ling</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Zhao</surname>
          </string-name>
          ,
          <article-title>Grag: Graph retrieval-augmented generation</article-title>
          ,
          <source>arXiv preprint arXiv:2405.16506</source>
          (
          <year>2024</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>X.</given-names>
            <surname>He</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Tian</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Sun</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N. V.</given-names>
            <surname>Chawla</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Laurent</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>LeCun</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Bresson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Hooi</surname>
          </string-name>
          , G-retriever:
          <article-title>Retrieval-augmented generation for textual graph understanding and question answering</article-title>
          ,
          <source>arXiv preprint arXiv:2402.07630</source>
          (
          <year>2024</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>S.</given-names>
            <surname>Pan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Luo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <article-title>Unifying large language models and knowledge graphs: A roadmap, IEEE Transactions on Knowledge and Data Engineering (</article-title>
          <year>2024</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>J.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Lin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Han</surname>
          </string-name>
          ,
          <string-name>
            <surname>L</surname>
          </string-name>
          . Sun,
          <article-title>Benchmarking large language models in retrieval-augmented generation</article-title>
          ,
          <source>in: Proceedings of the AAAI Conference on Artificial Intelligence</source>
          , volume
          <volume>38</volume>
          ,
          <year>2024</year>
          , pp.
          <fpage>17754</fpage>
          -
          <lpage>17762</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>O. M.</given-names>
            <surname>Group</surname>
          </string-name>
          ,
          <article-title>Business process model</article-title>
          and notation,
          <year>2014</year>
          . URL: https://www.omg.org/spec/BPMN/ 2.0.2#document-metadata.
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>M.</given-names>
            <surname>Minor</surname>
          </string-name>
          , E. Kaucher,
          <article-title>Retrieval augmented generation with llms for explaining business process models</article-title>
          ,
          <source>in: International Conference on Case-Based Reasoning</source>
          , Springer,
          <year>2024</year>
          , pp.
          <fpage>175</fpage>
          -
          <lpage>190</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>S.</given-names>
            <surname>Feuerriegel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Hartmann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Janiesch</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Zschech</surname>
          </string-name>
          , Generative ai,
          <source>Business &amp; Information Systems Engineering</source>
          <volume>66</volume>
          (
          <year>2024</year>
          )
          <fpage>111</fpage>
          -
          <lpage>126</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>F.</given-names>
            <surname>Monti</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Leotta</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Mangler</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Mecella</surname>
          </string-name>
          , S. Rinderle-Ma,
          <article-title>NL2ProcessOps: Towards LLM-Guided Code Generation for Process Execution</article-title>
          , in: A.
          <string-name>
            <surname>Marrella</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Resinas</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Jans</surname>
          </string-name>
          , M. Rosemann (Eds.),
          <source>Business Process Management Forum</source>
          , volume
          <volume>526</volume>
          , Springer Nature Switzerland, Cham,
          <year>2024</year>
          , pp.
          <fpage>127</fpage>
          -
          <lpage>143</lpage>
          . URL: https://link.springer.
          <source>com/10.1007/978-3-031-70418-5_8. doi:10.1007/ 978-3-031-70418-5_8, series Title: Lecture Notes in Business Information Processing.</source>
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>T.</given-names>
            <surname>Tel</surname>
          </string-name>
          ,
          <string-name>
            <surname>M.</surname>
          </string-name>
          <article-title>Minor, Utilizing the structure of process models for guided generation of explanatory texts</article-title>
          ,
          <source>in: International Conference on Case-Based Reasoning</source>
          , Springer,
          <year>2025</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>