<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>U. Simsek, E. Kärle, K. Angele, E. Huaman, J. Opdenplatz, D. Sommer, J. Umbrich, D. Fensel,
A knowledge graph perspective on knowledge engineering, SN Computer Science</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <article-id pub-id-type="doi">10.1007/s42979-022-01429-x</article-id>
      <title-group>
        <article-title>Towards Eficient Field Service Engineering for Powertrains via LLM-generated Knowledge Graphs</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Prerna Juhlin</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Rana Hussein</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Nicolai Schoch</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>ABB AG Corporate Research Center Germany</institution>
          ,
          <addr-line>Kallstadter Strasse 1, D-68309 Mannheim</addr-line>
          ,
          <country country="DE">Germany</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2025</year>
      </pub-date>
      <volume>4</volume>
      <issue>2022</issue>
      <fpage>3</fpage>
      <lpage>5</lpage>
      <abstract>
        <p>A knowledge integration framework is presented for eficient field service engineering in the context of powertrains. The framework enables semantic integration of otherwise siloed data from diferent operations and maintenance system databases. Starting with an expert-curated and database schema-aligned ontology, and using LLMs, a knowledge graph is created from unstructured data sources. A comparison of two leading LLM-based approaches is provided for unstructured data integration in knowledge graphs, namely LangChain LLM Graph Transformer and Microsoft GraphRAG. We also suggest customizations for fine-tuning of the generation process. Besides eficient powertrain fault resolutions, potential applications include Root Cause Analysis (RCA), Failure Mode and Efects Analysis (FMEA), as well as prescriptive maintenance.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;powertrain</kwd>
        <kwd>field service engineering</kwd>
        <kwd>RCA</kwd>
        <kwd>FMEA</kwd>
        <kwd>prescriptive maintenance</kwd>
        <kwd>LLM Graph Transformer</kwd>
        <kwd>GraphRAG</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>for diferent FSE applications; Section 3 presents a comparison of leading LLM-based approaches for
unstructured data together with a comparative evaluation, findings and related customizations; Section
4 concludes the paper with remarks and a future outlook.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Powertrain Knowledge Integration Framework</title>
      <p>The overall powertrain knowledge integration architecture is organized in layers as shown in Figure 1.
Diferent information such as service instructions, installed base data, service history, customer requests,
maintenance history and operational data are available via the various specialized databases from
multiple vendors in ‘Powertrain Operations and Maintenance Systems’. Due to the siloed information
sources and heterogeneous data models, relations across information entities can easily be missed. Field
service engineers looking for information about a particular asset on site must refer to various databases
and manually find relevant interconnections, which tends to be a time-consuming and error-prone
process. As a result, services such as fault resolution and root cause analyses requiring integrated
information from diferent operations and maintenance systems are dificult to perform.</p>
      <p>
        The ‘Powertrain Knowledge Graph Layer’ consists of an ontology with TBox statements and a
knowledge graph with instances corresponding to ABox statements [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ], based on the underlying data
sources for powertrain. The ontology is created based on expert curation of classes and relationships,
which is subsequently aligned to type- and attribute-level information from diferent database schemas.
In particular, relationships across information entities not captured in any single database are possible
to represented in the ontology based on expert knowledge. For example, service cases in a CRM system
about a certain drive product family are linked via the drive fault code to recommended actions in the
corresponding drive service manual. The ontology is created using the W3C Resource Description
Framework (RDF) family of standards [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ], which allows semantic representations as well as the possibility
to process using LLMs, e.g., in Turtle syntax [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]. The ontology serves as a basis for constructing the
‘Powertrain Knowledge Graph’ using LLMs for unstructured data sources, as described in Section 3.
Due to the potential for errors in LLM outputs, which can vary based on selected approach as explained
in Section 3, the generated knowledge graph is reviewed by experts for quality assurance in a final step.
      </p>
      <p>
        Troubleshooting, resolution and maintenance are among the primary tasks of FSEs, see e.g., [
        <xref ref-type="bibr" rid="ref8 ref9">8, 9</xref>
        ].
Related use cases requiring integrated information are enabled based on the constructed knowledge
graph in combination with querying, e.g., using the RDF query language SPARQL [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ]:
1. Fault Resolution FSEs dealing with faulty equipment require assessing the current situation,
considering operational history and maintenance records, and performing relevant standard
checks and procedures. Based on the connected information stored in the knowledge graph, the
process can be guided through resolution step based on recommended actions from the service
manuals. In case of a novel resolution means not hereto recorded in the manuals, the fix can be
stored in the knowledge graph as future reference for future cases.
2. Root Cause Analysis and Failure Mode and Effects Analysis (FMEA) Successful
failure resolution typically also involves performing the Root Cause Analysis (RCA) for preventive
maintenance [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ]. The knowledge graph provides operational data, case history, maintenance
records and service instructions from manuals with relevant connections, enabling easier access
to information with their context for eficient investigations. The knowledge graph can also
enable design-phase FMEA used to anticipate failure in an early phase based on the integration of
history across operations and maintenance together with actual service resolutions and service
instructions, recorded for diferent powertrain instances. FSE feedback can be recorded in the
knowledge graph for a constantly up-to-date integrated knowledge base.
3. Powertrain Prescriptive Maintenance Beyond preventive maintenance, prescriptive
maintenance services could also be provided based on mapped future scenario possibilities with
related instructions or resolutions stored in the knowledge graph.
      </p>
    </sec>
    <sec id="sec-3">
      <title>3. LLM-based Approaches for Knowledge Graph Generation</title>
      <p>
        Field service engineers typically rely on service manuals to perform maintenance tasks. These manuals
are in the form of unstructured data and are usually complex containing domain specific concepts and
terminology. Capturing the meanings in such unstructured text and understanding the connections
between diferent entities is not a trivial task. Large Language Models (LLMs) have recently shown
significant capabilities in several natural language processing research areas. Among these areas is
converting unstructured data into entities and relationships to construct knowledge graphs for large
amounts of unstructured text, e.g., as investigated in [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ]. The LLM-based approaches are particularly
valuable when it comes to scaling the work of structuring large amounts of unstructured texts [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ]. A
well-known challenge is that LLM models are prone to hallucinations [
        <xref ref-type="bibr" rid="ref13 ref14">13, 14</xref>
        ], i.e. LLMs may fabricate
responses that are factually incorrect. As a result, the knowledge graph can contain an inaccurate
representation of service instructions which hinders the quality of maintenance activities.
      </p>
      <p>In this section, we explore diferent LLM techniques and frameworks to convert unstructured text into
knowledge graphs. We highlight our findings and present customizations to overcome some challenges.</p>
      <sec id="sec-3-1">
        <title>3.1. LLM Techniques</title>
        <p>We investigate three techniques to explore LLM capabilities in converting large amounts of unstructured
text to entities and relationships.</p>
        <sec id="sec-3-1-1">
          <title>3.1.1. Schema-less technique</title>
          <p>A schema-less technique is an unstructured approach that allows LLMs to build knowledge graphs
without relying on any pre-defined graph structure. Here, the input to the LLM is only the unstructured
text. The LLM subsequently generates a schema representing the underlying entities and relationships
and populates instances accordingly.</p>
        </sec>
        <sec id="sec-3-1-2">
          <title>3.1.2. Schema-based technique</title>
          <p>In a schema-based technique, a structured representation of the TBox statements in the form of an
ontology is provided to the LLM to act as a guide for capturing entities and relationships. The objective
of the LLM is to find triples in the text conforming with the given ontology schema. Here, the
expertcurated ontology aligned to database schemas in a second step, as shown in Figure 1, is provided.
Example nodes include fault codes, which are identifiers representing specific faults generated by a
drive or a motor as well as likely causes and instruction explaining how to resolve such faults. Examples
of relationships include linking a drive product family to a fault code, by specifying that a drive generates
a fault code and a fault code occurs due to a likely cause.</p>
        </sec>
        <sec id="sec-3-1-3">
          <title>3.1.3. Prompt design</title>
          <p>
            Providing additional prompt instructions is a widely used approach to provide clear guidance to the LLM
[
            <xref ref-type="bibr" rid="ref15 ref16">15, 16</xref>
            ]. This can further enhance the generation of more accurate responses and reduce hallucinations.
We investigate the benefit of providing such prompts by designing a set of detailed instructions that
contain contextual information and demonstrations of domain specific input and output examples. An
example of such instructions is including a fault code number representation in the entity text. We
provide examples to guide the LLM on how to handle unseen fault codes encountered in unstructured
text. Handling such domain specific details is not obvious without providing additional instructions.
Another example is instructing the LLM to include the full text of a sentence in the entity text as
opposed to summarizing or cutting the sentence. This is crucial in our case, as the text often contains
detailed instructions from a manual on how to address a particular drive fault.
          </p>
        </sec>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. LLM Frameworks</title>
        <p>
          We investigate two widely used LLM frameworks to generate knowledge graphs: LangChain LLM
Graph Transformer and Microsoft GraphRAG. To drive a fair comparison, we use GPT-4 model [
          <xref ref-type="bibr" rid="ref17">17</xref>
          ] for
both frameworks.
        </p>
        <sec id="sec-3-2-1">
          <title>3.2.1. LLM Graph Transformer</title>
          <p>
            LLM Graph Transformer [
            <xref ref-type="bibr" rid="ref18">18</xref>
            ] is a framework developed by Langchain [19] and Neo4j [20]. It extracts
structured information from text using LLMs. Text is split into several chunks according to a given
window size. LLM models are then used to extract entities and relationships from text chunks. The
extracted information is stored in Neo4j graph database to facilitate downstream RAG applications.
LLM Graph Transformer provides several customizations to tailor the LLM depending on a given use
case. First, it provides the ability to define a graph schema in a precise way, which provides guidance
and structure to the LLM. The graph schema is a set of entity types and relationships in the form of
triples: source node, relationship type, and target node. Second, the framework has the flexibility to use
a variety of LLM models. Third, a prompt with additional instructions can be passed to the LLM for
further customization.
          </p>
        </sec>
        <sec id="sec-3-2-2">
          <title>3.2.2. Microsoft GraphRAG</title>
          <p>Microsoft’s GraphRAG [21] extracts structured entities and relationships from natural language text
using LLMs. A graph index is built and an entity knowledge graph is derived. GraphRAG also leverages
global information in the graph by using community detection to cluster closely related communities
together. This is then used for summarization and RAG retrieval applications. This framework supports
providing an ontology to guide the structuring of the knowledge graph. Several prompts are provided
for extracting the graph, clustering communities, and performing local and global searches on the
knowledge graph.</p>
        </sec>
      </sec>
      <sec id="sec-3-3">
        <title>3.3. Findings</title>
        <p>Below we present our findings for the three LLM techniques used to construct the knowledge graph.
We also compare the two LLM frameworks using several aspects.</p>
        <sec id="sec-3-3-1">
          <title>3.3.1. Evaluating LLM techniques</title>
          <p>We evaluate the knowledge graph generated from unstructured text in manuals. The manuals contain a
description of faults encountered by drives, their likely causes, and resolution steps. Figure 2 shows
the schemas generated from LLM models when using each of the three LLM techniques. First, for the
schema-less technique, although the LLM has more freedom and flexibility to construct the knowledge
graph, we observe that it fails in capturing the underlying information in the unstructured text. This
leads to misrepresenting entities and relationships in the knowledge graph as well as ignoring relevant
concepts in the text, such as likely causes of drive faults. Second, in the schema-based technique, we
observe a more informative structure in the knowledge graph. The entities and relationships accurately
represent the domain information present in the unstructured text. However, we still notice the LLM
missing essential semantic connections present in the text. Finally, when providing detailed prompt
instructions in addition to the ontology, we observe an improvement in the quality of the generated
schema and instances. The LLM is able to accurately capture the required domain knowledge in the text
and form meaningful semantic connections. For example, fault code entity representation accurately
follows the provided prompt examples. This is crucial for service engineers to be able to accurately
recognize faults generated by drives. Also, when providing instructions to the LLM, full instructions
from the manuals are included in the instance graphs. This is important for efective diagnoses and
repair of faults.</p>
        </sec>
        <sec id="sec-3-3-2">
          <title>3.3.2. Evaluating LLM frameworks</title>
          <p>We compare both frameworks against several aspects in Table 1. The first aspect is stable runs. Both
frameworks use LLMs, as a consequence the graph generation process is non-deterministic which may
lead to variant output among diferent runs. Producing diferent output is problematic in real-world
scenarios where a consistent data generation workflow is required. To assess the magnitude of the
variation for both frameworks, we perform 10 runs using the same input and prompts. We examine the
consistency of the output in terms of the following: the total number of generated entities/relationships
and the number of generated entities per each type in the ontology schema. When using GraphRAG,
we notice consistent output across multiple runs. However, when using LLM Graph Transformer, a
substantial variation of results is observed. The second aspect of the comparison is related to the
schema-based technique. LLM Graph Transfomer ofers more control over specifying the ontology
structure. A detailed specification of relationship types and properties can be provided to guide the LLM.
Next, we notice both frameworks fail to semantically link and understand domain specific terminology.
Also, both frameworks are prone to hallucinations. LLM Graph Transformer exhibits a hallucination
rate of ∼ 15%, while for GraphRAG, the hallucination rate is ∼ 12%. In Section 3.4 we further discuss the
concept of hallucinations. Hence, summarizing, our preference based on scalability-relevant criteria
such as ‘Stable runs’ and ‘Triple hallucination avoidance’ is to use GraphRAG as it performs better
compared to the LLM Graph Transformer approach. Additionally, the ‘Provide summary descriptions’
and ‘Generate communities’ features can aid in the future for processing of the constructed knowledge
graph.</p>
        </sec>
      </sec>
      <sec id="sec-3-4">
        <title>3.4. Customizations</title>
        <sec id="sec-3-4-1">
          <title>3.4.1. Hallucinations</title>
          <p>LLM approaches are prone to hallucinations. We observe several examples of such behavior. First,
hallucinations exist in the form of new relationship types that do not exist in the ontology schema. This
leads to incorrect semantic linking between entities. Second, hallucinations exist in the form of triple
hallucinations. Although the type of relationship as well as the types of generated entities exist in the
ontology schema, the LLM fabricates and mixes false associations between entity types that are not
consistent with the domain knowledge. Next, we observe cases where the triple types and associations
follow the ontology schema, however, the generated instances are not present in the input data. Finally,
in some cases, entities are generated with no assigned types. In this case, the LLM fails to understand
the meaning of such entities. These examples are flagged in the graph for further domain expert review.</p>
        </sec>
        <sec id="sec-3-4-2">
          <title>3.4.2. Knowledge Graph verification and alignment to the ontology</title>
          <p>Diferent from hallucinations, LLMs may generate triples that are slightly deviated from the ontology
structure. For example, a relationship can be reversed between two entities. We adjust such
inconsistencies to ensure that the instances are aligned with the defined ontology concepts. The verification
step involves the experts towards the final ‘Expert-Reviewed Knowledge Graph’ of Figure 1.</p>
        </sec>
        <sec id="sec-3-4-3">
          <title>3.4.3. Mapping layer</title>
          <p>We provide the capability to store the output of the LLM, in the form of entities and relationships, in
diferent RDF-native and labeled property graph (LPG) databases. To achieve this, a mapping layer
is needed to transform the entities and relationships to diferent formats via the RDF representation,
supported widely by diferent LPG databases as mentioned in Section 2. Our mapping layer generates
queries using diferent graph query-programming languages to store and manipulate data in various
graph databases.</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Conclusion</title>
      <p>FSEs stand to benefit from integrated knowledge in a variety of ways for providing services in an eficient
manner based on up-to-date information. The described LLM-based Knowledge Graph generation based
on unstructured data sources together with related customizations is promising, particularly for scaling
knowledge graphs for powertrain knowledge integration. In the future, LLMs could also be utilized
for automatically suggesting ontology extensions for TBox alignments with database schemas, for an
initial reviewing step where LLMs act as a judge [22], as well as for enabling a chat-based user interface
with translation of natural language questions to graph queries for knowledge retrieval and response.
Beyond powertrain manufacturing, the proposed architecture and evaluation methodology could also
be applied to further verticals involving field service engineering such as productive robotics, oil and
gas, aerospace and healthcare.</p>
    </sec>
    <sec id="sec-5">
      <title>Declaration on Generative AI</title>
      <p>The author have not employed any Generative AI tools in creating the paper.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1] ABB Press, ABB Ability™
          <article-title>Digital Powertrain for eficient, safe</article-title>
          and reliable operations,
          <year>2019</year>
          . URL: https://new.abb.com/news/detail/17846/ abb-ability
          <article-title>-digital-powertrain-for-eficient-safe-and-reliable-operations.</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <surname>C. P</surname>
          </string-name>
          ,
          <article-title>Addressing the top 5 concerns of field service workers</article-title>
          ,
          <year>2023</year>
          . URL: https://librestream.com/ blog/addressing
          <article-title>-the-top-5-concerns-of-field-service-workers/.</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>W.</given-names>
            <surname>Schell</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Leoncio</surname>
          </string-name>
          ,
          <article-title>Building massive knowledge graphs using automated ETL pipelines</article-title>
          ,
          <year>2024</year>
          . URL: https://blog.metaphacts.
          <article-title>com/ building-massive-knowledge-graphs-using-automated-etl-pipelines.</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>B.</given-names>
            <surname>Stieger</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Aleksy</surname>
          </string-name>
          ,
          <article-title>Utilization of knowledge management for service business processes improvement</article-title>
          , in: 2009
          <source>International Multiconference on Computer Science and Information Technology</source>
          ,
          <year>2009</year>
          , pp.
          <fpage>171</fpage>
          -
          <lpage>175</lpage>
          . doi:
          <volume>10</volume>
          .1109/IMCSIT.
          <year>2009</year>
          .
          <volume>5352730</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <surname>Wikipedia</surname>
          </string-name>
          , Description logic,
          <year>2025</year>
          . URL: https://en.wikipedia.org/w/index.php?title=Description_ logic&amp;oldid=1283591955,
          <string-name>
            <surname>page Version</surname>
            <given-names>ID</given-names>
          </string-name>
          :
          <fpage>1283591955</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>World</given-names>
            <surname>Wide Web</surname>
          </string-name>
          <article-title>Consortium (W3C), Resource description framework (rdf) - semantic web standards</article-title>
          ,
          <year>2014</year>
          . URL: https://www.w3.org/RDF/.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>J.</given-names>
            <surname>Frey</surname>
          </string-name>
          , L.-P. Meyer,
          <string-name>
            <given-names>N.</given-names>
            <surname>Arndt</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Brei</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Bulert</surname>
          </string-name>
          ,
          <article-title>Benchmarking the abilities of large language models for rdf knowledge graph creation and comprehension: How well do llms speak turtle?</article-title>
          ,
          <source>in: CEUR Workshop Proceedings</source>
          , Athens, Greece,
          <year>2023</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>Teal</given-names>
            <surname>Labs</surname>
          </string-name>
          , Inc.,
          <source>What is a field service engineer?</source>
          ,
          <year>2025</year>
          . URL: https://www.tealhq.com/career-paths/ ifeld-service-engineer#:~:text=
          <source>What%20does%20a%20Field%20Service,rigorous%20demands% 20of%20various%20industries.</source>
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <surname>ORACLE</surname>
          </string-name>
          ,
          <article-title>Making sense of proactive maintenance in field service</article-title>
          ,
          <year>2021</year>
          . URL: https://www.oracle. com/a/ocom/docs/field-service
          <article-title>-proactive-maintenance</article-title>
          .pdf.
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <surname>W. W. W. C.</surname>
          </string-name>
          (
          <article-title>W3C), Sparql 1.1 query language</article-title>
          ,
          <year>2013</year>
          . URL: https://www.w3.org/TR/sparql11-query/.
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>A.</given-names>
            <surname>Parashar</surname>
          </string-name>
          ,
          <article-title>What is Root Cause Analysis (RCA) in Maintenance</article-title>
          and Why It's Important,
          <year>2025</year>
          . URL: https://www.fieldcircle.com/blog/root-cause
          <article-title>-analysis-in-maintenance/#:~:text=It's%20a% 20systematic%20process%20that,intervene%20to%20prevent%20the%20failure.</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>N.</given-names>
            <surname>Mihindukulasooriya</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Tiwari</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C. F.</given-names>
            <surname>Enguix</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Lata</surname>
          </string-name>
          ,
          <article-title>Text2kgbench: A benchmark for ontologydriven knowledge graph generation from text</article-title>
          , in: International semantic web conference, Springer,
          <year>2023</year>
          , pp.
          <fpage>247</fpage>
          -
          <lpage>265</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Bang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Cahyawijaya</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Dai</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Su</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Wilie</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Lovenia</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Ji</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Yu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Chung</surname>
          </string-name>
          , et al.,
          <article-title>A multitask, multilingual, multimodal evaluation of chatgpt on reasoning, hallucination, and interactivity</article-title>
          ,
          <source>arXiv preprint arXiv:2302.04023</source>
          (
          <year>2023</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>L.</given-names>
            <surname>Huang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Yu</surname>
          </string-name>
          , W. Ma,
          <string-name>
            <given-names>W.</given-names>
            <surname>Zhong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Feng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Peng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Feng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Qin</surname>
          </string-name>
          , et al.,
          <article-title>A survey on hallucination in large language models: Principles, taxonomy, challenges, and open questions</article-title>
          ,
          <source>ACM Transactions on Information Systems</source>
          <volume>43</volume>
          (
          <year>2025</year>
          )
          <fpage>1</fpage>
          -
          <lpage>55</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>T.</given-names>
            <surname>Brown</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Mann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Ryder</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Subbiah</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. D.</given-names>
            <surname>Kaplan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Dhariwal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Neelakantan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Shyam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Sastry</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Askell</surname>
          </string-name>
          , et al.,
          <article-title>Language models are few-shot learners</article-title>
          ,
          <source>Advances in neural information processing systems</source>
          <volume>33</volume>
          (
          <year>2020</year>
          )
          <fpage>1877</fpage>
          -
          <lpage>1901</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>B.</given-names>
            <surname>Lester</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Al-Rfou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Constant</surname>
          </string-name>
          ,
          <article-title>The power of scale for parameter-eficient prompt tuning</article-title>
          ,
          <source>arXiv preprint arXiv:2104.08691</source>
          (
          <year>2021</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>J.</given-names>
            <surname>Achiam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Adler</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Agarwal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Ahmad</surname>
          </string-name>
          ,
          <string-name>
            <given-names>I.</given-names>
            <surname>Akkaya</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F. L.</given-names>
            <surname>Aleman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Almeida</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Altenschmidt</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Altman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Anadkat</surname>
          </string-name>
          , et al.,
          <source>Gpt-4 technical report, arXiv preprint arXiv:2303.08774</source>
          (
          <year>2023</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <surname>Langchain</surname>
          </string-name>
          , Llm graph transformer,
          <year>2023</year>
          . URL: https://python.langchain.
          <source>com/v0</source>
          .2/api_
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>