<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>PrO-KGC: Prompt Optimization for LLM-Based Knowledge Graph Completion</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Amel Gader</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Alsayed Algergawy</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Chair of Data and Knowledge Engineering, University of Passau</institution>
          ,
          <addr-line>Passau</addr-line>
          <country country="DE">Germany</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Knowledge Graphs (KGs) are integral to multiple applications which makes their completeness a key research focus. Large Language Models (LLMs) have been increasingly leveraged alongside embedding models to enhance Knowledge Graph Completion (KGC). However, efectively transforming contextual features from KGs into wellstructured textual prompts for LLMs remains a challenge. In this work, we propose PrO-KGC, a framework for Prompt Optimization in the context of LLM-Based KGC. PrO-KGC aims to enhance LLM-based KGC performance by refining and optimizing input prompts through multiple steps. This process includes enriching prompts with additional information, incorporating structural composition patterns and facts, and refining relation representations to be more LLM-friendly. Experimental results demonstrate the efectiveness of our approach, achieving improvements across three common benchmarks compared to vanilla models. Our code is available at https://github.com/amal-gader/PrO-KGC.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Knowledge Graph Completion</kwd>
        <kwd>Large Language Models</kwd>
        <kwd>Knowledge Graph to Text</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>Knowledge graphs form the backbone of the semantic web and are among the most widely used
knowledge representation techniques across various domains. As a result, extensive research has been
conducted to address the incompleteness of knowledge graphs through diferent tasks and settings
like the link prediction task which consists of predicting a tail entity t for a given head entity h and a
relation r (h, r, ?).</p>
      <p>
        Knowledge graph completion was initially tackled using knowledge graph embedding (KGE) models
such as TransE [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] and RotatE [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. Later, embedding models evolved to incorporate semantics and
textual information alongside structural connections in knowledge graphs [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. With the rise of large
language models (LLMs), tasks like link prediction were reformulated as sequence-to-sequence problems.
Models like KG-BERT [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] and KGT5 [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] demonstrated the potential of LLMs in knowledge graph
completion (KGC). However, research consistently finds that LLMs alone struggle to capture the
intricate relationships and structured nature of knowledge graphs [
        <xref ref-type="bibr" rid="ref3 ref6">6, 3</xref>
        ]. Instead of viewing KGE
models and LLMs as competing approaches, recent studies like [
        <xref ref-type="bibr" rid="ref10 ref7 ref8 ref9">7, 8, 9, 10</xref>
        ] have explored their synergy,
leveraging LLMs in a second step to rerank and refine KGE model predictions.
      </p>
      <p>
        In-context learning techniques were used to craft rich prompts by including entity descriptions, entity
classes and relevant relations providing the model with structured hints about potential predictions
[
        <xref ref-type="bibr" rid="ref11 ref12">11, 12</xref>
        ]. However, prompts can still be further optimized to maximize the integration of textual features
and better adapt the task to LLMs’ input processing patterns, especially that LLMs have shown reasoning
capabilities [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ]. One interesting aspect that was not well explored is the incorporation of composition
patterns in the given prompt. A composition pattern is a set of relations where one relation logically
implies another. In other terms, if A has relation 1 with B and B has relation 2 with C , then we can
infer that A has relation 3 with C . An intuitive example is illustrated in Figure 1: if Dr. John Smith
works at the University of Edinburgh and the University of Edinburgh is in Edinburgh, Scotland, then
we may reasonably infer that Dr. John Smith lives in Edinburgh, Scotland. These facts help resolve
ambiguities and improve the model’s ability to generate accurate predictions. However, not all multi-hop
pattern relations hold consistently, which can sometimes be misleading. For example, the following
path: Bob is friends with Alice, and Alice is friends with Charlie does not necessarily imply that Bob is
friends with Charlie. This highlights the need for a meticulous approach in selecting composition facts
from the knowledge graph to ensure reliability and avoid incorrect inferences.
      </p>
      <p>In this paper, we propose PrO-KGC a framework to systematically extract and integrate textual
features from knowledge graphs like descriptions, composition patterns and related contextual facts
into LLMs’ prompts, transforming the task into an NLP problem that aligns more closely with LLMs’
expected inputs. We also suggest methods to refine and adapt textual inputs bridging the gap between
symbolic reasoning in knowledge graphs and the language-driven capabilities of LLMs. Our method
consists of three main components: Composition Pattern Extraction (CPE), which extracts and validates
relational patterns from the KG to retrieve supporting facts, Description Generation (DG), which
generates missing entity descriptions and Relation Transformation (RT), which transforms no-verbal
relations into more intuitive natural language forms.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related Work</title>
      <p>In this section we provide a brief survey of the latest state-of-the-art research combining Large Language
Models (LLMs) and Knowledge Graph Completion (KGC) models and methods to improve the integration
of structured knowledge with LLMs for KGC tasks.</p>
      <sec id="sec-2-1">
        <title>2.1. State of the Art in KGC with LLMs</title>
        <p>
          The KGC task has evolved significantly with the integration of Knowledge Graph Embeddings (KGE)
and LLMs. Traditional KGE models such as TransE [
          <xref ref-type="bibr" rid="ref1">1</xref>
          ], RotatE [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ], and TuckER [
          <xref ref-type="bibr" rid="ref14">14</xref>
          ] learn structural
connections from the training data but remain limited by their inability to leverage textual information.
On the other hand, LLM-based models rely on rich textual features overlooking the structured nature
of knowledge graphs. Several recent approaches have tried to address this gap by combining the
strengths of both paradigms. One such method, KICGPT [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ], introduces a framework that leverages
both structural and textual information side by side. KICGPT incorporates a two-stage process where a
triple-based retriever model extracts structural knowledge, followed by an LLM-based reranker that
iteratively refines the candidate predictions. This method requires training only the retriever module,
while the reranking is handled by an external LLM, ensuring eficiency. However, the iterative reranking
introduces additional computational costs. To reduce this overhead and hallucinations associated with
LLMs, DIFT [
          <xref ref-type="bibr" rid="ref10">10</xref>
          ] proposes an alternative by finetuning an LLM specifically for the reranking task.
Instead of treating the LLM as a generic reasoning engine, this approach injects structured embeddings
of both the query and candidate entities into the model, making it more aligned with the KG’s structure.
Additionally, it employs a refined sampling strategy to filter and select only the most relevant contextual
facts, thereby improving eficiency and reliability.
        </p>
        <p>
          Another direction explored in KGR3 [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ] is addressing the variability between the predefined semantics
of a KG and the open-ended nature of generative LLM predictions. This method follows a three-step
process where a retrieval module extracts structural information from the KG and generates a list of
candidate answers, a reasoning module uses an LLM to infer additional candidates, and a reranking
module to combine and rerank these results.
        </p>
        <p>
          A related efort, KLR-KGC [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ], aims to leverage analogical reasoning by extracting structural patterns
and subgraphs from the KG. The extracted analogical knowledge helps filter KGE predictions based
on similarities to known triples, while the subgraph information provides additional context for the
reranking step. By introducing structural analogies, this approach ensures that predictions abide by the
underlying KG structure.
        </p>
      </sec>
      <sec id="sec-2-2">
        <title>2.2. Integrating Structured Knowledge into prompts for LLM-Based KGC</title>
        <p>
          Beyond the integration of LLMs with KGE models, recent work has also explored optimizing textual
representations to better guide LLMs in KGC tasks. One example is MPIKGC [
          <xref ref-type="bibr" rid="ref11">11</xref>
          ] which focuses on
designing efective prompts that are adaptable across diferent LLMs. By expanding entity and relational
descriptions both globally (from the KG perspective) and locally (from the triple perspective), it improves
the contextual richness of the input. However, it only relies on keyword-based fact extraction.
        </p>
        <p>
          To improve multi-hop link prediction, KG-LLM [
          <xref ref-type="bibr" rid="ref15">15</xref>
          ] formulates knowledge graph structures as natural
language inputs. By extracting all possible paths using depth-first search (DFS) and categorizing them
based on whether the first and last nodes are connected, it enables the model to learn logical chains of
inference. However, the model may apply learned decision paths even when the pattern does not hold
due to the absence of a filtering step leading to overgeneralization.
        </p>
        <p>
          Another approach, MuKDC [
          <xref ref-type="bibr" rid="ref12">12</xref>
          ], aims to enhance few-shot KGC by generating additional knowledge
for long-tailed entities. It utilizes an LLM to generate triplets, attributes, and decision paths based on
a set of instructions and prompt templates, enriching entities with relevant contextual information.
To mitigate potential inaccuracies introduced by the LLM, a consistency check is applied using a KGE
model to score and validate new facts. However, this method is relying on LLMs for structural context
generation, which may not always align with the KG’s factual constraints.
        </p>
        <p>Overall, these works ofer valuable contributions and point out the ongoing eforts to integrate
structured knowledge with LLMs’ capabilities. While some remain loyal to structural patterns through
retrievers and analogical reasoning, others focus on refining textual inputs to better fit LLMs’
expectations. The synergy between structured KGE models and unstructured LLMs’ reasoning remains a key
area of research to get the best blend of both paradigms. In this work we focus on the latter approach
and leave the exploration of the former for future work.</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Methodology</title>
      <sec id="sec-3-1">
        <title>In this section we explain in detail our PrO-KGC approach.</title>
        <sec id="sec-3-1-1">
          <title>3.1. Problem Definition</title>
          <p>A knowledge graph (KG) is a structured representation of knowledge, defined as  = (E , R, F , D ),
where E is the set of entities, R is the set of relations, F is the set of factual triples, and D is
the set of textual descriptions associated with entities. Each fact  ∈ F is represented as a triplet:
(h, r , t ) ∈ E × R × E , where h, t are the head and tail entities respectively connected by the relation
r . An entity may have a description denoted as  ∈ D .</p>
          <p>Knowledge Graph Completion (KGC) includes tasks such as link prediction, relation prediction and
instance completion aiming all to enrich KGs with new inferred facts. Traditional triple-based models
do not leverage the set of descriptions D and other potential textual features, while LLM-based methods
often overlook the structural connections within the KG. The PrO-KGC framework addresses this
limitation and suggests a set of steps to preserve structual information and optimize the LLM input
prompt for KGC tasks.</p>
        </sec>
        <sec id="sec-3-1-2">
          <title>3.2. Overview</title>
          <p>Our approach consists of two main steps, with the first being the primary focus of this work.</p>
          <p>Step 1: Prompt Generation. The goal of this step is to extract and integrate the most relevant
knowledge from the KG into the prompt to enhance the performance of the language model. First, if
a head entity description is available in the KG, we incorporate it directly. Otherwise, we generate it
using a backbone Large Language Model (LLM). Next, we extract composition patterns from the KG,
validating them with the same LLM to extract relevant supporting facts. If no composition patterns are
available, we retrieve a fact associated with a neighboring entity that shares the same relation. If such
a fact is also unavailable, we select any fact involving the same relation. To address cases where the
relation is ambiguous or dificult to interpret, we transform it into a more intuitive verb-based form
using the LLM. All these components are then combined into a structured prompt, compensating for
the language model’s lack of direct access to the structural and relational information in the KG. All of
these steps are detailed in the following subsections and depicted in Figure 2.</p>
          <p>Step 2: Link Prediction. We evaluate the efectiveness of our framework on the Link Prediction
task. We use the constructed prompt as the input for a finetuned LLM.</p>
          <p>Link prediction focuses on inferring missing entities in triples of the form (h,r,?) or (?,r,t), where the
goal is to predict the missing tail t or head h, respectively. In this work, we focus on tail prediction,
noting that the two cases are structurally symmetric.</p>
        </sec>
        <sec id="sec-3-1-3">
          <title>3.3. Prompt Generation</title>
          <p>Following we detail the steps of the proposed PrO-KGC approach to generate a clear and informative
prompt for the model.</p>
          <p>Composition Pattern Extraction (CPE) To provide the model with a structured subgraph containing
relevant facts that support the inference of implicit knowledge, we extract all possible continuous paths
of length  using a depth-first search (DFS) algorithm, as detailed in Algorithm 1. The hyperparameter 
determines the maximum allowed path length and can be set empirically. A multi-hop path of length 
consists of a sequence of  interconnected nodes and relations, where each consecutive pair of nodes is
linked by a relation. These paths connect a subject to an object and are grounded in actual entities and
facts from the knowledge graph.
A composition pattern (CP), on the other hand, is defined as the ordered sequence of relations that
interconnect the nodes within a multi-hop path. It represents an abstract relational template derived
from the structure of the graph. However, not all extracted composition patterns are semantically
valid or reliable. For example, if the USA has a diplomatic relation with Germany, and Germany has
a diplomatic relation with France, then it might be reasonable to assume that the USA also has a
diplomatic relation with France (which is also a valid fact). In this case, the algorithm would extract
the composition pattern (hasDiplomaticRelation, hasDiplomaticRelation, hasDiplomaticRelation) as
depicted in Figure 3. The same pattern could misleadingly suggest that the USA has a diplomatic relation
with North Korea, given that the USA has a diplomatic relation with China, and China has a diplomatic
relation with North Korea.</p>
          <p>Since our goal is to improve the quality of the prompt rather than introduce misleading information,
we employ an LLM to validate and filter out erroneous patterns that could negatively impact the model’s
reasoning. The prompt used is reported in Appendix A.1.</p>
          <p>Descriptions Generation (DG) The head entity’s description plays a crucial role in disambiguating
entity types and names, ensuring clarity and reducing ambiguity. It also allows the Large Language
Model (LLM) to retrieve and reinforce relevant knowledge about the entity from its internal knowledge
base. Modern knowledge graphs often include descriptions for entities, but some may be missing. Given
that LLMs have been trained on extensive knowledge bases and hence have broad general knowledge,
we leverage them as an external resource to generate missing descriptions. We address cases where an
entity ENT lacks a description by prompting our backbone model with the following instruction. If an
entity is not recognized, we instruct the model to return a blank response to reduce post-processing
eforts and limit the risk of hallucinations:</p>
          <p>Prompt:
You are a knowledge graph. Provide a concise description of the given entity. Return only the
description without any additional text. If the entity is unknown, return a blank response: {ENT}</p>
        </sec>
      </sec>
      <sec id="sec-3-2">
        <title>Algorithm 1 Extraction of Composition Patterns from a Knowledge Graph</title>
        <p>1: Initialize pattern dictionary  ← ∅ to store relation sequences and examples.
2: For each entity node  ∈ , perform DFS to explore paths of length .
3: If a valid path is found:</p>
        <p>Let the path be: (1, 1, 2), (2, 2, 3), ..., (− 1, − 1, )</p>
        <p>Check if a direct relation exists between 1 and : (1,  , ) ∈ 
if  exists then</p>
        <p>Form pattern: (1, 2, ..., − 1 →  )
Store two examples of supporting facts:</p>
        <p>Observed: (, , +1) for  = 1, ...,  − 1</p>
        <p>Inferred: (1,  , )
end if
4: Repeat for all nodes.
5: Until no more paths exist.
6: Validate extracted patterns using an LLM:</p>
        <p>validated ← LLM( )
7: Return validated.</p>
        <p>
          Relation Transformation (RT) In some datasets, relations are expressed in a structured or a
nonnatural language format, making them dificult to interpret and incorporate into an LLM-friendly
prompt. For example in the Freebase-extracted FB15K-237 dataset [
          <xref ref-type="bibr" rid="ref16">16</xref>
          ], relations are represented as a
chain of concatenated categories like: /sports/sports_position/players./sports/sports_team_roster/position
which could be simplified to plays as. To ensure prompt readability and clarity, we introduce a step to
simplify relation representation and semantics.
        </p>
        <p>This is achieved by instructing our backbone LLM with the following prompt:
Prompt:
You will get a relation from the Freebase knowledge graph along with an example. Your task is to
convert the relation into a simple verb form while preserving its meaning and key terms. Return
only the converted relation. Example: Input: /__ Output:   Relation:
{REL}
Contextual Facts Retrieval (CFR) In this step, we leverage the composition patterns extracted and
validated in the first stage. If the relation in the given triplet matches one of these patterns, we retrieve
the associated nodes from the graph following the same relational structure. For more details refer to
the algorithm in Appendix A.3. When no matching pattern is found, we instead retrieve a neighboring
fact that shares the same relation. In such cases, neighboring nodes often reveal similar properties,
which can support inference. If neither approach is applicable, we include a fact containing the same
relation to provide contextual hints about entity categories to reduce ambiguity. It’s worth noting that
during the training phase, we strictly limit the CFR module to access only the training and validation
splits to prevent any risk of data leakage.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Experiments</title>
      <p>In this section, we present the experimental setup, including datasets, evaluation metrics, and results.</p>
      <sec id="sec-4-1">
        <title>4.1. Setups</title>
        <p>
          Datasets We evaluate our approach on three widely used benchmark datasets: FB15k-237 by [
          <xref ref-type="bibr" rid="ref16">16</xref>
          ], an
improved iteration of FB15k [
          <xref ref-type="bibr" rid="ref16">16</xref>
          ] that removes inverse relations to prevent data leakage. WN18RR [
          <xref ref-type="bibr" rid="ref17">17</xref>
          ]
is as well a refined version of WN18 [
          <xref ref-type="bibr" rid="ref16">16</xref>
          ], a dataset based on WordNet. And CoDEx-M [
          <xref ref-type="bibr" rid="ref18">18</xref>
          ] which is
extracted from Wikidata and Wikipedia. Detailed statistics of these datasets are provided in Table 1.
Baselines To assess the impact of our framework’s components on KGC tasks, particularly link
prediction, we establish a baseline using the vanilla model (i.e., before integrating our framework
enhancements). Additionally, we benchmark our approach against leading triple-based models, including
TuckER [
          <xref ref-type="bibr" rid="ref14">14</xref>
          ], TransE [
          <xref ref-type="bibr" rid="ref1">1</xref>
          ], and ComplEx[
          <xref ref-type="bibr" rid="ref19">19</xref>
          ], as well as the earliest implemented LLM-based methods:
KG-BERT [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ], KGT5[
          <xref ref-type="bibr" rid="ref5">5</xref>
          ], and GenKGC[
          <xref ref-type="bibr" rid="ref20">20</xref>
          ].
        </p>
        <p>Implementation Details We conduct our experiments using two diferent GPUs: an NVIDIA A100
(80GB) and an NVIDIA RTX A6000 (50GB).</p>
        <p>
          For the knowledge base LLM in the DG module, we use LLaMA 3.1 70B1. In the CPE filtering step
and RT module, we employ GPT-4-Turbo via the OpenAI API2, with a temperature of 1 and top-p set to
1. The size of composition patterns is set to 3. For the link prediction task, we use two diferent models:
T5-Large3 from [
          <xref ref-type="bibr" rid="ref21">21</xref>
          ]. and LLaMA 3 (8B) from Unsloth4, which enables 2× faster fine-tuning, 70% lower
memory usage, and 2× faster inference.
        </p>
        <p>
          To optimize GPU memory usage and eficiency, we apply LoRA [
          <xref ref-type="bibr" rid="ref22">22</xref>
          ] and QLoRA [
          <xref ref-type="bibr" rid="ref23">23</xref>
          ] in the fine-tuning
process. As decoding techniques, we use top-k sampling (top_k=50) and beam search (num_beams=10)
to get the top-10 predictions. The remaining hyperparameters are reported in Appendix A, Table 5.
Metrics To evaluate the link prediction task, we use standard evaluation metrics: Hits@k (for
 ∈ {1, 3, 10}), which measures the proportion of cases where the correct entity appears in the top-k
predictions. Additionally, we use the Mean Reciprocal Rank (MRR), which computes the average of the
inverse ranks of the correct predictions.
        </p>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. Main Results</title>
        <p>The main results of our experiments are reported in Table 2. We observe that our enhanced model
competes with PLM-based models; however, KGT5 continues to outperform on WN18RR. This can be
explained by the fact that KGT5 was trained exclusively on KGC benchmarks. Pretrained weights were
discarded and the objective function was replaced with a link prediction one. This probably reduced
potential hallucinations, as the model encounters entity names for the first time within the KG. In</p>
        <p>MRR
contrast, LLMs, which were initially pretrained on large text corpora, may show biases toward the
expected answer format from their pretraining. Particularily, relations and taxonomy in WN18RR
might difer from those in the generic pretraining datasets used for LLMs. For example: according to
the WordNet Glossary5, the relation "derivationally related form" refers to terms in diferent syntactic
categories that share the same root form and are semantically related. However, in WN18RR, we find
entities connected through this relation which are semantically related but do not share the same root.
e.g: ("cover", "derivationally related form", "spread over")</p>
        <p>
          Regarding triple-based models, we observe that our model outperforms them on CoDEx with a 6.9%
Hits@1 improvement, while TuckER achieves better results on WN18RR and FB15k-237, which might be
explained by the inherent nature of these datasets. CoDEx, based on Wikidata, primarily contains binary
relations, unlike FB15k-237, which is derived from Freebase, where relations are n-nary and require
an understanding of the KG structure [
          <xref ref-type="bibr" rid="ref18">18</xref>
          ], such as /media_common/netflix_genre/titles . Deterministic
binary relations help LLMs in making more accurate predictions.
        </p>
        <p>
          According to [
          <xref ref-type="bibr" rid="ref25">25</xref>
          ], WN18RR is the most structured KG among these datasets, which could explain
the strong performance of triple-based models on it.
        </p>
        <p>Our overarching goal is to improve LLM-based KGC performance. The results show that our approach
outperforms the vanilla models across all three benchmarks, with a notable 9.7% improvement in Hits@1
on WN18RR, 3.8% on CoDEx, and a slight enhancement on FB15k-237 of 1.4% in Hits@1 and 2.1% in
Hits@10. This confirms our hypothesis that enriching the prompt with structured knowledge enhances
LLM-based KGC performance.</p>
      </sec>
      <sec id="sec-4-3">
        <title>4.3. Ablation Study</title>
        <p>To gauge the impact of the diferent components on the model’s performance, we conduct an ablation
study. We start with the vanilla model where we only feed the model with the query including the
head entity and the relation and we instruct the model to predict the tail entity. Then, we gradually
add the diferent components starting with the Description Generation (DG) component and then the
Contextual facts Retrieval (CFR) component. The results of this study are reported in Table 3 and show
that the more components we add the better the model’s performance gets.</p>
        <p>
          In fact, the incorporation of textual descriptions in CoDEx boosts performance by 3.4% in Hits@10,
while the addition of structural information further increases this improvement to 4.7%. On WN18RR,
we observe a 9.4% increase in Hits@10 with the addition of contextual facts, comparing the vanilla
5WordNet Glossary
model to the improved version. As aforementioned, WN18RR is a typical structured network, which
probably contributes to this boost. For FB15k-237, however, incorporating descriptions does not
significantly enhance performance, while adding contextual facts improves Hits@10 by 2.1%. One
possible explanation for the limited impact of descriptions is their length: 141.7 words on average per
description [
          <xref ref-type="bibr" rid="ref26">26</xref>
          ]. These lengthy descriptions may contain too much information, which may cause
important facts to lose their relevance.
        </p>
        <p>Model Variant</p>
        <p>MRR
+PrO-KGC (Full Model) 0.293
w/ Desc, w/o CP 0.272
w/o CP and DG (Vanilla) 0.271</p>
        <p>FB15k-237
H@1 H@3
0.241 0.317
0.226 0.290
0.227 0.294</p>
        <p>To better understand the influence of multi-hop decision paths on model performance, we measure
accuracy on a subset of CoDEx test triples that are supported by such paths, and compare it to the
overall performance on the full test dataset. As shown in Table 4, the performance is significantly
better on this subset, with Hits@1 and Hits@3 increased by 14.5% and 15.9%, respectively. These results
indicate that the presence of structured decision paths plays a significant role in guiding the model’s
predictions.</p>
        <p>Hits@10
0.331</p>
      </sec>
      <sec id="sec-4-4">
        <title>4.4. Model Comparison</title>
        <p>In this subsection, we compare the performance of the diferent models fine-tuned for the link prediction
task based on two key dimensions: Accuracy and Eficiency, as depicted in Figure 4. The models used
are, as aforementioned, T5-large and LLaMA 3 8B, with only a small percentage of their parameters
ifne-tuned: 4.7M (0.64%) for T5 and 42M (0.92%) for LLaMA.</p>
        <p>We observe that, overall, LLaMA outperforms T5 in terms of accuracy and is more eficient in training
time for WN18RR and CoDEx. However, on FB15k-237, T5 is 70% faster (21 vs. 29 hours). One possible
explanation for this diference is the nature of the dataset and the complexity of its n-nary relations.
In this case, decoder-only models like LLaMA may struggle with complex multi-entity relationships,
requiring more updates to learn meaningful patterns.</p>
        <p>Regarding performance, LLaMA has more parameters and was trained on larger datasets, which could
explain its superior results, particularly in Hits@1. For example, on CoDEx, T5 achieves 0.212, whereas
LLaMA reaches 0.331, an 11.9% improvement. However, in terms of Hits@10, the T5-based model
outperforms LLaMA (0.584 vs. 0.561, a 2.3% diference). One possible reason for this is the diference
in decoding methods; for LLaMA, beam search is not available in the used library, so sampling was
employed instead.</p>
        <p>In terms of the impact of additional prompt components, we observe that LLaMA’s performance
on CoDEx and WN18RR significantly decreases when prompt components are ablated (9.4% and 4.7%
reductions, respectively), whereas T5 remains relatively stable (2.4% and 2.5% reductions). One possible
explanation is T5’s 512-token context length limit, prompts exceeding this limit are truncated, which
could lead to information loss.</p>
      </sec>
      <sec id="sec-4-5">
        <title>4.5. Limitations</title>
        <p>In this section, we discuss the major limitations of our work from our perspective.</p>
        <p>Firstly, we believe that applying the framework to other Knowledge Graph Completion (KGC) tasks
beyond link prediction, such as triple classification and relation prediction, may reveal additional
insights and further validate the efectiveness of our approach.</p>
        <p>For composition patterns, the length  was fixed at 3, but we believe that longer paths could lead to
greater improvements, particularly for multi-hop relations. Additionally, composition patterns cannot
be applied to all samples; only a small percentage of queries can leverage them (e.g., in the FB15K-237
test set, only 10% of queries benefit from contextual paths), which may obscure their true impact.
A potential solution could be utilizing an LLM to generate decision paths when compositions are
unavailable.</p>
        <p>Another limitation is that the reported results in Table 2 were not extracted under the same settings,
which is not ideal for a fair comparison. Additionally, the extraction of contextual facts could benefit
from refinements, such as selecting the most semantically similar facts to the query.</p>
        <p>A further arguable point is the limited improvement of our framework on the FB15k-237 dataset
and the overall performance lagging behind knowledge graph embeddings. This could be attributed to
the reduction of trainable parameters to 42M (0.92%), which may limit the LLM’s ability to memorize
newly learned facts or patterns. This reinforces the idea that LLMs alone cannot fully replace KGC
models but should rather be optimized to complement KGE models.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Conclusion</title>
      <p>In this work, we introduced PrO-KGC, a framework for prompt optimization in LLM-based
Knowledge Graph Completion (KGC). PrO-KGC follows a structured approach to extracting and integrating
structural context from the KG into prompts, leveraging LLMs to bridge the gap between structured
knowledge and LLM capabilities.</p>
      <p>We first extract composition patterns, validate them using a knowledge base LLM, and harness them
to support link prediction queries with relevant subgraphs. In addition, we generate new descriptions
when needed and refine relation representations to make the prompt more natural, language-friendly,
and semantically rich.</p>
      <p>Through a set of experiments, we demonstrated the efectiveness of our approach across three
diferent benchmarks. However, we also conclude that LLMs are still far from replacing embedding
models; rather, they should be used jointly. Our work contributes to fostering a more efective synergy
between LLMs and Knowledge Graph Embedding (KGE) models.</p>
      <p>For future work, we plan to assess the impact of varying composition pattern lengths on model
performance and explore transforming the PrO-KGC textual input into embeddings as an initial input
for embedding-based models. Another promising direction is evaluating our approach on multiple KGC
tasks and investigating new strategies to further optimize prompt design.</p>
    </sec>
    <sec id="sec-6">
      <title>6. Declaration on Generative AI</title>
      <p>During the preparation of this work, the author(s) used ChatGPT in order to: spelling check, paraphrase
and reword. After using these tool(s)/service(s), the author(s) reviewed and edited the content as needed
and take(s) full responsibility for the publication’s content.</p>
    </sec>
    <sec id="sec-7">
      <title>A. Appendix</title>
      <sec id="sec-7-1">
        <title>A.1. Prompt used for Pattern Validation</title>
        <p>Pattern Validation Prompt
Instruction: Given a relation pattern (r1, r2, r3) and two examples of paths using these
relations, determine whether this represents a valid or invalid composition pattern.
A valid composition pattern means that Fact1 and Fact2 necessarily imply Fact3. If this
implication does not universally hold or if there are exceptions, classify it as invalid.
Your response should be in the following format: Valid/Invalid. Explanation:
Pattern: (diplomatic relation, member of, member of)
Example 1:
Fact1: (Senegal, diplomatic relation, Germany)
Fact2: (Germany, member of, International Telecommunication Union)
Fact3: (Senegal, member of, International Telecommunication Union)
Example 2:
Fact1: (Senegal, diplomatic relation, Germany)
Fact2: (Germany, member of, African Development Bank)
Fact3: (Senegal, member of, African Development Bank)
Response: Invalid. Explanation: If a country has a diplomatic relation with another that is
a member of an organization, it does not necessarily imply that it is also part of that organization.</p>
      </sec>
      <sec id="sec-7-2">
        <title>A.2. Hyperparameters of Models used for Link Prediction</title>
      </sec>
      <sec id="sec-7-3">
        <title>A.3. Extraction of supporting facts from the KG</title>
        <sec id="sec-7-3-1">
          <title>Algorithm 2 Extract Supporting Facts from a Graph</title>
          <p>Input: A query (ℎ, , ?), a set of composition patterns , and a knowledge graph .
Output: A supporting triplet or triplet pair.</p>
          <p>Identify patterns where relation  appears.
for each matching pattern do</p>
          <p>Determine the position of  in the pattern.
if  appears as the last relation then</p>
          <p>Retrieve the preceding relations (1, 2): (1, 2, ) ← .</p>
          <p>Find entities 1 connected to ℎ via 1 in .
for each 1 do</p>
          <p>Store (ℎ, 1, 1) as a potential supporting fact.</p>
          <p>Check if 1 is linked to another entity 2 via 2.
if (1, 2, 2) exists in  then</p>
          <p>Store the pair ((ℎ, 1, 1), (1, 2, 2)).</p>
          <p>end if
end for
end if
end for
if ∃(ℎ, 1, 1), (1, 2, 2) ∈  then return randomly chosen ((ℎ, 1, 1), (1, 2, 2)).
else if ∃(ℎ, 1, 1) ∈  then return randomly chosen (ℎ, 1, 1).
elsereturn None.
end if</p>
        </sec>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>A.</given-names>
            <surname>Bordes</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Usunier</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Garcia-Duran</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Weston</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Yakhnenko</surname>
          </string-name>
          ,
          <article-title>Translating embeddings for modeling multi-relational data</article-title>
          ,
          <source>Advances in neural information processing systems</source>
          <volume>26</volume>
          (
          <year>2013</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Sun</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.-H.</given-names>
            <surname>Deng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.-Y.</given-names>
            <surname>Nie</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Tang</surname>
          </string-name>
          ,
          <article-title>Rotate: Knowledge graph embedding by relational rotation in complex space</article-title>
          , arXiv preprint arXiv:
          <year>1902</year>
          .
          <volume>10197</volume>
          (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>L.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Zhao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Wei</surname>
          </string-name>
          , J. Liu, Simkgc:
          <article-title>Simple contrastive knowledge graph completion with pre-trained language models</article-title>
          ,
          <source>arXiv preprint arXiv:2203.02167</source>
          (
          <year>2022</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>L.</given-names>
            <surname>Yao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Mao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Luo</surname>
          </string-name>
          , KG-BERT:
          <article-title>BERT for knowledge graph completion</article-title>
          , CoRR abs/
          <year>1909</year>
          .03193 (
          <year>2019</year>
          ). URL: http://arxiv.org/abs/
          <year>1909</year>
          .03193. arXiv:
          <year>1909</year>
          .03193.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>A.</given-names>
            <surname>Saxena</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Kochsiek</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Gemulla</surname>
          </string-name>
          ,
          <article-title>Sequence-to-sequence knowledge graph completion and question answering</article-title>
          ,
          <year>2022</year>
          . URL: https://arxiv.org/abs/2203.10321. arXiv:
          <volume>2203</volume>
          .
          <fpage>10321</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>X.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Gao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Zhu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Tang</surname>
          </string-name>
          ,
          <article-title>Kepler: A unified model for knowledge embedding and pre-trained language representation</article-title>
          ,
          <year>2020</year>
          . URL: https://arxiv.org/abs/
          <year>1911</year>
          .06136. arXiv:
          <year>1911</year>
          .06136.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Wei</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Huang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , J. Kwok, Kicgpt:
          <article-title>Large language model with knowledge in context for knowledge graph completion, in: Findings of the Association for Computational Linguistics: EMNLP 2023, Association for Computational Linguistics</article-title>
          ,
          <year>2023</year>
          , p.
          <fpage>8667</fpage>
          -
          <lpage>8683</lpage>
          . URL: http://dx.doi. org/10.18653/v1/
          <year>2023</year>
          .findings-emnlp.
          <volume>580</volume>
          . doi:
          <volume>10</volume>
          .18653/v1/
          <year>2023</year>
          .findings-emnlp.
          <volume>580</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>S.</given-names>
            <surname>Ji</surname>
          </string-name>
          , L. Liu,
          <string-name>
            <given-names>J.</given-names>
            <surname>Xi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <article-title>Klr-kgc: Knowledge-guided llm reasoning for knowledge graph completion</article-title>
          ,
          <source>Electronics</source>
          <volume>13</volume>
          (
          <year>2024</year>
          ). URL: https://www.mdpi.com/2079-9292/13/24/5037. doi:
          <volume>10</volume>
          .3390/electronics13245037.
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>M.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Jiang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Qi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Guo</surname>
          </string-name>
          , H. fung
          <article-title>Leung, I. King, Retrieval, reasoning, reranking: A context-enriched framework for knowledge graph completion</article-title>
          ,
          <year>2024</year>
          . URL: https: //arxiv.org/abs/2411.08165. arXiv:
          <volume>2411</volume>
          .
          <fpage>08165</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Tian</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Sun</surname>
          </string-name>
          , W. Hu,
          <article-title>Finetuning generative large language models with discrimination instructions for knowledge graph completion</article-title>
          ,
          <year>2024</year>
          . URL: https://arxiv.org/abs/2407.16127. arXiv:
          <volume>2407</volume>
          .
          <fpage>16127</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>D.</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Lin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Zhu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Zhao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zheng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Chen</surname>
          </string-name>
          <article-title>, Multi-perspective improvement of knowledge graph completion with large language models</article-title>
          ,
          <source>arXiv preprint arXiv:2403</source>
          .
          <year>01972</year>
          (
          <year>2024</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>Q.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Ji</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Jiang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <article-title>Llm-based multi-level knowledge generation for few-shot knowledge graph completion</article-title>
          ,
          <source>in: Proceedings of the Thirty-Third International Joint Conference on Artificial Intelligence</source>
          , volume
          <volume>271494703</volume>
          ,
          <year>2024</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>J.</given-names>
            <surname>Huang</surname>
          </string-name>
          ,
          <string-name>
            <surname>K. C.-C. Chang</surname>
          </string-name>
          ,
          <article-title>Towards reasoning in large language models: A survey</article-title>
          ,
          <year>2023</year>
          . URL: https://arxiv.org/abs/2212.10403. arXiv:
          <volume>2212</volume>
          .
          <fpage>10403</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>I.</given-names>
            <surname>Balažević</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Allen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T. M.</given-names>
            <surname>Hospedales</surname>
          </string-name>
          , Tucker:
          <article-title>Tensor factorization for knowledge graph completion</article-title>
          , arXiv preprint arXiv:
          <year>1901</year>
          .
          <volume>09590</volume>
          (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>D.</given-names>
            <surname>Shu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Jin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Du</surname>
          </string-name>
          ,
          <string-name>
            <surname>Y. Zhang,</surname>
          </string-name>
          <article-title>Knowledge graph large language model (kg-llm) for link prediction</article-title>
          ,
          <source>arXiv preprint arXiv:2403.07311</source>
          (
          <year>2024</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>A.</given-names>
            <surname>Bordes</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Usunier</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Garcia-Duran</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Weston</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Yakhnenko</surname>
          </string-name>
          ,
          <article-title>Translating embeddings for modeling multi-relational data</article-title>
          , in: C.
          <string-name>
            <surname>Burges</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <string-name>
            <surname>Bottou</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Welling</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          <string-name>
            <surname>Ghahramani</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          Weinberger (Eds.),
          <source>Advances in Neural Information Processing Systems</source>
          , volume
          <volume>26</volume>
          ,
          <string-name>
            <surname>Curran</surname>
            <given-names>Associates</given-names>
          </string-name>
          , Inc.,
          <year>2013</year>
          . URL: https://proceedings.neurips.cc/paper_files/paper/2013/file/ 1cecc7a77928ca8133fa24680a88d2f9-Paper.pdf.
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>T.</given-names>
            <surname>Dettmers</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Minervini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Stenetorp</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Riedel</surname>
          </string-name>
          ,
          <article-title>Convolutional 2d knowledge graph embeddings</article-title>
          ,
          <source>in: Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence and Thirtieth Innovative Applications of Artificial Intelligence Conference and Eighth AAAI Symposium on Educational Advances in Artificial Intelligence</source>
          , AAAI'18/IAAI'18/EAAI'18, AAAI Press,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>T.</given-names>
            <surname>Safavi</surname>
          </string-name>
          , D. Koutra,
          <article-title>CoDEx: A Comprehensive Knowledge Graph Completion Benchmark</article-title>
          ,
          <source>in: Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP)</source>
          ,
          <article-title>Association for Computational Linguistics</article-title>
          , Online,
          <year>2020</year>
          , pp.
          <fpage>8328</fpage>
          -
          <lpage>8350</lpage>
          . URL: https:// www.aclweb.org/anthology/2020.emnlp-main.
          <volume>669</volume>
          . doi:
          <volume>10</volume>
          .18653/v1/
          <year>2020</year>
          .emnlp-main.
          <volume>669</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>T.</given-names>
            <surname>Trouillon</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C. R.</given-names>
            <surname>Dance</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Welbl</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Riedel</surname>
          </string-name>
          , Éric Gaussier, G. Bouchard,
          <article-title>Knowledge graph completion via complex tensor factorization</article-title>
          ,
          <year>2017</year>
          . URL: https://arxiv.org/abs/1702.06879. arXiv:
          <volume>1702</volume>
          .
          <fpage>06879</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>X.</given-names>
            <surname>Xie</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Deng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Xiong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <article-title>From discrimination to generation: Knowledge graph completion with generative transformer</article-title>
          ,
          <source>in: Companion Proceedings of the Web Conference</source>
          <year>2022</year>
          , WWW '22,
          <string-name>
            <surname>ACM</surname>
          </string-name>
          ,
          <year>2022</year>
          . URL: http://dx.doi.org/10.1145/3487553. 3524238. doi:
          <volume>10</volume>
          .1145/3487553.3524238.
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>C.</given-names>
            <surname>Rafel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Shazeer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Roberts</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Narang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Matena</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. J.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <article-title>Exploring the limits of transfer learning with a unified text-to-text transformer</article-title>
          ,
          <year>2023</year>
          . URL: https://arxiv.org/ abs/
          <year>1910</year>
          .10683. arXiv:
          <year>1910</year>
          .10683.
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>E. J.</given-names>
            <surname>Hu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Shen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Wallis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Allen-Zhu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Chen</surname>
          </string-name>
          , Lora:
          <article-title>Low-rank adaptation of large language models</article-title>
          ,
          <year>2021</year>
          . URL: https://arxiv.org/abs/2106.09685. arXiv:
          <volume>2106</volume>
          .
          <fpage>09685</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <given-names>T.</given-names>
            <surname>Dettmers</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Pagnoni</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Holtzman</surname>
          </string-name>
          , L. Zettlemoyer, Qlora: Eficient finetuning of quantized llms,
          <year>2023</year>
          . URL: https://arxiv.org/abs/2305.14314. arXiv:
          <volume>2305</volume>
          .
          <fpage>14314</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [24]
          <string-name>
            <given-names>D.</given-names>
            <surname>Nathani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Chauhan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Sharma</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Kaul</surname>
          </string-name>
          ,
          <article-title>Learning attention-based embeddings for relation prediction in knowledge graphs</article-title>
          , in: A.
          <string-name>
            <surname>Korhonen</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Traum</surname>
          </string-name>
          , L. Màrquez (Eds.),
          <article-title>Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, Association for Computational Linguistics</article-title>
          , Florence, Italy,
          <year>2019</year>
          , pp.
          <fpage>4710</fpage>
          -
          <lpage>4723</lpage>
          . URL: https://aclanthology.org/ P19-1466/. doi:
          <volume>10</volume>
          .18653/v1/
          <fpage>P19</fpage>
          -1466.
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          [25]
          <string-name>
            <given-names>S.</given-names>
            <surname>Faralli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Lenzi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Velardi</surname>
          </string-name>
          ,
          <article-title>A benchmark study on knowledge graphs enrichment and pruning methods in the presence of noisy relationships</article-title>
          ,
          <source>J. Artif. Int. Res</source>
          .
          <volume>78</volume>
          (
          <year>2024</year>
          ). URL: https://doi.org/10. 1613/jair.1.14494. doi:
          <volume>10</volume>
          .1613/jair.1.14494.
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          [26]
          <string-name>
            <given-names>X.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>He</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Liang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Xiao</surname>
          </string-name>
          ,
          <article-title>Language models as knowledge embeddings, 2023</article-title>
          . URL: https: //arxiv.org/abs/2206.12617. arXiv:
          <volume>2206</volume>
          .
          <fpage>12617</fpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>