<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Fairness Analysis of GPT-2's Interpretability with GNN-Generated Knowledge Graph Embeddings Delivered through Soft Prompts</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Ahmad Khalidi</string-name>
          <email>ahmad.khalidi@haw-Hamburg.de</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Thomas Clemen</string-name>
          <email>thomas.clemen@haw-hamburg.de</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Hamburg University of Applied Sciences (HAW Hamburg)</institution>
          ,
          <addr-line>5 Berliner Tor, Hamburg, 20099</addr-line>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>LLM</institution>
          ,
          <addr-line>Knowledge Graph, Graph Representation Learning, Explainable AI, Fairness, Soft Prompt</addr-line>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2026</year>
      </pub-date>
      <abstract>
        <p>Naive fairness criteria applied to machine learning (ML) models, such as low feature importance of sensitive features, are not suficient criteria for formal group fairness. On the contrary, the inclusion of group membership can be justified in order to achieve group fairness. A currently relevant model architecture combining large language model (LLMs) and graph neural networks (GNNs) is suitable for investigating this phenomenon. Not only is this architecture showing promising results in the processing of knowledge graphs (KGs), it also combines two diferent algorithmic perspectives on underlying data. In this paper, we stage a scenario in which a given training dataset contains systematic bias. We assume that when ML models are exposed to this bias during training, their fairness is afected. We propose that a fairness criterion with regard to this bias should take into account systematic processing within the models. We argue that if an ML model is exposed to systematic bias, it should be better able to process this systematic bias in a In a simple scenario using explainable AI (XAI) methods, we show that the hybrid GNN-LLM architecture processes systematic bias in a structured manner.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        LLMs such as Llama3 [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], GPT4 [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ], Gemma2 [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] and DeepSeek [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] are already showing amazing
results in natural language processing (NLP). The processing of structured texts, such as in KGs, is also
being investigated further. In particular, the combination of GNNs and transformer models is enjoying
great popularity and promising success [
        <xref ref-type="bibr" rid="ref5 ref6 ref7">5, 6, 7</xref>
        ]. However, this technology requires transparency and
explainability in order to create trust and fairness and to comply with regulations [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ].
XAI methods can be used under strict conditions for the investigation of fairness in LLMs. The basis
for the successful application of XAI to measure fairness are suitable fairness criteria combined with a
technical understanding of XAI methods and their significance [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]. The fairness criteria applied in this
paper refer to group fairness. First, we demonstrate that a significant proportion of all input edges are
directed toward a small number of nodes (its group membership). We then use XAI methods to show
that this bias is interpreted in a structured manner in the GraphPrompter model. We argue that this
property should be taken into account when evaluating the fairness of this model.
For our GraphPrompter implementation, we decided to use GraphSage [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] for the GNN and
GPT2Medium1 [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] for the Transformer. As a down-stream task, we train the models on link prediction in the
MovieLens dataset [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ]. The structured form of the dataset allows us to design the applied XAI methods
in a concept-based way [
        <xref ref-type="bibr" rid="ref13 ref8">8, 13</xref>
        ]. Our implementation also has tendencies of explainable-by-design, as
the concept-based approach is trained in fine-tuning with the help of segments.
Bologna
CEUR
Workshop
      </p>
      <p>ISSN1613-0073</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related Work</title>
      <sec id="sec-2-1">
        <title>2.1. LLMs (Transformer)</title>
        <p>
          In 2017, [
          <xref ref-type="bibr" rid="ref14">14</xref>
          ] proposed the application of the Transformer architecture in NLP and showed great success
compared to the trends of sequential models. The architecture is largely based on the so-called
multihead-attention mechanism. This allows the learning of relationships between all positions within the
texts and thus creates a semantic understanding of words and their context. The original form of this
transformer architecture consisted of an encoder and decoder. The encoder was intended for embedding
the input text in the vector space and the decoder for the autoregressive generation of the output text.
Modern generative transformer architectures are usually based only on the decoder [
          <xref ref-type="bibr" rid="ref1 ref2 ref3 ref4">1, 2, 3, 4</xref>
          ]. GPT2
[
          <xref ref-type="bibr" rid="ref11">11</xref>
          ] is a decoder-only model that can efectively be understood as a multitask model. The Hugging Face
Transformers [
          <xref ref-type="bibr" rid="ref15">15</xref>
          ] implementation of GPT22 supports text and positional encodings as well as segment
encodings as proposed in BERT [
          <xref ref-type="bibr" rid="ref16">16</xref>
          ].
        </p>
      </sec>
      <sec id="sec-2-2">
        <title>2.2. Graph Representation Learning</title>
        <p>
          Mapping graph structures to low-dimensional vectors [
          <xref ref-type="bibr" rid="ref17 ref18">17, 18</xref>
          ] allows back-end models to process the
complex relationships between entities [
          <xref ref-type="bibr" rid="ref19">19</xref>
          ]. Since KGs encode structured factual knowledge with
multirelational edges [
          <xref ref-type="bibr" rid="ref20">20</xref>
          ], multimodal nodes [
          <xref ref-type="bibr" rid="ref21">21</xref>
          ] and follow logical rules, constraints and ontologies,
we call low-dimensional vector representations of KGs Knowledge Graph Embeddings (KGEs) [
          <xref ref-type="bibr" rid="ref20">20</xref>
          ].
KGs are structured data where entities are related to each other and facts (data points) are referred to as
triple (h,r,t), where h represents the head (source node), r the relationship (edge) between nodes and t
the tail (target node) [
          <xref ref-type="bibr" rid="ref22">22</xref>
          ].
        </p>
        <p>
          In the link prediction task, missing entities (h,r,?) or (?,r,t) are defined as fact triples [
          <xref ref-type="bibr" rid="ref23">23</xref>
          ]. Link prediction
can be used for recommender systems [
          <xref ref-type="bibr" rid="ref24 ref25">24, 25</xref>
          ] or KG completion [
          <xref ref-type="bibr" rid="ref26">26</xref>
          ], among other tasks. This decision
is often based on the similarity between the KGEs of two nodes, whereby nodes that share an edge are
closer to each other than if they do not share an edge [
          <xref ref-type="bibr" rid="ref27">27</xref>
          ]. A typical measure for the similarity of two
KGEs is the dot product [
          <xref ref-type="bibr" rid="ref20">20</xref>
          ], whereby cosine similarity is widely used [
          <xref ref-type="bibr" rid="ref28">28</xref>
          ].
        </p>
        <p>
          GNNs often aggregate the node embeddings of their neighborhood using the message passing concept
[
          <xref ref-type="bibr" rid="ref18">18, 29</xref>
          ]. GraphSage [
          <xref ref-type="bibr" rid="ref10">10</xref>
          ] interprets the message passing concept in the following way
        </p>
        <p>′
ℎ =  1ℎ +  2 ∗ 
∈ () ℎ

matrices  1 and  2.</p>
        <p>
          Where ℎ’ are the KGEs after aggregation over the sampled neighborhood  ()
with the trainable weight
A combination of GNN and Transformer can happen under various aspects. The GraphPrompter
architecture [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ] interprets the GNN as an encoder and the Transformer as a predictor. The KGEs generated
in the GNN are transferred to the Transformer via soft prompts [ 30] within the input embeddings.
The trainable weights of the transformer are frozen during fine-tuning and only the weights of the
GNN are trained. The authors of the paper demonstrate an increase in performance of this
architecture in graph representation learning tasks, but do not show the internal learned strategies of the models.
In a Pytorch Geometrics (PyG) tutorial for Link Prediction3, the authors implement a GraphSage
model [
          <xref ref-type="bibr" rid="ref10">10</xref>
          ], and train it on the MovieLens dataset [
          <xref ref-type="bibr" rid="ref12">12</xref>
          ]. The model was trained to generate KGEs of
nodes connected by an edge with high similarity and KGEs of nodes not connected by an edge with
low similarity. During preprocessing of the MovieLens dataset, all users and movies are assigned
consecutively numbered IDs. The tags are also ignored and the ratings are replaced with a label has
rated and has not rated. The dataset is then split into training, test and validation, whereby some of the
training data is excluded for message passing.
2https://huggingface.co/docs/transformers/v4.52.3/en/model_doc/gpt2#transformers.models.gpt2.modeling_gpt2
3Link accessed 09.08.25, released under MIT License:
https://colab.research.google.com/drive/1xpzn1Nvai1ygd_P5Yambc_oe4VBPK_ZT?usp=sharing
        </p>
      </sec>
      <sec id="sec-2-3">
        <title>2.3. XAI in Transformers</title>
        <p>
          Despite their supposed neutrality, KGs exhibit biases [31]. These can manifest themselves in the models
trained on them [32]. To strengthen trust in this technology, ensure fairness and comply with
regulations, experts and laypersons must be able to understand and explain the underlying mechanisms and
decision-making processes of AI models [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ].
        </p>
        <p>
          XAI methods can be divided into local and global methods, among others, in which the behavior of a
model is described for a specific data point or in general [ 33]. With concept-based XAI methods, input
texts can be mapped to concepts and allow us to summarize positions within the models [
          <xref ref-type="bibr" rid="ref13 ref8">8, 13</xref>
          ]. With
classifier-based probing, a classifier can be trained on a specific task in order to reveal properties in the
internal vector representation [
          <xref ref-type="bibr" rid="ref8">8, 34</xref>
          ].
        </p>
        <p>T-SNE [35] maps high-dimensional vector representations to two- or three-dimensional vector
representations. The mapping is non-linear and keeps vectors close to each other that are close to each other
in high-dimensional space.</p>
        <p>
          In [
          <xref ref-type="bibr" rid="ref13">13</xref>
          ] the authors combined local XAI methods, such as attention maps, with concept-based XAI
methods to aggregate groups of concepts instead of individual items. These concept-based attention
maps allow global insights to be generated in a further aggregation over the entire data set.
        </p>
      </sec>
      <sec id="sec-2-4">
        <title>2.4. Fairness</title>
        <p>
          A common fairness criteria is the absence of discrimination, which is often referred to as statistical
parity [36, 37, 38]. In this context, individuals of a population are divided into groups based on sensitive
features, like gender or race [
          <xref ref-type="bibr" rid="ref9">9, 36, 39</xref>
          ]. Statistical parity in ML is defined as the independence of group
membership in predictions [36]. A common misconception is that removing sensitive features from
the model improves their fairness. This misconception is refereed to as fairness through unawareness.
Sometimes, group membership can be predicted from other attributes [
          <xref ref-type="bibr" rid="ref9">9, 36, 37</xref>
          ] or including group
membership in the decision process has some legitimate and causal cause [
          <xref ref-type="bibr" rid="ref9">9, 36</xref>
          ].
        </p>
        <p>Fair recommender systems operate in a field of tension between utility and exposure [ 40, 41, 42]. From
the perspective of users, a fair recommender system provides them with a wide range of mainstream,
niche and user specific information. From the perspective of providers, a fair recommender system
exposes items equally, independent of the producers’ popularity [41, 42]. Studies that examine the
fairness of recommender systems often focus on the exposure of items with low popularity [ 41, 43, 44],
mitigating the Matthew efect [ 45] and popularity bias [46].</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Model Architecture</title>
      <p>
        The GraphPrompter architecture [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] is based on a GNN and a Transformer model. The GNN generates
KGEs based on a KG and transfers them to the Transformer. This configures the transferred KGEs into
the input embeddings and generates an output based on this. Our implementation of the GNN in link
prediction is based on the PyG tutorial for link prediction. For each link prediction, this means that two
KGEs are produced by GNN and passed to the transformer, one for the start node and one for the target
node.
      </p>
      <p>
        We use the Hugging Face Transformers [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ] implementation of GPT2-medium SequenceClassification 4
as the transformer. To do this, we expand the set of expected segments to include the concepts that we
want to distinguish in our XAI methods. In fine-tuning, we train the transformer to use the segment
embeddings, which, among other things, mark the positions of the KGEs. We base this idea on [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ], in
which the authors have a structured data set in which recurring elements are expected in each data
point. Figure 1 describes the GraphPrompter architecture as we have implemented it for the MovieLens
data set. The movies are reduced to the properties (Movie-ID, Title and Genres). Each position in the
internal vector representation is assigned a segment according to its afiliation. For example, each
position that encodes a separator token is assigned an arbitrary number  _ . The fact quadruples
4https://huggingface.co/docs/transformers/v4.52.3/en/model_doc/gpt2#transformers.GPT2ForSequenceClassification
individual items within the input prompt.
we calculate the word-embedding W⃗ that embeds the KGEs with
batch length,   ∈ ℝ&gt;0 is the sequence length and   ∈ ℝ&gt;0 the hidden size (
Let S⃗ ∈ ℝ  ×  be the segments in batch and W⃗ ∈ ℝ  ×  ×  be word embeddings, while   ∈ ℝ&gt;0 is the
 = 1024 for   2
), then
T⃗ = {
1, if S⃗ = S
0, otherwise
      </p>
      <p>M⃗ = T⃗  1⃗d
⃖K⃖⃖⃖G⃖⃖⃖E⃗
 = ⃖k⃖⃖⃖g⃖⃖e⃗ ⃗
1



W⃗ = ( W⃗ ⊙ (1⃗ − M⃗ ) + ⃖K⃖⃖⃖G⃖⃖⃖E⃗ ⊙ M⃗ ) ⊙ (1⃗ − M⃗  ) + ⃖K⃖⃖⃖G⃖⃖⃖E⃗ ⊙ M⃗

(1)
(2)
(3)
(4)
while
 ∈ [, ]
⃗
1


is the unit vector of dimension   ,</p>
      <p>is an index for either user or movie,
S⃗ ∈ ℝ are the segments for user and movie (here 2 and 3),
Deep Bidirectional Transformers for Language Understanding, by Devlin et al., 2019</p>
      <p>is the unit vector of dimension   ,
1⃗ is the unit vector of dimension (  ×   ×   ).
⃖k⃖⃖⃖g⃖⃖e⃗ ∈ ℝ  ×  are the user and movie KGEs of dimension   distributed over the entire batch   and
In equation 1, we filter the segments in the batch by the segments of the KGEs and thus create a mask for
the respective positions. Then we scale both segment masks T⃗ via the hidden states   and both KGEs
via the sequence length   in the equations 2 and 3 by repeating. In equation 4, we then superimpose
the masked and scaled KGEs with the inverse-masked word embeddings and thus obtain the word
embedding with the KGEs at the placeholder positions.</p>
    </sec>
    <sec id="sec-4">
      <title>4. Segment-based XAI</title>
      <p>
        following applies
In this paper, we first focus on the XAI methods dimension-reduction of the hidden states. As suggested
in [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ], concept-based XAI methods can be used to aggregate all positions of the same segments in
order to obtain a semantic-oriented view. For the hidden states, we calculate the average of all positions
in which the sequence is the same. Let 
be the set of all segments,  ∈ 
an element from

,   ∈ ℝ&gt;0 the number of layers (here 24), H⃗ ∈ ℝ  ×  ×  ×  the hidden states of a batch, then the
H⃗ =
∑ ( H⃗ ⊙ (M⃗  1⃗

 ))

∑ (T⃗ 1⃗d )


(5)
In equation 5, we first mask the hidden states
      </p>
      <p>H⃗ and then add them up over the entire sequence length
 . We then divide the sum by the number of all positions at which the segment is located and thus

obtain the average of all hidden states from the same segment H⃗. Figure 2 illustrates a segment-based
XAI method in aggregating hidden states of the transformer model. Let’s assume that the movie title is
“SomeTitle” and will be divided into two parts in the vector representation: “Some##” and “##Title”. To
summarize and examine the internal vector representation of all positions related to the title, we filter
the segments and calculate the average using the masked output embeddings of the transformer.</p>
    </sec>
    <sec id="sec-5">
      <title>5. Experiments</title>
      <p>5.1. Goal
We will prove that there is a systematic bias in the MovieLens dataset, namely that the distribution
of the incoming edges of all movie nodes is subject to strong fluctuations. We use XAI methods to
demonstrate that the GraphPrompter model systematically processes the systematic bias and follows a
subjectively reasonable processing strategy.
5.2. Set-Up</p>
      <sec id="sec-5-1">
        <title>5.2.1. Hyper Parameters and Hardware</title>
        <p>
          For GraphSage, we adopt the architecture of the PyG tutorial, matching the layer size to GPT2-medium’s
hidden size 1024 on the output layer and thus also increasing the input layer size to 256. For the
linkedneighborhood sampler we keep 20 neighbors in the first hop and 10 neighbors in the second hop. For
training the GNN, we increase the batch size to 25600 in order to speed up training process The models
were each trained for 2 epochs. We chose the largest possible batch size for the LLM so that it can still
be trained on a single GPU which was 128. Training and forwarding took place on an Nvidia A100
SXM4 40 GB and lasted approximately four days in total.
5.2.2. Dataset
The authors of the MovieLens dataset [
          <xref ref-type="bibr" rid="ref12">12</xref>
          ] collected movie ratings over a publicly available website5.
MovieLens 32M includes 32 million movie ratings from 87,585 movies and 200,948 independent users
collected in 2023. Each user is identified by a random (user) ID. Each movie is represented by an
arbitrary (movie) ID, a title with year of release and a list of genres and tags. The users added tags and
ratings to the movies, resulting in facts of the form (user ID, movie ID, rating, timestamp).
        </p>
      </sec>
      <sec id="sec-5-2">
        <title>5.2.3. Data Preprocessing</title>
        <p>The preprocessing is largely based on the steps in the PyG Link Prediction Tutorial. First, new sequential
IDs are assigned to the users and movies. The genres are then one-hot encoded and transferred together
with the user and movie IDs and the labels has rated into a heterogeneous graph dataset. This is then
split 70-15-15 into training, test and validation and the test and validation splits are uniformly assigned
negative samples of not rated in a 1:1 ratio. In the PyG tutorial, 70% of the training data is only separated
for message passing. We refrain from this separation so that we can perform a uniform mapping of the
data set into the NLP data set.</p>
        <p>The NLP data set is made up of the newly assigned IDs, the movie titles and genres in a fixed structure.
To enable the Transformer model to make decisions with or without KGEs, we introduce a special
placeholder token &lt;user KGE&gt; and &lt;movie KGE&gt;, as well as a separator token &lt;SEP&gt; for separation. This
results in the schema:
&lt;user_KGE&gt;&lt;SEP&gt;user ID&lt;SEP&gt;&lt;movie_KGE&gt;&lt;SEP&gt;movie ID&lt;SEP&gt;movie title&lt;SEP&gt;[movie genres]&lt;SEP&gt;.
We create the segments for each of these data points. We must note that not every feature has the
same length in encoded form. During tokenization, we also create the corresponding segments. The
segments that we want to keep apart are the separator tokens &lt;SEP&gt;, those of the KGEs, the user ID and
the NLP movie features movie-ID, title and genres. The NLP dataset is then compared with the graph
dataset and split accordingly.</p>
      </sec>
      <sec id="sec-5-3">
        <title>5.2.4. Training</title>
        <p>First, we train the GraphSage model and the transformer-Only model for two epochs and evaluate
their accuracy against the test dataset. Since our GraphSage model is trained as a regression problem,
5Permalink: https://grouplens.org/datasets/movielens/32m/
(a) The comulative distribution of in- (b) Training losses of Transformer- (c) Training Accuracy on the test
coming (positive) edges on the Only and GraphPrompter models dataset after every epoch of
trainset of all movies ing GNN, Transformer-Only and
GraphPrompter
we assume discrete thresholds that best map the regression to the classification. Finally, we train and
evaluate the Graphprompter model based on the pre-trained GraphSage and Transformer-only model
over two epochs. From the training process, we map the loss and the accuracy on a graph.</p>
        <sec id="sec-5-3-1">
          <title>5.3. Results</title>
        </sec>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>6. Discussion</title>
      <p>
        The experiments carried out in this paper show that there is a systematic bias in the MovieLens dataset
[
        <xref ref-type="bibr" rid="ref12">12</xref>
        ]. A few movies hold a majority of the ratings. This bias was established in our GraphPrompter
architecture [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] and can be clearly distinguished in its internal vector representation from the KGEs
passed in soft prompts. We interpret this separation as a sign that the model is able to interpret the
underlying bias in a systematic way. We can assume that this distinction can help the model to decide
between the features produced in the GNN and those produced in the transformer. We assume that the
GNN learns (global) neighborhood structures during training with neighborhood Sampling. We also
assume that the transformer tends to learn local contexts. We further assume that the Transformer-Only
model has a harder time deciding on movies with high input degree, since a global view is needed to
diferentiate movies with high input degree. In the GraphPrompter model, we combine the strategies of
the local and global view of both model architectures. It is possible for the GraphPrompter model to
refer to the features generated in the GNN if it makes the decision for a movie with a high input degree
and to the transformer-generated features when making decision for a movie with low input degree.
If the GraphPrompter model actually learns such a strategy, then this architecture may lend itself to the
(a) Hidden states of transformer-only model
      </p>
      <p>(b) Hidden states of GraphPrompter
use of a bias mitigation strategy or other fairness strategies. The discrimination in the model can also
be used to favor underrepresented movies, e.g. to promote cultural diversity.</p>
    </sec>
    <sec id="sec-7">
      <title>7. Limitations</title>
      <p>Figure 4 could be misinterpreted as indicating a lack of structured processing of bias. However, we have
only demonstrated that this processing is unlikely to occur in this particular segment. Comparatively
structured processing could still occur in other segments of the transformer. We can therefore only
conclude that certain properties are present within the GraphPrompter model.</p>
      <p>The behaviors of GNN, transformer and GraphPrompter model assumed in this work still have many
gaps. The assumption that our transformer learns fewer global structures is based solely on the
assumption that it only ever sees one data point at each time during training. Though, the transformer
in this experiment is significantly more powerful than the GNN. Nevertheless, an investigation of the
actual scopes of all models is appropriate.</p>
      <p>The question remains whether the structured processing of an underlying bias in the dataset is more
appropriate for fair processing. We assume that such a crucial bias will inevitably have an impact on
models trained on it and that structured processing tends to make the system more fair. However,
for a more thorough discussion of the fairness of this GraphPrompter implementation, more concrete
fairness criteria are needed.</p>
    </sec>
    <sec id="sec-8">
      <title>8. Conclusion</title>
      <p>The systematic bias in the MovieLens dataset prevails in the internal vector representation of our
GraphPrompter model. We argue that the structured processing of this bias rather indicates behavior
that increases group fairness. However, further experiments and more concrete fairness criteria are
needed to evaluate the practical implications of this perspective.</p>
    </sec>
    <sec id="sec-9">
      <title>Declaration on Generative AI</title>
      <p>During the preparation of this work, the authors used ChatGPT (GPT-4o) for generating small chunks
of Python code, which were adjusted and used in the experiments. Further, the author(s) used DeepL
for the translation from German into English. After using these services, the authors reviewed and
edited the content as needed and take full responsibility for the publication’s content.
in Natural Language Processing and the 9th International Joint Conference on Natural Language
Processing (EMNLP-IJCNLP), Association for Computational Linguistics, Hong Kong, China, 2019,
pp. 3982–3992. doi:10.18653/v1/D19-1410.
[29] S. Liang, Knowledge graph embedding based on graph neural network, in: 2023 IEEE 39th
International Conference on Data Engineering (ICDE), 2023, pp. 3908–3912. doi:10.1109/ICDE55515.
2023.00379.
[30] B. Lester, R. Al-Rfou, N. Constant, The power of scale for parameter-eficient prompt tuning, in:
M.F. Moens, X. Huang, L. Specia, S. W.-t. Yih (Eds.), Proceedings of the 2021 Conference on Empirical
Methods in Natural Language Processing, Association for Computational Linguistics, Online and
Punta Cana, Dominican Republic, 2021, pp. 3045–3059. doi:10.18653/v1/2021.emnlp-main.243.
[31] A. Kraft, R. Usbeck, The lifecycle of “facts”: A survey of social bias in knowledge graphs, in: Y. He,
H. Ji, S. Li, Y. Liu, C.-H. Chang (Eds.), Proceedings of the 2nd Conference of the Asia-Pacific Chapter
of the Association for Computational Linguistics and the 12th International Joint Conference on
Natural Language Processing (Volume 1: Long Papers), Association for Computational Linguistics,
Online only, 2022, pp. 639–652. doi:10.18653/v1/2022.aacl-main.49.
[32] R. Fernando, I. Norton, P. Dogra, R. Sarnaik, H. Wazir, Z. Ren, N. S. Gunda, A. Mukhopadhyay,
M. Lutz, Quantifying bias in agentic large language models: A benchmarking approach, in:
2024 5th Information Communication Technologies Conference (ICTC), 2024, pp. 349–353. doi:10.
1109/ICTC61510.2024.10601938.
[33] T. Speith, A review of taxonomies of explainable artificial intelligence (xai) methods, in:
Proceedings of the 2022 ACM Conference on Fairness, Accountability, and Transparency,
FAccT ’22, Association for Computing Machinery, New York, NY, USA, 2022, p. 2239–2250.
doi:10.1145/3531146.3534639.
[34] E. N. Volkov, A. N. Averkin, Local explanations for large language models: a brief review of
methods, in: 2024 XXVII International Conference on Soft Computing and Measurements (SCM),
2024, pp. 189–192. doi:10.1109/SCM62608.2024.10554222.
[35] L. van der Maaten, G. E. Hinton, Visualizing data using t-sne, Journal of Machine Learning</p>
      <p>Research 9 (2008) 2579–2605.
[36] S. Barocas, M. Hardt, A. Narayanan, Fairness and Machine Learning: Limitations and Opportunities,</p>
      <p>MIT Press, 2023.
[37] C. Dwork, M. Hardt, T. Pitassi, O. Reingold, R. Zemel, Fairness through awareness, in: Proceedings
of the 3rd Innovations in Theoretical Computer Science Conference, ITCS ’12, Association for
Computing Machinery, New York, NY, USA, 2012, p. 214–226. URL: https://doi.org/10.1145/2090236.
2090255. doi:10.1145/2090236.2090255.
[38] L. Baresi, C. Criscuolo, C. Ghezzi, Understanding fairness requirements for ml-based software,
in: 2023 IEEE 31st International Requirements Engineering Conference (RE), 2023, pp. 341–346.
doi:10.1109/RE57278.2023.00046.
[39] Z. Chu, Z. Wang, W. Zhang, Fairness in large language models: A taxonomic survey, SIGKDD
Explor. Newsl. 26 (2024) 34–48. URL: https://doi.org/10.1145/3682112.3682117. doi:10.1145/3682112.
3682117.
[40] A. Singh, T. Joachims, Fairness of exposure in rankings, in: Proceedings of the 24th ACM SIGKDD
International Conference on Knowledge Discovery &amp; Data Mining, KDD ’18, Association for
Computing Machinery, New York, NY, USA, 2018, p. 2219–2228. URL: https://doi.org/10.1145/
3219819.3220088. doi:10.1145/3219819.3220088.
[41] Y. Deldjoo, D. Jannach, A. Bellogin, A. Difonzo, D. Zanzonelli, Fairness in recommender systems:
research landscape and future directions, User Modeling and User-Adapted Interaction 34 (2023)
59–108. URL: https://doi.org/10.1007/s11257-023-09364-z. doi:10.1007/s11257-023-09364-z.
[42] Y. Wang, W. Ma, M. Zhang, Y. Liu, S. Ma, A survey on the fairness of recommender systems, ACM</p>
      <p>Trans. Inf. Syst. 41 (2023). URL: https://doi.org/10.1145/3547333. doi:10.1145/3547333.
[43] Q. Dong, S.-S. Xie, W.-J. Li, User-item matching for recommendation fairness, IEEE Access 9 (2021)
130389–130398. doi:10.1109/ACCESS.2021.3113975.
[44] Y. Li, H. Chen, Z. Fu, Y. Ge, Y. Zhang, User-oriented fairness in recommendation, in: Proceedings of
the Web Conference 2021, WWW ’21, Association for Computing Machinery, New York, NY, USA,
2021, p. 624–632. URL: https://doi.org/10.1145/3442381.3449866. doi:10.1145/3442381.3449866.
[45] R. K. Merton, The matthew efect in science, Science 159 (1968) 56–63. URL:
https://www.science.org/doi/abs/10.1126/science.159.3810.56. doi:10.1126/science.159.3810.
56. arXiv:https://www.science.org/doi/pdf/10.1126/science.159.3810.56.
[46] Z. Zhu, Y. He, X. Zhao, Y. Zhang, J. Wang, J. Caverlee, Popularity-opportunity bias in collaborative
ifltering, in: Proceedings of the 14th ACM International Conference on Web Search and Data
Mining, WSDM ’21, Association for Computing Machinery, New York, NY, USA, 2021, p. 85–93.
URL: https://doi.org/10.1145/3437963.3441820. doi:10.1145/3437963.3441820.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>Llama</given-names>
            <surname>Team</surname>
          </string-name>
          ,
          <article-title>AI @ Meta, The llama 3 herd of models</article-title>
          ,
          <source>AI @ Meta</source>
          (
          <year>2024</year>
          ). URL: https://ai.meta.com/ research/publications/the-llama-3
          <string-name>
            <surname>-</surname>
          </string-name>
          herd-of-models/.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2] OpenAI, Gpt-4
          <source>technical report</source>
          ,
          <source>OpenAI</source>
          (
          <year>2023</year>
          ). URL: https://cdn.openai.com/papers/gpt-4.pdf.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>Gemma</given-names>
            <surname>Team</surname>
          </string-name>
          ,
          <source>Google DeepMind, Gemma</source>
          <volume>2</volume>
          :
          <article-title>Improving open language models at a practical size</article-title>
          ,
          <source>ArXiv preprint arXiv:2408.00118</source>
          (
          <year>2024</year>
          ). URL: https://arxiv.org/pdf/2408.00118.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <surname>DeepSeek-AI</surname>
          </string-name>
          ,
          <article-title>Deepseek-r1: Incentivizing reasoning capability in llms via reinforcement learning</article-title>
          ,
          <source>ArXiv preprint arXiv:2501.12948</source>
          (
          <year>2025</year>
          ). URL: https://github.com/deepseek-ai/DeepSeek-R1.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>B.</given-names>
            <surname>Jin</surname>
          </string-name>
          , G. Liu, C. Han,
          <string-name>
            <given-names>M.</given-names>
            <surname>Jiang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Ji</surname>
          </string-name>
          , J. Han,
          <article-title>Large language models on graphs: A comprehensive survey</article-title>
          ,
          <source>IEEE Transactions on Knowledge and Data Engineering</source>
          <volume>36</volume>
          (
          <year>2024</year>
          )
          <fpage>8622</fpage>
          -
          <lpage>8642</lpage>
          . doi:
          <volume>10</volume>
          .1109/ TKDE.
          <year>2024</year>
          .
          <volume>3469578</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>L.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Shen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Guo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Gao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Pei</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Long</surname>
          </string-name>
          ,
          <article-title>Graph neural networks for natural language processing: A survey</article-title>
          ,
          <source>Foundations and Trends® in Machine Learning</source>
          <volume>16</volume>
          (
          <year>2023</year>
          )
          <fpage>119</fpage>
          -
          <lpage>328</lpage>
          . URL: http://dx.doi.org/10.1561/2200000096. doi:
          <volume>10</volume>
          .1561/2200000096.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>He</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Tian</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N. V.</given-names>
            <surname>Chawla</surname>
          </string-name>
          ,
          <article-title>Can we soft prompt llms for graph learning tasks?</article-title>
          ,
          <source>in: Companion Proceedings of the ACM Web Conference</source>
          <year>2024</year>
          , WWW '24,
          <string-name>
            <surname>Association</surname>
          </string-name>
          for Computing Machinery, New York, NY, USA,
          <year>2024</year>
          , p.
          <fpage>481</fpage>
          -
          <lpage>484</lpage>
          . URL: https://doi.org/10.1145/3589335.3651476. doi:
          <volume>10</volume>
          .1145/3589335.3651476.
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>H.</given-names>
            <surname>Zhao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Deng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Cai</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Yin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Du</surname>
          </string-name>
          ,
          <article-title>Explainability for large language models: A survey</article-title>
          ,
          <source>ACM Trans. Intell. Syst. Technol</source>
          .
          <volume>15</volume>
          (
          <year>2024</year>
          ). doi:
          <volume>10</volume>
          .1145/3639372.
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>L.</given-names>
            <surname>Deck</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Schoefer</surname>
          </string-name>
          ,
          <string-name>
            <surname>M. De-Arteaga</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          <string-name>
            <surname>Kühl</surname>
          </string-name>
          ,
          <article-title>A critical survey on fairness benefits of explainable ai</article-title>
          ,
          <source>in: Proceedings of the 2024 ACM Conference on Fairness, Accountability, and Transparency</source>
          , FAccT '24,
          <string-name>
            <surname>Association</surname>
          </string-name>
          for Computing Machinery, New York, NY, USA,
          <year>2024</year>
          , p.
          <fpage>1579</fpage>
          -
          <lpage>1595</lpage>
          . doi:
          <volume>10</volume>
          . 1145/3630106.3658990.
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>W. L.</given-names>
            <surname>Hamilton</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Ying</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Leskovec</surname>
          </string-name>
          ,
          <article-title>Inductive representation learning on large graphs</article-title>
          ,
          <source>in: Proceedings of the 31st International Conference on Neural Information Processing Systems</source>
          , NIPS'17, Curran Associates Inc.,
          <string-name>
            <surname>Red</surname>
            <given-names>Hook</given-names>
          </string-name>
          ,
          <string-name>
            <surname>NY</surname>
          </string-name>
          , USA,
          <year>2017</year>
          , p.
          <fpage>1025</fpage>
          -
          <lpage>1035</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>A.</given-names>
            <surname>Radford</surname>
          </string-name>
          ,
          <string-name>
            <surname>J. Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Child</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Luan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Amodei</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Sutskever</surname>
          </string-name>
          ,
          <article-title>Language models are unsupervised multitask learners</article-title>
          ,
          <source>OpenAI</source>
          (
          <year>2019</year>
          ). URL: https://cdn.openai.
          <article-title>com/better-language-models/ language_models_are_unsupervised_multitask_learners</article-title>
          .pdf.
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>F. M.</given-names>
            <surname>Harper</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. A.</given-names>
            <surname>Konstan</surname>
          </string-name>
          ,
          <article-title>The movielens datasets: History and context</article-title>
          ,
          <source>ACM Trans. Interact. Intell. Syst</source>
          .
          <volume>5</volume>
          (
          <year>2015</year>
          ). URL: https://doi.org/10.1145/2827872. doi:
          <volume>10</volume>
          .1145/2827872.
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>A. H.</given-names>
            <surname>Mohammadkhani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Tantithamthavorn</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Hemmatif</surname>
          </string-name>
          ,
          <article-title>Explaining transformer-based code models: What do they learn? when they do not work?</article-title>
          ,
          <source>in: 2023 IEEE 23rd International Working Conference on Source Code Analysis and Manipulation (SCAM)</source>
          ,
          <year>2023</year>
          , pp.
          <fpage>96</fpage>
          -
          <lpage>106</lpage>
          . doi:
          <volume>10</volume>
          .1109/SCAM59687.
          <year>2023</year>
          .
          <volume>00020</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>A.</given-names>
            <surname>Vaswani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Shazeer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Parmar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Uszkoreit</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Jones</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. N.</given-names>
            <surname>Gomez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Kaiser</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Polosukhin</surname>
          </string-name>
          ,
          <article-title>Attention is all you need</article-title>
          ,
          <source>in: Proceedings of the 31st International Conference on Neural Information Processing Systems</source>
          , NIPS'17, Curran Associates Inc.,
          <string-name>
            <surname>Red</surname>
            <given-names>Hook</given-names>
          </string-name>
          ,
          <string-name>
            <surname>NY</surname>
          </string-name>
          , USA,
          <year>2017</year>
          , p.
          <fpage>6000</fpage>
          -
          <lpage>6010</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>T.</given-names>
            <surname>Wolf</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Debut</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Sanh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Chaumond</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Delangue</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Moi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Cistac</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Rault</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Louf</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Funtowicz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Davison</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Shleifer</surname>
          </string-name>
          , P. von Platen, C. Ma,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Jernite</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Plu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T. Le</given-names>
            <surname>Scao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Gugger</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Drame</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Lhoest</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Rush</surname>
          </string-name>
          , Transformers:
          <article-title>State-of-the-art natural language processing</article-title>
          , in: Q. Liu, D. Schlangen (Eds.),
          <source>Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: System Demonstrations</source>
          ,
          <article-title>Association for Computational Linguistics</article-title>
          , Online,
          <year>2020</year>
          , pp.
          <fpage>38</fpage>
          -
          <lpage>45</lpage>
          . doi:
          <volume>10</volume>
          .18653/v1/
          <year>2020</year>
          .emnlp-demos.
          <volume>6</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>J.</given-names>
            <surname>Devlin</surname>
          </string-name>
          , M.-
          <string-name>
            <given-names>W.</given-names>
            <surname>Chang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Toutanova</surname>
          </string-name>
          , BERT:
          <article-title>Pre-training of deep bidirectional transformers for language understanding</article-title>
          , in: J.
          <string-name>
            <surname>Burstein</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Doran</surname>
          </string-name>
          , T. Solorio (Eds.),
          <source>Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies</source>
          , Volume
          <volume>1</volume>
          (Long and Short Papers),
          <source>Association for Computational Linguistics</source>
          , Minneapolis, Minnesota,
          <year>2019</year>
          , pp.
          <fpage>4171</fpage>
          -
          <lpage>4186</lpage>
          . doi:
          <volume>10</volume>
          .18653/v1/
          <fpage>N19</fpage>
          -1423.
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>W.</given-names>
            <surname>Ju</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Fang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Gu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Long</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Qiao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Qin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Shen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Sun</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Xiao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Yuan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Luo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <article-title>A comprehensive survey on deep graph representation learning</article-title>
          ,
          <source>Neural Networks</source>
          <volume>173</volume>
          (
          <year>2024</year>
          )
          <article-title>106207</article-title>
          . doi:https://doi.org/10.1016/j.neunet.
          <year>2024</year>
          .
          <volume>106207</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>S.</given-names>
            <surname>Khoshraftar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>An</surname>
          </string-name>
          ,
          <article-title>A survey on graph representation learning methods</article-title>
          ,
          <source>ACM Trans. Intell. Syst. Technol</source>
          .
          <volume>15</volume>
          (
          <year>2024</year>
          ). doi:
          <volume>10</volume>
          .1145/3633518.
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>K.</given-names>
            <surname>Liang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Meng</surname>
          </string-name>
          , M. Liu,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Tu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Sun</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>He</surname>
          </string-name>
          ,
          <article-title>A survey of knowledge graph reasoning on graph types: Static, dynamic, and multi-modal</article-title>
          ,
          <source>IEEE Transactions on Pattern Analysis and Machine Intelligence</source>
          <volume>46</volume>
          (
          <year>2024</year>
          )
          <fpage>9456</fpage>
          -
          <lpage>9478</lpage>
          . doi:
          <volume>10</volume>
          .1109/TPAMI.
          <year>2024</year>
          .
          <volume>3417451</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>J.</given-names>
            <surname>Cao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Fang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Meng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Liang</surname>
          </string-name>
          ,
          <article-title>Knowledge graph embedding: A survey from the perspective of representation spaces</article-title>
          ,
          <source>ACM Comput. Surv</source>
          .
          <volume>56</volume>
          (
          <year>2024</year>
          ). doi:
          <volume>10</volume>
          .1145/3643806.
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>R.</given-names>
            <surname>Sun</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Cao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Wan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <surname>K.</surname>
          </string-name>
          <article-title>Zheng, Multi-modal knowledge graphs for recommender systems</article-title>
          ,
          <source>in: Proceedings of the 29th ACM International Conference on Information &amp; Knowledge Management, CIKM '20</source>
          ,
          <string-name>
            <surname>Association</surname>
          </string-name>
          for Computing Machinery, New York, NY, USA,
          <year>2020</year>
          , p.
          <fpage>1405</fpage>
          -
          <lpage>1414</lpage>
          . doi:
          <volume>10</volume>
          .1145/3340531.3411947.
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>S.</given-names>
            <surname>Ji</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Pan</surname>
          </string-name>
          , E. Cambria,
          <string-name>
            <given-names>P.</given-names>
            <surname>Marttinen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. S.</given-names>
            <surname>Yu</surname>
          </string-name>
          ,
          <article-title>A survey on knowledge graphs: Representation, acquisition, and applications</article-title>
          ,
          <source>IEEE Transactions on Neural Networks and Learning Systems</source>
          <volume>33</volume>
          (
          <year>2022</year>
          )
          <fpage>494</fpage>
          -
          <lpage>514</lpage>
          . doi:
          <volume>10</volume>
          .1109/TNNLS.
          <year>2021</year>
          .
          <volume>3070843</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <given-names>F.</given-names>
            <surname>Akrami</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Guo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Hu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <article-title>Re-evaluating embedding-based knowledge graph completion methods</article-title>
          ,
          <source>in: Proceedings of the 27th ACM International Conference on Information and Knowledge Management</source>
          , CIKM '18,
          <string-name>
            <surname>Association</surname>
          </string-name>
          for Computing Machinery, New York, NY, USA,
          <year>2018</year>
          , p.
          <fpage>1779</fpage>
          -
          <lpage>1782</lpage>
          . doi:
          <volume>10</volume>
          .1145/3269206.3269266.
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [24]
          <string-name>
            <given-names>X.</given-names>
            <surname>Han</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L</given-names>
            .
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. N.</given-names>
            <surname>Han</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Crespi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Farahbakhsh</surname>
          </string-name>
          ,
          <article-title>Link prediction for new users in social networks</article-title>
          ,
          <source>in: 2015 IEEE International Conference on Communications (ICC)</source>
          ,
          <year>2015</year>
          , pp.
          <fpage>1250</fpage>
          -
          <lpage>1255</lpage>
          . doi:
          <volume>10</volume>
          .1109/ICC.
          <year>2015</year>
          .
          <volume>7248494</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          [25]
          <string-name>
            <given-names>X.</given-names>
            <surname>Song</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Lian</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Huang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Jin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Xie</surname>
          </string-name>
          ,
          <article-title>Friend recommendations with self-rescaling graph neural networks</article-title>
          ,
          <source>in: Proceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining</source>
          , KDD '22,
          <string-name>
            <surname>Association</surname>
          </string-name>
          for Computing Machinery, New York, NY, USA,
          <year>2022</year>
          , p.
          <fpage>3909</fpage>
          -
          <lpage>3919</lpage>
          . URL: https://doi.org/10.1145/3534678.3539192. doi:
          <volume>10</volume>
          .1145/3534678.3539192.
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          [26]
          <string-name>
            <given-names>F.</given-names>
            <surname>Akrami</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Guo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Hu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <article-title>Re-evaluating embedding-based knowledge graph completion methods</article-title>
          ,
          <source>in: Proceedings of the 27th ACM International Conference on Information and Knowledge Management</source>
          , CIKM '18,
          <string-name>
            <surname>Association</surname>
          </string-name>
          for Computing Machinery, New York, NY, USA,
          <year>2018</year>
          , p.
          <fpage>1779</fpage>
          -
          <lpage>1782</lpage>
          . doi:
          <volume>10</volume>
          .1145/3269206.3269266.
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          [27]
          <string-name>
            <given-names>M.</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Yin</surname>
          </string-name>
          ,
          <article-title>A similarity index algorithm for link prediction</article-title>
          ,
          <source>in: 2017 12th International Conference on Intelligent Systems and Knowledge Engineering (ISKE)</source>
          ,
          <year>2017</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>6</lpage>
          . doi:
          <volume>10</volume>
          .1109/ ISKE.
          <year>2017</year>
          .
          <volume>8258724</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          [28]
          <string-name>
            <given-names>N.</given-names>
            <surname>Reimers</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Gurevych</surname>
          </string-name>
          , Sentence-BERT:
          <article-title>Sentence embeddings using Siamese BERT-networks</article-title>
          , in: K. Inui,
          <string-name>
            <given-names>J.</given-names>
            <surname>Jiang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Ng</surname>
          </string-name>
          ,
          <string-name>
            <surname>X.</surname>
          </string-name>
          Wan (Eds.),
          <source>Proceedings of the 2019 Conference on Empirical Methods</source>
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>