<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>Visits</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Path of Time: Explanations for Temporal Knowledge Graph Completion through Chronological Regulation</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Lukas Gehring</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Moritz Blum</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Basil Ell</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Philipp Cimiano</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Bielefeld University</institution>
          ,
          <addr-line>CITEC, Inspiration 1, 33619, Bielefeld</addr-line>
          ,
          <country country="DE">Germany</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>University of Oslo</institution>
          ,
          <addr-line>Problemveien 11, 0313 Oslo</addr-line>
          ,
          <country country="NO">Norway</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2024</year>
      </pub-date>
      <volume>202</volume>
      <fpage>4</fpage>
      <lpage>12</lpage>
      <abstract>
        <p>Temporal Knowledge Graph Completion (TKGC) uses the facts available in a TKG to make it less incomplete. Stateof-the-art Graph Neural Networks (GNNs) for TKGC are black boxes that provide results without explanations. Existing explanation methods for static KGC are dificult to transfer to TKGC as they do not capture the temporal properties and likely generate large explanation graphs. As the chronological order of facts is relevant for TKGC, we infuse this characteristic into the explanation subgraphs. In this work, we (i) propose a regulation method that incentivizes a chronological order in the explanations and (ii) investigate the efect of the chronological regulation on the explanations of two state-of-theart TKGC models. Our results show that in most scenarios, the chronological regulation can improve explanations of TKGCs. For example, we observe an improvement of the fidelity characterization score by up to 2% and significant improvements for small explanations.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;XAI</kwd>
        <kwd>Temporal Knowledge Graph Completion</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>Temporal Knowledge
Graph Completion</p>
      <p>Interpolation
Extrapolation</p>
      <p>Timestamps
Dependent</p>
      <p>based
Timestamps-specific
functions-based
Deep Learning-based</p>
      <p>Angela
Merkel</p>
      <p>Consult
2014-05-01</p>
      <p>
        0.7
Consult
path-based explanations used for XAI of static KG approaches [
        <xref ref-type="bibr" rid="ref5 ref6">5, 6</xref>
        ], our regulation method
encourages a chronological order of the facts in the explanations; iii) investigate the efect of chronological
regulation on the explanations of two state-of-the-art TKGC models. We evaluate explanation quality
using common metrics and introduce a new metric better suited for graphs in the temporal setting than
existing metrics.
      </p>
    </sec>
    <sec id="sec-2">
      <title>2. Foundations</title>
      <sec id="sec-2-1">
        <title>2.1. Temporal Knowledge Graph Completion</title>
        <p>A Knowledge Graph (KG) stores facts as triples (, , ), where  ∈ ℰ is called the subject,  ∈ ℛ the
relation, and  ∈ ℰ the object. ℰ and ℛ are finite sets of entity and relation identifiers, respectively.</p>
        <p>A Temporal Knowledge Graph (TKG) is a KG extended by the temporal information about the facts.
 is a specific point in time (e.g. 05-11-2014) from a finite set of timestamps  .</p>
        <p>Facts in a TKG are represented as quadruples (, , , ), with  ∈  adding time information to the fact.</p>
        <p>Temporal Knowledge Graph Completion (TKGC) is about adding missing quadruples to a TKG.
TKGC models predict a missing entity of a given query  = (, , ?, ) or  = (?, , , ), where
, , ? ∈ ℰ ,  ∈ ℛ, and  ∈  .</p>
      </sec>
      <sec id="sec-2-2">
        <title>2.2. Graph Neural Networks for TKGC</title>
        <p>neighborhood.
updated node features x^ as</p>
        <sec id="sec-2-2-1">
          <title>Graph Neural Networks (GNNs)</title>
          <p>are a type of neural network designed to process graphs as input.</p>
          <p>
            The core concept of a GNN is message-passing, first introduced by Gilmer et al. [
            <xref ref-type="bibr" rid="ref7">7</xref>
            ], which enables
the GNN to learn node embeddings that capture its features but also include information from the
Given a node , with its features x and its incoming neighborhood , message passing computes
︂(
x^ =  x, ⊕
 (x, x ) ,
          </p>
          <p>
            ︂)
∈
(1)
where message function  and update function  are trainable functions and ⊕ denotes a nonparametric
operation such as sum, mean, or maximum [
            <xref ref-type="bibr" rid="ref8">8</xref>
            ].
          </p>
          <p>In a GNN, this message passing scheme is typically repeated layerwise and can include edge features
e for each edge connecting node  and . Two commonly used GNNs are the Graph Attention Networks
(GATs) and the Graph Convolution Networks (GCNs).</p>
        </sec>
      </sec>
      <sec id="sec-2-3">
        <title>2.3. GNN Explainability</title>
        <p>We focus on post-hoc explanations, which are generated for a target model ℳ. We utilize a
perturbationbased method that modifies the model’s input by masking to identify minimal subgraphs that explain
the model’s prediction.</p>
        <p>
          One of the initial and well-known perturbation-based GNN explainer models for non-temporal KGs is
the GNNExplainer [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ]. This model is designed to identify the subgraph and node features most relevant
to a GNN’s prediction, by applying a learnable mask  ∈ [
          <xref ref-type="bibr" rid="ref1 ref21">0, 1</xref>
          ]|ℰ|×|ℛ|×|ℰ| to the graph’s adjacency
matrix , to minimize the following cross-entropy objective:
min −

∑︁ 1[ = ] log ℳ( = | =  ⊙  ( )).
=1
(2)
Here, 1[ = ] is the indicator function for the target class , ℳ is the probability of target model ℳ
predicting , and  ( ) maps the mask to a continuous range [
          <xref ref-type="bibr" rid="ref1 ref21">0, 1</xref>
          ]. The framework learns the mask 
to minimize the conditional entropy of the predictions when restricted to the masked subgraph. Sparse
explanations are encouraged through regularization terms, and additional thresholds can be applied to
refine the resulting subgraph, retaining only the most important edges and nodes.
        </p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Related Work</title>
      <p>XAI aims to help humans understand the predictions of neural networks and other AI models, which
are normally considered black boxes. Given a target model ℳ, the goal of XAI is to provide a
humanunderstandable textual or visual explanation of ℳ’s predictions.</p>
      <p>Perturbation-based instance-level explanation methods investigate the behavior of the target model’s
predictions on varying inputs to identify a subgraph relevant to the prediction, which then serves as an
explanation.</p>
      <p>
        A perturbation-based instance-level explanation should reflect the model’s prediction, i. e., the
explanation graph should only contain information important for the prediction. Similarly, the result
should change if crucial information is removed from the input. At the same time, an explanation
should be suficiently sparse to be interpretable by a human [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ].
      </p>
      <p>
        The GNNExplainer proposed by Ying et al. [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] is one of the most well-known and initial approaches
in explainability for GNNs.
      </p>
      <p>
        Recently, path-based explanations for KGC have gained attention [
        <xref ref-type="bibr" rid="ref5 ref6">5, 6</xref>
        ]. Instead of subgraphs, they
generate a set of paths connecting the query entities, naturally capturing their connections. Such
explanations are expected to be better interpretable and more user-friendly.
      </p>
      <p>While TKGC and XAI are well-researched subjects, there is still little literature on using XAI in
TKGC.</p>
      <p>
        Some works combine TKGC and TKGC Explainability in a single model, also known as self-interpretable
models. Examples of self-interpretable models are xERTE [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] and T-GAP [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ], which both construct a
subgraph using attention propagation during inference that can also serve as an explanation graph.
However, in this work, we are interested in model-agnostic explainers, i. e., explainers that can be used
for diferent target models ℳ without large modifications.
      </p>
      <p>
        The perturbation-based explainer Temporal Motifs Explainer (TempME) proposed by Chen et al. [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ]
identifies the most important recurring temporal patterns of connections in a TKG.
      </p>
      <p>Visits
2024-12-27</p>
      <p>Charlie</p>
      <p>
        He et al. [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ] extend an existing explainer to the temporal setting. First, the TKG is divided into
several non-temporal KGs, i. e., a sequence of KGs. Second, a non-temporal explainer is then used
to explain the instance on each static snapshot. Finally, a time-aware explanation is constructed by
combining the most dominant static explanations.
      </p>
      <p>The existing TKGC explainers fail to incorporate the temporal aspect suficiently, address the unique
challenges posed by the TKG graph characteristics, or tailor explanations to meet user requirements for
time-based interpretations.</p>
    </sec>
    <sec id="sec-4">
      <title>4. Method</title>
      <p>
        GNN approaches for TKGC learn how information evolves over time to predict new facts. Since the
temporal order of facts conveys information, models process the graph in chronological order rather than
random order to leverage the causal relationships and temporal dependencies [
        <xref ref-type="bibr" rid="ref14 ref15">14, 15</xref>
        ]. In consequence,
this should be captured in explanations, too. E. g., if we want to explain why a person visits a doctor, it
can be interesting to know what happened the days before or after. This can build up a temporal chain
of facts, see Fig. 2.
      </p>
      <p>
        Inspired by how path-based explanations incorporate connections between query entities in the
explanations [
        <xref ref-type="bibr" rid="ref5 ref6">5, 6</xref>
        ], we propose chronological paths to infuse temporal properties into the explanation
of non-temporal explainers. We propose a chronological regulation that favours the temporal chain of
facts. To reinforce the efect, relations not being on such a path might be penalized.
      </p>
      <sec id="sec-4-1">
        <title>4.1. Chronological Path</title>
        <p>A chronological path is a path in a graph with chronologically ascending or descending timestamps
along its edges. For any two consecutive edges on a chronologically ascending path, the timestamp of
the second edge is greater than or equal to the timestamp of the first edge. Chronological descending is
defined analogously. Given a TKG  and the target models predicted entity ′ for query  = (, , ?, )
with entity , relation  and timestamp , we denote (,′) as a chronological path from  to ′.
The set of all chronological paths between  and ′ is defined as</p>
        <p>()
(,′) = {(,′) |  ≤ max},
(3)
where max is the maximal length of the chronological paths.</p>
        <p>We set max = 3 for all experiments as longer paths may be less relevant, and most TKG models
only consider a maximum of 3 hops around a query.</p>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. Chronological Regulation</title>
        <p>Chronological regulation rewards edges on chronological paths between the query’s subject  and the
target model’s predicted object ′ and might penalize edges that are not. We propose two methods to
implement chronological regulation.</p>
        <p>
          ⎧
⎪⎪⎩0,
otherwise
Loss Regulation Given all chronological paths (,′) from  to the target model’s prediction ′,
chronological regulation can be applied to the edge mask  ( ) by defining a regulation loss that
measures the distance between  ( ) and the optimal edge mask regarding the chronological paths
(loss_reg) ∈ [
          <xref ref-type="bibr" rid="ref1 ref21">0, 1</xref>
          ]. For each incoming edge  in  that connects the nodes  and  ∈  with the
relation  at time , we define the optimal edge mask regarding the chronological paths as
log (||min) , if (, , , ) ∈  ∈ (,′),
(loss_reg) = ⎪⎪⎨1 −  · log (max)
where  ∈ [
          <xref ref-type="bibr" rid="ref1 ref21">0, 1</xref>
          ] is a hyperparameter that determines the logarithmic value decrease for edges on
chronological paths that exceed a length of 1 and ||min is the length of the shortest chronological path
between  and ′. If  = 0, (loss_reg) is equal to 1 regardless of the length of the chronological path. If
(loss_reg) is 0 for paths of length . Note that we only reward
 = 1, the decrease is maximum and 
edges in the direct neighborhood of , if they lie on a chronological path. We expect that we can guide
the explanation in the direction of the chronological paths, and not regulate them individually.
Now we can define a loss ℓreg( ( ), (loss_reg)) between  ( ) and (loss_reg) which we can add to the
explainer loss. We choose the mean absolute error.
        </p>
        <p>ℓreg( ( ), (loss_reg)) = mean({1, ...,  }),  =  · |  ( ) − (loss_reg)|
(5)
We use a hyperparameter  , to scale the strength of the regulation.</p>
        <p>Gradient Regulation The second regulation method directly applies the regulation to the gradients.
(grad_reg) that rewards or penalizes the gradients
The chronological paths are used to create a function 
of the mask. This function is similar to the one used in Eq. 4, but with hyperparameter  scaling the
maximum reward and penalty.</p>
        <p>⎧ (︂
⎪
(grad_reg) = ⎨⎪

⎪⎪⎩− ,</p>
        <p>log (||min) )︂
1 −  · log (max)
, if (, , , ) ∈  ∈ (,′),
otherwise
Let Θ be the edge mask parameters and ∇ℓ(Θ) the computed gradients. To regulate the edge mask,
before doing gradient descent, the computed gradient is subtracted by (grad_reg). This regulation
increases the gradient for edges on a chronological path and decreases it otherwise.</p>
        <p>∇ℓ(Θ) = ∇ℓ(Θ) − (grad_reg)
Note that we do not need a scaling parameter  as in the previous method since we can scale (grad_reg)
directly with  .</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Experiments</title>
      <sec id="sec-5-1">
        <title>5.1. Datasets &amp; Target Models</title>
        <p>
          Commonly used real-world benchmark datasets for TKGC are subsets of ICEWS1 and WIKIDATA [
          <xref ref-type="bibr" rid="ref16">16</xref>
          ].
We utilize ICEWS14 and WIKIDATA11K [
          <xref ref-type="bibr" rid="ref17">17</xref>
          ].
        </p>
        <p>
          ICEWS14 contains socio-political events. The entities are, for example, countries, institutions, or
persons; the relations are predicates like Consult or Make statement, and the timestamps are the dates
on which the event occurred [
          <xref ref-type="bibr" rid="ref17">17</xref>
          ].
1https://www.lockheedmartin.com/en-us/capabilities/research-labs/advanced-technology-labs/icews.html (last visit september
30, 2024)
(4)
(6)
(7)
        </p>
        <p>
          WIKIDATA11K contains entities such as historical figures, places, and artifacts, connected by relations
like Was born in or Founded. The characteristics of both datasets can be found in App. A Tab. 2.
In this work, we use two state-of-the-art TKGC models for predictions. TARGCN [
          <xref ref-type="bibr" rid="ref1">1</xref>
          ] aggregates a subset
of the temporal neighborhood with a single GCN layer to compute the time-dependent representation
of an entity. T-GAP [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ] utilizes multiple GNN layers and attention-based subgraph sampling to account
for distant nodes, which increases the representativeness of predictions due to increased information
lfow. A detailed description of both target models can be found in appendix C. We then explain these
target models using the GNNExplainer.
        </p>
      </sec>
      <sec id="sec-5-2">
        <title>5.2. Metrics for Graph Neural Network Explainers</title>
        <p>
          Following [
          <xref ref-type="bibr" rid="ref18">18</xref>
          ], we evaluate our explanations using Fidelity, charact, and Sparsity, as well as with our
proposed SparseFid, which combines fidelity and sparsity.
        </p>
        <p>
          Fidelity [
          <xref ref-type="bibr" rid="ref19">19</xref>
          ] measures the faithfulness of an explanation to the target model. This means the model’s
prediction should change if important entities or relations are removed from the explanation graph (fid +).
However, if unimportant entities or relations are removed, the prediction should remain the same (fid − ).
 + = 1 −
1 ∑︁ 1(^
        </p>
        <p>∖ = ),
|| ∈
 − = 1 −
1 ∑︁ 1(^</p>
        <p>
          = )
|| ∈
If fid − is close to 0, the provided explanation is suficient , and if fid + is close to 1, the explanation is
necessary. An explanation should be suficient and necessary. A metric to combine fid + and fid − is the
charact score [
          <xref ref-type="bibr" rid="ref19">19</xref>
          ].
        </p>
        <p>+ + −
ℎ = ifd++ + 1−  fid− −
, with + + − = 1
We give equal weight to fid + and fid .</p>
        <p>Sparsity is also an important pro p−erty of explanation graphs to provide human-understandable
explanations as TKGs often have a high avg. node degree and high information density due to the
additional temporal information compared to static KGs. We define the sparsity of an explanation
subgraph  as
 =
1 ∑︁ (︂
 =0
1</p>
        <p>(| | + 1) )︂ ,
− (| | + 1)
(8)
(9)
(10)
(11)
where | | denotes the number of edges in the explanation subgraph and | | the number of edges
in the computation graph. Note that we are taking the  of the number of edges because we want to
focus on explanations that are as small as possible. Reducing an already small explanation has a greater
efect on the sparsity than reducing a large explanation.</p>
        <p>Sparse-Fidelity Finally, we propose a new combined metric based on the charact score and sparsity.
As with the charact score, we calculate the harmonic mean between charact and sparsity.</p>
        <p>+ 
 
charact + sparsity
  =
, with  +  = 1
With  =  = 0.5 we give equal weight to charact and sparsity.</p>
      </sec>
      <sec id="sec-5-3">
        <title>5.3. Baseline</title>
        <p>
          We use the GNNExplainer without temporal edge mask regularization as the baseline to compare the
proposed temporal regulation methods. Although the GNNExplainer was originally developed for
static KGs only, it can be adapted to the temporal setting by extending the edge mask to the temporal
adjacency matrix. The use of this inflated mask  ∈ [
          <xref ref-type="bibr" rid="ref1 ref21">0, 1</xref>
          ]|ℰ|×|ℛ|×|ℰ|×| | allows the GNNExplainer to
indirectly model the temporal information with the edge mask since each relation between two entities
can be considered independently at all possible times. This is indirect because the timestamp is not
masked independently of the edge type, and the temporal information might also be utilized in other
model components not afected by the edge mask. A description of how the edge mask can be applied
to the target models TARGCN and T-GAP can be found in App. C.
        </p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>6. Results</title>
      <p>This chapter provides the evaluation results of the GNNExplainer w. and w/o. temporal regulation
on TKGC. We apply the GNNExplainer to the two target models TARGCN and T-GAP. Details about
the hyperparameter-tuning can be found in App. B. Tab. 1 shows an overview of the results using the
edge mask and the proposed edge mask regulation methods loss and gradient regulation. We report all
metrics with a threshold of 100 edges for each mask.2</p>
      <p>We observe that the explanations for the predictions of the target model TARGCN achieve notably
better scores compared to those for T-GAP. Using the target model TARGCN, the proposed regulation
methods can outperform the edge mask, with loss regulation providing the best results. In contrast, the
best results for the target model T-GAP are obtained with gradient regulation, while loss regulation
cannot improve upon the baseline. Since the sparsity of the explanations remains constant at a fixed
threshold for the edge mask, we observe the same values with and without regulation.</p>
      <sec id="sec-6-1">
        <title>Case Study: Impact of Chronological Regulation on Edge Mask Evolution: We investigate the</title>
        <p>evolution of masks on one randomly selected ICEWS14 quadruple to see how the regulation methods
influence edge mask learning.</p>
        <p>The mask history using TARGCN as the target model can be seen in Fig. 3. We highlighted two edges
for better visualization: one on a chronological path (black) and one that is not (red). When using loss
regulation, an initial increase in the mask value can be observed for the edge not on a chronological path,
followed by a steep decrease after about 60 epochs. The explainer seems to have found a minimum for
the loss after a few epochs. An instant decrease of the edge mask for edges not lying on a chronological
path can be observed using the gradient regulation. Since this method does not minimize a loss, the
influence of the regulation is immediately apparent, which generally seems to lead to a more precise
separation of important and unimportant edges. Please note that this behavior does not apply to all
2The source code and target model checkpoints for our experiments are publicly available at
https://anonymous.4open.science/r/ExplainableTKGC-1908/
on chr. path not on chr. path</p>
        <p>Loss-Regulation
10 30 50 70 90</p>
        <p>Threshold
10 30 50 70 90</p>
        <p>Threshold
edges on a non-chronological path. If we look at the same sample with the target model T-GAP, which
can be found in App. D Fig. 5, we see that using the threshold has already removed all edges that the
explainer considers unimportant. It can be seen that, compared to the target model TARGCN, the
loss regulation seems not to influence the edge weights. For the gradient regulation, some edges are
influenced.</p>
        <p>Edge Masks Across Diferent Thresholds: The previous results show explanations with an edge
mask threshold of 100. However, since we are not only interested in the fidelity of the explanation but
also in achieving a high degree of sparsity, we have also evaluated low thresholds for the masks. The
lower part of Tab. 1 reports the results of the GNNExplainer for both target models using the edge mask
with and without the two regulations with the best threshold regarding the SparseFid score. We observe
that the best threshold for all methods is below 100. For the target model TARGCN on the ICEWS14
dataset, the edge mask can achieve the highest SparseFid score. On the WIKIDATA11K dataset, loss
regulation is still the best method. Gradient regulation also remains the best method for the target
model T-GAP.</p>
        <p>In the following, we look at the results for the target models i) TARGCN and ii) T-GAP in detail.
i) TARGCN The comparison of loss and gradient regulation to the baseline on the ICEWS14 dataset
in Fig. 4 shows a similar trend of the scores depending on the threshold. However, the baseline can
provide better results for lower thresholds. This is also indicated by the smaller optimal threshold of
the baseline compared to the regulation methods. Loss and gradient regulation can only improve the
baseline with thresholds of 50 or higher. Gradient regulation, in particular, struggles with high fidelity
for very small thresholds.</p>
        <p>The results on the WIKIDATA11K dataset show a significant improvement of the baseline for small
thresholds using loss and gradient regulation, as can be observed in Fig. 6a in the appendix. This results
in the optimal threshold being improved from 30 to 20 for both regulation methods. The charact score
of the loss regulation is superior to the baseline for every threshold. Thus, the charact score of the
loss regulation at a threshold of 30 is already above the baseline score with the maximum threshold of
100. Gradient regulation, on the other hand, can outperform the baseline for small thresholds. Above a
threshold of 40, the improvements are minimal.
ii) T-GAP Since the results of the GNNExplainer for the target model T-GAP using the loss regulation
for diferent thresholds show no diference to the baseline performance, we only report the results of
the edge and node mask and the gradient regulation.</p>
        <p>We report the results of T-GAP in the appendix in Fig. 6. In comparison to the baseline, the gradient
regulation can only achieve very small improvements for a threshold of 100 on ICEWS14, as can be
seen in Tab. 1. However, if we look at smaller explanations in Fig. 6b, we see an improvement in the
charact score when using the gradient regulation. For a threshold of 30 to 60, a noticeable improvement
can be observed compared to the baseline. The optimal threshold can be decreased to 80 using gradient
regulation.</p>
        <p>A very similar behavior can be found in the results on the WIKIDATA11K dataset in Fig. 6c. The
optimal threshold for the edge mask without regulation is 70 but can be reduced to 50 using gradient
regulation.</p>
      </sec>
    </sec>
    <sec id="sec-7">
      <title>7. Discussion</title>
      <p>Our results show improvements through temporal regulation for all models on all datasets in most
scenarios. Often, we observe improvements through both regulation methods, or at least through one
of them.</p>
      <p>With an edge mask threshold of 100, the GNNExplainer can obtain the best charact score through loss
regulation for TARGCN and gradient regulation for T-GAP. While the GNNExplainer for TARGCN can
achieve an improvement with the gradient regulation compared to the edge mask without regulation,
the loss regulation for T-GAP had no noticeable influence on the quality of the explanations according
to the metrics used.</p>
      <p>The explainer uses a significantly smaller edge mask for TARGCN than for T-GAP, which may be
easier to optimize. This is because message passing is only performed for a sampled temporal 1-hop
neighborhood of the subject node in this model. Since TARGCN limits this neighborhood to a maximum
of 100 edges, the edge mask includes, at most, 100 parameters to optimize. In contrast, with T-GAP,
message passing is performed for each edge in the graph, which means that the number of parameters
in the edge mask is significantly larger than with TARGCN. This might cause the loss regulation to
have only a small impact on explanations of T-GAP’s predictions.</p>
      <p>We can observe that the explanation quality seems to depend highly on the target model to be
explained. Tab. 1 shows that TARGCN explanations are considerably better than explanations for
T-GAP. This might be caused by i) the larger neighborhood context of T-GAP and the resulting complex
inference of T-GAP compared to TARGCN ; ii) a diference in the TKGC prediction quality as TARGCN
performs better than T-GAP on both datasets,3 which makes explanations more dificult.</p>
      <p>Furthermore, the optimal size of the edge mask seems to depend heavily on the underlying dataset
and target model. The explainer consistently achieves a smaller optimal explanation threshold for
TARGCN than for T-GAP. One reason could be that T-GAP considers the 3-hop neighborhood around
the query node for its prediction, while TARGCN only considers the direct neighborhood. Therefore,
T-GAP generally requires more edges to provide a reliable prediction than TARGCN. This is further
3With the original source code, we reproduced the original experiments and achieved TKGC scores close the ones publications
with the models. TARGCN: 0.606 MRR on ICEWS14, 0.715 MRR on WIKIDATA11K; T-GAP: 0.56 MRR on ICEWS14, 0.663
MRR on WIKIDATA11K.
supported by the observation of the charact score curve in relation to the threshold shown in Fig. 4 and
Fig. 6b in the appendix.</p>
      <p>We evaluated the models following common methods and standards in XKGC and introduced a new
metric to better reflect the explanations’ size. A human evaluation to verify a model’s capabilities in
realworld scenarios is not common in XKGC tue to open challenges, especially with TKGs, as i) the standard
KGC benchmark datasets require human experts in the respective domains, ii) no commonly accepted
dataset for X(T)KGC exists, iii) existing state-of-the-art TKGC models require large subgraphs to make
TKGC predictions. Even though our chronological regulation can reduce the size of explanations, a
human evaluation still poses significant challenges and would be an interesting topic for future work.</p>
    </sec>
    <sec id="sec-8">
      <title>8. Conclustion</title>
      <p>In this work, we address the explainability of GNN-based TKGC models. We implement a baseline for
GNN-based TKGC explanations using non-temporal GNN explainers and report the explanation quality
according to established metrics. Furthermore, we proposed a regulation method that incentivizes a
chronological order in the explanations to improve explanations over TKCs. We see this in improved
explainability scores in most scenarios across models and datasets, e. g., with fidelity characterization
scores increased by up to 2% compared to the baselines. We observe that the regulation methods can
reduce the size of the explanation graph while maintaining the same explanation quality according to
explainability metrics in most scenarios.</p>
    </sec>
    <sec id="sec-9">
      <title>Acknowledgments</title>
      <p>This work was supported by the Research Council of Norway through its Centres of Excellence scheme,
Integreat – Norwegian Centre for Knowledge-driven Machine Learning, project number 332645.</p>
    </sec>
    <sec id="sec-10">
      <title>Declaration on Generative AI</title>
      <p>We used ChatGPT and Grammarly to check grammar and spelling and to make minor rephrasings for
improved clarity. All changes were reviewed by us, and we take full responsibility for the content of
this publication.</p>
    </sec>
    <sec id="sec-11">
      <title>A. Dataset Statistics</title>
    </sec>
    <sec id="sec-12">
      <title>B. Hyper-parameter Search</title>
      <p>The proposed chronological regulation methods add new hyperparameters to the GNNExplainer.</p>
      <p>For all other parameters added for the chronological regulation, we use grid-search hyperparameter
tuning with the parameters reported in Tab. 3 on 1000 samples for TARGCN and 500 for T-GAP. Note that
it is also possible to optimize the hyperparameters for each sample individually since the GNNExplainer
has to be trained separately for each sample by default. The best hyperparameters are determined by
the charact score. However, as this can be artificially inflated with a very large explanation, we limit
the explanation size to 100 edges. Except for the number of training epochs (200 for TARGCN and 100
for T-GAP), we do not change any default hyperparameters of the GNNExplainer.</p>
    </sec>
    <sec id="sec-13">
      <title>C. TKGC Models</title>
      <sec id="sec-13-1">
        <title>Time-aware Relational Graph Convolutional Network (TARGCN) [1] is based on a single</title>
        <p>GCN layer to aggregate graph neighborhood information. For every query  = (, , ?, ), the model
samples the temporal neighborhood ¯ (, ) ⊆  (, ) of the query node  at time . Then, a
GCN layer is used to aggregate information of ¯ (, ) to encode the time-aware representation of
entity  at time , by combining time-invariant representations of relation , entity  and implicit
time diference information from the subset of all temporal neighbors.</p>
        <p>h(,) =
1</p>
        <p>∑︁
¯
| (,)| (,)∈¯ (,)</p>
        <p>W(h(,)||h),
where h denotes the time-invariant embedding of relation  and h(,) the time-aware entity embedding
¯
h(,) for (, ) ∈  (,).</p>
        <p>
          For each possible candidate object ′, a simplified time-aware representation is compared to  using
DistMult decoding [
          <xref ref-type="bibr" rid="ref20">20</xref>
          ].
        </p>
        <p>To apply the edge mask to TARGCN, we need to adjust Eq. 12 as follows:
h(,) =
1</p>
        <p>∑︁
¯
| (,)| (,)∈¯ (,)</p>
        <p>W((h(,)||h) ⊙  ( )(,)),
(12)
(13)
where  ( )(,) is the sigmoid applied edge mask parameter for the edge connecting  with  at time
.  masks a feature that is based on the time-aware entity embedding and the time-invariable relation
embedding .</p>
        <p>
          Temporal GNN with Attention Propagation (T-GAP)
[
          <xref ref-type="bibr" rid="ref2">2</xref>
          ], another state-of-the-art TKGC model,
considers distant nodes for encoding through multiple GNN layers. This allows the model to capture a
richer context and potentially increase representativeness due to the increased information flow. T-GAP
iteratively samples a subgraph based on node and edge attention values. Starting from a single node,
each iteration adds nodes and edges based on their attention values to the subgraph. To complete
the query, the node within the subgraph with the highest attention is predicted. T-GAP performs
message-passing initially for each edge of the graph, as well as for all edges of the sampled subgraph in
each iteration. While the weights vary across diferent layers and may also depend on the timestamp,
the following message-passing scheme can always be found:
        </p>
        <p>m = W(h + p +  |Δ|),
embedding.
where h denotes the node features, p the relation embedding, and  |Δ| a temporal displacement
The implementation of the edge mask in T-GAP is similar to TARGCN. The message passing from
Eq. 14 is modified as follows:
m = W
︁(
(h + p +  |Δ|) ⊙  ( ) ,
︁)
(14)
(15)
where the edge mask  is multiplied with each of the messages between node  and node .</p>
      </sec>
    </sec>
    <sec id="sec-14">
      <title>D. Further results</title>
      <p>Fig. 6 shows the performance of the explainer with and without regulation at diferent thresholds.
on chr. path</p>
      <p>not on chr. path
without regulation for a random sample from ICEWS14.</p>
    </sec>
    <sec id="sec-15">
      <title>E. Computing Resource</title>
      <p>We ran the experiments on our GPU cluster with Nvidia A40 GPUs (older GPUs with less VRAM,
e. g., Nvidia Tesla cards, are suficient, too). For both target model training and the hyperparameter
tuning, we used approx. 450h GPU hours. Note that our approach does not substantially increase
the computation time of the existing GNNExplainer. We evaluated our approaches on existing TKGC
datasets for comparability. These datasets were not developed for XAI and, therefore, contain large test
sets that cause the runtime of our experiments. Furthermore, the large computation times are related to
the target model T-GAP and are thus independent of our proposed approach.</p>
      <p>1.0
0.8
k
sa0.6
egd0.4
0.2
0.0
M</p>
      <p>M</p>
      <p>M
E</p>
      <p>E</p>
      <p>E
1.0
0.8
charact
fid +
0.78
charact
fid +
10
30
charact
fid +
SparseFid
charact
30</p>
      <p>50 70
Threshold
90
10
30</p>
      <p>50 70
Threshold</p>
      <p>90
charact
fid +
charact
{0.05, 0.1, 0.2, 0.4, 0.8}
{0.0, 0.33, 0.66, 1.0}
0.8
0.0</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Ding</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Ma</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>He</surname>
          </string-name>
          ,
          <string-name>
            <surname>J. Wu</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          <string-name>
            <surname>Han</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          <string-name>
            <surname>Tresp</surname>
            ,
            <given-names>A Simple</given-names>
          </string-name>
          <string-name>
            <surname>But</surname>
          </string-name>
          <article-title>Powerful Graph Encoder for Temporal Knowledge Graph Completion</article-title>
          ,
          <source>in: Intelligent Systems and Applications</source>
          , Springer Nature Switzerland,
          <year>2024</year>
          , pp.
          <fpage>729</fpage>
          -
          <lpage>747</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>J.</given-names>
            <surname>Jung</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Jung</surname>
          </string-name>
          ,
          <string-name>
            <given-names>U.</given-names>
            <surname>Kang</surname>
          </string-name>
          ,
          <article-title>Learning to Walk across Time for Interpretable Temporal Knowledge Graph Completion</article-title>
          ,
          <source>in: Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery &amp; Data Mining, Association for Computing Machinery</source>
          ,
          <year>2021</year>
          , p.
          <fpage>786</fpage>
          -
          <lpage>795</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>J.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Cao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. C. K.</given-names>
            <surname>Cheung</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W. L.</given-names>
            <surname>Hamilton</surname>
          </string-name>
          ,
          <article-title>TeMP: Temporal message passing for temporal knowledge graph completion</article-title>
          , in: B.
          <string-name>
            <surname>Webber</surname>
            , T. Cohn,
            <given-names>Y.</given-names>
          </string-name>
          <string-name>
            <surname>He</surname>
          </string-name>
          , Y. Liu (Eds.),
          <source>Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP)</source>
          ,
          <article-title>Association for Computational Linguistics</article-title>
          , Online,
          <year>2020</year>
          , pp.
          <fpage>5730</fpage>
          -
          <lpage>5746</lpage>
          . URL: https://aclanthology.org/
          <year>2020</year>
          . emnlp-main.
          <volume>462</volume>
          /. doi:
          <volume>10</volume>
          .18653/v1/
          <year>2020</year>
          .emnlp-main.
          <volume>462</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>H.</given-names>
            <surname>Yuan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Yu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Gui</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Ji</surname>
          </string-name>
          ,
          <article-title>Explainability in Graph Neural Networks: A Taxonomic Survey</article-title>
          ,
          <source>IEEE Transactions on Pattern Analysis and Machine Intelligence</source>
          <volume>45</volume>
          (
          <year>2023</year>
          )
          <fpage>5782</fpage>
          -
          <lpage>5799</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>H.</given-names>
            <surname>Chang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Ye</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Lopez-Avila</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Du</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <article-title>Path-based explanation for knowledge graph completion</article-title>
          ,
          <source>in: Proceedings of the 30th ACM SIGKDD Conference on Knowledge Discovery and Data Mining</source>
          , KDD '24,
          <string-name>
            <surname>Association</surname>
          </string-name>
          for Computing Machinery, New York, NY, USA,
          <year>2024</year>
          , p.
          <fpage>231</fpage>
          -
          <lpage>242</lpage>
          . URL: https://doi.org/10.1145/3637528.3671683. doi:
          <volume>10</volume>
          .1145/3637528.3671683.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>S.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Song</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Adeshina</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Zheng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Faloutsos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Sun</surname>
          </string-name>
          , Page-link:
          <article-title>Path-based graph neural network explanation for heterogeneous link prediction</article-title>
          ,
          <source>in: Proceedings of the ACM Web Conference</source>
          <year>2023</year>
          , WWW '23,
          <string-name>
            <surname>Association</surname>
          </string-name>
          for Computing Machinery, New York, NY, USA,
          <year>2023</year>
          , p.
          <fpage>3784</fpage>
          -
          <lpage>3793</lpage>
          . URL: https://doi.org/10.1145/3543507.3583511. doi:
          <volume>10</volume>
          .1145/3543507.3583511.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>J.</given-names>
            <surname>Gilmer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. S.</given-names>
            <surname>Schoenholz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. F.</given-names>
            <surname>Riley</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Vinyals</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G. E.</given-names>
            <surname>Dahl</surname>
          </string-name>
          , Neural Message Passing for Quantum Chemistry,
          <source>in: Proceedings of the 34th International Conference on Machine Learning, PMLR</source>
          ,
          <year>2017</year>
          , pp.
          <fpage>1263</fpage>
          -
          <lpage>1272</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <surname>M. M. Bronstein</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Bruna</surname>
            , T. Cohen,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Veličković</surname>
          </string-name>
          , Geometric Deep Learning: Grids, Groups, Graphs, Geodesics, and
          <string-name>
            <surname>Gauges</surname>
          </string-name>
          ,
          <year>2021</year>
          . URL: https://arxiv.org/abs/2104.13478. arXiv:
          <volume>2104</volume>
          .
          <fpage>13478</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Ying</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Bourgeois</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>You</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Zitnik</surname>
          </string-name>
          ,
          <string-name>
            <surname>J. Leskovec,</surname>
          </string-name>
          <article-title>GNNExplainer: Generating Explanations for Graph Neural Networks</article-title>
          ,
          <source>in: Advances in Neural Information Processing Systems</source>
          , volume
          <volume>32</volume>
          ,
          <string-name>
            <surname>Curran</surname>
            <given-names>Associates</given-names>
          </string-name>
          , Inc.,
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>J.</given-names>
            <surname>Kakkad</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Jannu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Sharma</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Aggarwal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Medya</surname>
          </string-name>
          ,
          <source>A Survey on Explainability of Graph Neural Networks</source>
          ,
          <year>2023</year>
          . URL: https://arxiv.org/abs/2306.
          <year>01958</year>
          . arXiv:
          <fpage>2306</fpage>
          .
          <year>01958</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Han</surname>
          </string-name>
          ,
          <string-name>
            <surname>P</surname>
          </string-name>
          . Chen,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Ma</surname>
          </string-name>
          , V. Tresp,
          <article-title>Explainable Subgraph Reasoning for Forecasting on Temporal Knowledge Graphs</article-title>
          , in: International Conference on Learning Representations,
          <year>2021</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>J.</given-names>
            <surname>Chen</surname>
          </string-name>
          , R. Ying,
          <article-title>TempME: Towards the Explainability of Temporal Graph Neural Networks via Motif Discovery</article-title>
          ,
          <source>in: Advances in Neural Information Processing Systems</source>
          , volume
          <volume>36</volume>
          ,
          <string-name>
            <surname>Curran</surname>
            <given-names>Associates</given-names>
          </string-name>
          , Inc.,
          <year>2023</year>
          , pp.
          <fpage>29005</fpage>
          -
          <lpage>29028</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>W.</given-names>
            <surname>He</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. N.</given-names>
            <surname>Vu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Jiang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. T.</given-names>
            <surname>Thai</surname>
          </string-name>
          ,
          <article-title>An Explainer for Temporal Graph Neural Networks</article-title>
          ,
          <source>in: GLOBECOM 2022 - 2022 IEEE Global Communications Conference</source>
          ,
          <year>2022</year>
          , pp.
          <fpage>6384</fpage>
          -
          <lpage>6389</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>W.</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Peng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Jia</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Peng</surname>
          </string-name>
          ,
          <article-title>Pre-trained language model with prompts for temporal knowledge graph completion</article-title>
          , in: A.
          <string-name>
            <surname>Rogers</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Boyd-Graber</surname>
          </string-name>
          , N. Okazaki (Eds.),
          <source>Findings of the Association for Computational Linguistics: ACL</source>
          <year>2023</year>
          ,
          <article-title>Association for Computational Linguistics</article-title>
          , Toronto, Canada,
          <year>2023</year>
          , pp.
          <fpage>7790</fpage>
          -
          <lpage>7803</lpage>
          . URL: https://aclanthology.org/
          <year>2023</year>
          .findings-acl.
          <volume>493</volume>
          /. doi:
          <volume>10</volume>
          . 18653/v1/
          <year>2023</year>
          .findings-acl.
          <volume>493</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>M.</given-names>
            <surname>Peng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Jiang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Zhu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Peng</surname>
          </string-name>
          ,
          <article-title>Deja vu: Contrastive historical modeling with prefix-tuning for temporal knowledge graph reasoning</article-title>
          , in: K. Duh,
          <string-name>
            <given-names>H.</given-names>
            <surname>Gomez</surname>
          </string-name>
          , S. Bethard (Eds.),
          <source>Findings of the Association for Computational Linguistics: NAACL</source>
          <year>2024</year>
          ,
          <article-title>Association for Computational Linguistics</article-title>
          , Mexico City, Mexico,
          <year>2024</year>
          , pp.
          <fpage>1178</fpage>
          -
          <lpage>1191</lpage>
          . URL: https://aclanthology. org/
          <year>2024</year>
          .findings-naacl.
          <volume>75</volume>
          /. doi:
          <volume>10</volume>
          .18653/v1/
          <year>2024</year>
          .findings-naacl.
          <volume>75</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>D.</given-names>
            <surname>Vrandečić</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Krötzsch</surname>
          </string-name>
          ,
          <article-title>Wikidata: a free collaborative knowledgebase</article-title>
          ,
          <source>Commun. ACM</source>
          <volume>57</volume>
          (
          <year>2014</year>
          )
          <fpage>78</fpage>
          -
          <lpage>85</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>A.</given-names>
            <surname>García-Durán</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Dumančić</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Niepert</surname>
          </string-name>
          ,
          <article-title>Learning Sequence Encoders for Temporal Knowledge Graph Completion</article-title>
          ,
          <source>in: Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, Association for Computational Linguistics</source>
          ,
          <year>2018</year>
          , pp.
          <fpage>4816</fpage>
          -
          <lpage>4821</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>K.</given-names>
            <surname>Amara</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Ying</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , Z. Han,
          <string-name>
            <surname>Y</surname>
          </string-name>
          . Shan,
          <string-name>
            <given-names>U.</given-names>
            <surname>Brandes</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Schemm</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , Graphframex:
          <article-title>Towards systematic evaluation of explainability methods for graph neural networks</article-title>
          ,
          <year>2024</year>
          . URL: https://arxiv.org/abs/2206.09677. arXiv:
          <volume>2206</volume>
          .
          <fpage>09677</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>K.</given-names>
            <surname>Amara</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Ying</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , Z. Han,
          <string-name>
            <surname>Y</surname>
          </string-name>
          . Zhao,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Shan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>U.</given-names>
            <surname>Brandes</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Schemm</surname>
          </string-name>
          ,
          <string-name>
            <surname>C. Zhang,</surname>
          </string-name>
          <article-title>GraphFramEx: Towards Systematic Evaluation of Explainability Methods for Graph Neural Networks</article-title>
          ,
          <source>in: The First Learning on Graphs Conference</source>
          ,
          <year>2022</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>B.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Yih</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>He</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Gao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Deng</surname>
          </string-name>
          , Embedding Entities and
          <article-title>Relations for Learning and Inference in Knowledge Bases</article-title>
          ,
          <source>in: Proceedings of the 3rd International Conference on Learning Representations</source>
          ,
          <year>2015</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          <source>e ro0.89 c S 0.83 1.00 0.94 e ro0.89 c S 0.83 1.00 0.94 e ro0.89 c S 0.83 0</source>
          .
          <fpage>78</fpage>
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>