<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Elastic Weight Consolidation for Knowledge Graph Continual Learning: An Empirical Evaluation</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Gaganpreet Jhajj</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Fuhua Lin</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>School of Computing and Information Systems, Athabasca University</institution>
          ,
          <country country="CA">Canada</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2026</year>
      </pub-date>
      <abstract>
        <p>Knowledge graphs (KGs) require continual updates as new information emerges, but neural embedding models sufer from catastrophic forgetting when learning new tasks sequentially. We evaluate Elastic Weight Consolidation (EWC), a regularization-based continual learning method, on KG link prediction using TransE embeddings on FB15k-237. Across multiple experiments with five random seeds, we find that EWC reduces catastrophic forgetting from 12.62% to 6.85%, a 45.7% reduction compared to naive sequential training. We observe that the task partitioning strategy afects the magnitude of forgetting: relation-based partitioning (grouping triples by relation type) exhibits 9.8 percentage points higher forgetting than randomly partitioned tasks (12.62% vs 2.81%), suggesting that task construction influences evaluation outcomes. While focused on a single embedding model and dataset, our results demonstrate that EWC efectively mitigates catastrophic forgetting in KG continual learning and highlight the importance of evaluation protocol design.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Continual Learning</kwd>
        <kwd>Knowledge Graphs</kwd>
        <kwd>Elastic Weight Consolidation</kwd>
        <kwd>Link Prediction</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Knowledge graphs (KGs) depict structured information as networks of entities and their relations [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ],
facilitating a variety of applications, including question answering [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] and recommendation systems [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]
and educational systems [
        <xref ref-type="bibr" rid="ref4 ref5 ref6 ref7 ref8 ref9">4, 5, 6, 7, 8, 9</xref>
        ]. These real-world KGs evolve continuously as new information
becomes available and existing knowledge is refined. Neural embedding models, such as TransE
[
        <xref ref-type="bibr" rid="ref10">10</xref>
        ], generate vector representations of entities and relations for link prediction. However, adapting
these models to accommodate new information while retaining prior knowledge presents a significant
challenge.
      </p>
      <p>
        Catastrophic forgetting occurs when neural networks, trained sequentially on multiple tasks,
experience significant performance degradation on earlier tasks after learning new ones [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ]. This
phenomenon poses particular challenges for KG embeddings, where maintaining consistent
representations across evolving information is essential. While continual learning methods have been developed
for image classification and natural language processing, their efectiveness on KG link prediction
remains underexplored.
      </p>
      <p>
        We investigate how Elastic Weight Consolidation (EWC) [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ], a regularization-based continual
learning method, performs on KG link prediction. EWC protects important parameters learned in
previous tasks by adding a quadratic penalty based on the Fisher Information Matrix, allowing networks
to learn new tasks while preserving performance on old ones. We evaluate EWC on TransE embeddings
using FB15k-237, a standard KG benchmark, and compare relation-based task partitioning (where triples
are grouped by relation type) to random partitioning.
      </p>
      <p>Our experiments reveal that EWC substantially reduces catastrophic forgetting. On relation-based
partitioned tasks, naive sequential training results in 12.62% forgetting (measured as MRR degradation
from post-task performance), while EWC with regularization strength  = 10 reduces this to 6.85%, a
45.7% reduction. This demonstrates that regularization-based continual learning efectively preserves
KG embeddings across sequential tasks.</p>
      <p>We also observe that the task partitioning strategy significantly afects the measured forgetting.
Naive sequential training on relation-based tasks exhibits 12.62% forgetting, compared to only 2.81% on
randomly partitioned tasks, a 9.8%point diference. This suggests that evaluation protocols, particularly
how tasks are constructed from datasets, influence continual learning metrics and should be carefully
explored when it comes time to design the study.</p>
      <p>Our study focuses on TransE embeddings for the FB15k-237 dataset across four relation-based tasks.
While this scope limits generalizability, it provides rigorous evidence that EWC reduces catastrophic
forgetting in KG continual learning and raises essential questions about task construction in continual
learning evaluation.</p>
      <p>This work addresses a critical challenge for knowledge graph-based AI agents: maintaining knowledge
representations as new information arrives. As agents increasingly rely on KG-based memory and
reasoning, the ability to incorporate new knowledge while preserving existing facts becomes essential
for long-term autonomous operation.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related Work</title>
      <p>
        KG Embeddings. TransE [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] represents relations as translations in embedding space, learning vectors
such that h + r ≈ t for true triples (ℎ, , ). Extensions include TransH [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ], RotatE [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ], and ComplEx
[
        <xref ref-type="bibr" rid="ref15">15</xref>
        ].
      </p>
      <p>
        Continual Learning. Methods for mitigating catastrophic forgetting include regularization
approaches like EWC [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ] and Learning without Forgetting [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ], replay-based methods that store and
revisit previous examples [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ], and architectural approaches that allocate separate parameters for
diferent tasks [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ]. EWC estimates parameter importance using the Fisher Information Matrix and
adds regularization penalties to protect important weights during subsequent task training.
      </p>
      <p>
        KG Continual Learning. Prior work has explored continual learning for KGs across various
contexts. Wang et al. [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ] studied multi-task feature learning for KG-enhanced recommendation.
Daruna et al. [
        <xref ref-type="bibr" rid="ref20">20</xref>
        ] demonstrated continual learning methods for KG embeddings in robotic manipulation
tasks, evaluating multiple architectures including TransE, DistMult, and ComplEx. Zhao et al. [
        <xref ref-type="bibr" rid="ref21">21</xref>
        ]
introduced the PS-CKGE benchmark, focusing on pattern shifts and demonstrating that these shifts
exacerbate catastrophic forgetting beyond what would be expected from simple data scaling. Recent
work has also examined embedding adaptation [
        <xref ref-type="bibr" rid="ref22">22</xref>
        ] and temporal KGs. While these comprehensive
benchmarking eforts establish the landscape of KG continual learning, a focused empirical analysis of
classic regularization methods like EWC, with explicit investigation of task partitioning efects, remains
limited, motivating our study.
is
(1)
(2)
      </p>
    </sec>
    <sec id="sec-3">
      <title>3. Methodology</title>
      <sec id="sec-3-1">
        <title>3.1. Problem Formulation</title>
        <p>A KG is  = (ℰ , ℛ,  ) where ℰ is the set of entities, ℛ is the set of relations, and  ⊆ ℰ × ℛ × ℰ
the set of true triples. TransE learns embeddings h, r, t ∈ R by minimizing:
ℒ =
∑︁</p>
        <p>∑︁
(ℎ,,)∈ (ℎ′,,′)∈ ′
max(0,  + (h + r, t) − (h
′ + r, t′))
where  ′ contains negative samples, (·, ·) is L2 distance, and  is the margin.</p>
        <p>In continual learning, we partition  into tasks 1, . . . ,  and train sequentially. After training on
task , we measure performance  on task  ≤ . Forgetting for task  after learning task  is:
 =  −   for  &gt; 
(3)
(4)
(5)</p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Elastic Weight Consolidation</title>
        <p>EWC protects important parameters by adding a regularization term to the loss when training on task :
where  *,1 are optimal parameters after task −1 ,  is the Fisher Information diagonal approximation,
and  contr− ols regularization strength. The Fisher Information Matrix diagonal is:</p>
        <p>ℒEWC = ℒ + 2 ∑︁ (  −  *,−1 )2
 = E(ℎ,,)∼ −1
[︃︂(  log (|; ) )︂ 2]︃
We report average forgetting at the end of training:</p>
        <p>We compute this empirically using all triples from the previous task, processed in mini-batches (see
Appendix A for implementation details).</p>
      </sec>
      <sec id="sec-3-3">
        <title>3.3. Task Partitioning</title>
        <p>We evaluate two task partitioning strategies to assess how task construction afects catastrophic
forgetting:</p>
        <p>Relation-based partitioning: We partition FB15k-237 by grouping all triples that share the same
relations. To create balanced task sizes, we sort the 237 relations by frequency (number of triples per
relation) and assign them to four tasks in round-robin order: the most frequent relation goes to Task
1, the second-most to Task 2, the third-most to Task 3, the fourth-most to Task 4, the fifth-most back
to Task 1, and so forth. This ensures each task receives approximately 59 relations with a mixture of
common and rare relation types, while maintaining relation-level coherence—all triples with the same
relation appear in the same task.</p>
        <p>Random partitioning: We randomly shufle all 272,115 training triples and divide them into four
equal chunks of approximately 68,000 triples each. This distributes relation types across all tasks,
creating relation-level overlap—most relations appear in multiple tasks. The key diference:
relationbased partitioning creates distinct distribution shifts between tasks (each task focuses on diferent
relation types). In contrast, random partitioning minimizes distribution shift (each task is a representative
sample of all relations). This allows us to isolate the efect of task boundary definition on the dificulty
of continual learning.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Experimental Setup</title>
      <p>
        Dataset and Partitioning. We use FB15k-237 [
        <xref ref-type="bibr" rid="ref23">23</xref>
        ], which contains 14,505 entities, 237 relations, and
272,115 triples. Using relation-based partitioning (round-robin assignment by relation frequency), we
create four tasks with approximately balanced sizes, each containing 59 relations and their associated
triples.
      </p>
      <p>
        Model Configuration. We use TransE with 50-dimensional embeddings, margin  = 1.0 , and
L2 distance. Training uses the Adam optimizer [
        <xref ref-type="bibr" rid="ref24">24</xref>
        ] with a learning rate of 0.001, a batch size of 256,
and 20 epochs per task. We found 20 epochs suficient for convergence on FB15k-237 in preliminary
experiments.
      </p>
      <p>Methods Evaluated. We compare naive sequential training (no continual learning) with EWC at
multiple regularization strengths ( ∈ {0.1, 1.0, 10.0} ), EWC combined with experience replay (500
examples per task), and replay-only baselines (random and wave-based sampling).</p>
      <p>Evaluation Protocol. We use Mean Reciprocal Rank (MRR) for link prediction, computing filtered
rankings that exclude known actual triples. For each task , we record MRR immediately after training
( ) and after each subsequent task ( for  &gt; ). We run five random seeds (42, 123, 456, 789, 2024)
and report means and standard deviations.</p>
      <p>Hardware. Experiments ran on NVIDIA RTX 3070 Ti (8GB) with 20 hours total computation for 80
experiments (8 methods × 5 seeds × 2 partitioning strategies).</p>
    </sec>
    <sec id="sec-5">
      <title>5. Results</title>
      <sec id="sec-5-1">
        <title>5.1. Classical Methods on Relation-Based Partitioning</title>
        <p>EWC with  = 10 achieves the best performance, reducing forgetting to 6.85% (std 0.33%), a 45.7%
reduction compared to naive training. This demonstrates that regularization-based protection of
important parameters efectively mitigates catastrophic forgetting in KG continual learning. Final MRR
also improves from 0.206 to 0.242, indicating that EWC preserves not only previous task performance
but also maintains overall embedding quality.</p>
        <p>Interestingly, replay-based methods underperform. Random replay achieves 13.78% forgetting, worse
than naive training, suggesting that simply revisiting old examples without principled parameter
protection may interfere with learning. Wave-based replay performs similarly (12.54% forgetting).
Combining EWC with wave replay (9.91% forgetting) improves over replay alone but underperforms
pure EWC, indicating that regularization is the primary driver of performance.</p>
        <p>Figure 1 visualizes these results, showing a clear separation between EWC and other methods.</p>
      </sec>
      <sec id="sec-5-2">
        <title>5.2. Efect of Task Partitioning</title>
        <p>We compare relation-based and random partitioning strategies to assess whether task construction
afects measured forgetting. Table 2 and Figure 2 show results.</p>
        <sec id="sec-5-2-1">
          <title>Partitioning</title>
          <p>Relation-based
Random</p>
        </sec>
        <sec id="sec-5-2-2">
          <title>Diference</title>
          <p>Naive Forgetting (%) EWC Forgetting (%)
12.62 ± 0.35
2.81 ± 0.34
9.81 pp
6.85 ± 0.33
5.08 ± 0.22
1.77 pp</p>
          <p>Relation-based partitioning results in substantially higher forgetting during naive training (12.62% vs
2.81%), a 9.8 percentage-point diference. We hypothesize this occurs because relation-based partitioning
creates task coherence: each task focuses on a distinct subset of relation types (approximately 59 relations</p>
          <p>Catastrophic Forgetting on Relation-Based Task Split
12.6%
10.4%
7.5%
6.8%
13.8%
per task), inducing larger distribution shifts when transitioning between tasks. When the model moves
from Task 1’s relations to Task 2’s completely diferent relation set, the parameter updates required are
more disruptive to previously learned representations. In contrast, random partitioning creates
relationlevel overlap: each task contains a representative sample of most relation types, so the distribution shift
between tasks is minimal. This naturally regularizes learning, as the model continually encounters all
relation types across tasks. This observation suggests that task construction significantly influences
the dificulty of continual learning and that evaluation protocols should explicitly consider and report
partitioning strategies.</p>
          <p>Notably, EWC reduces this gap: the diference between relation-based and random partitioning
under EWC is only 1.77 percentage points (6.85% vs 5.08%), suggesting that efective continual learning
methods can generalize across diferent task construction approaches.</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>6. Discussion</title>
      <p>KG embeddings have structured parameter spaces where specific dimensions encode semantic properties.
The Fisher Information Matrix identifies parameters critical for encoding relation types and entity
characteristics learned in previous tasks. By protecting these parameters, EWC enables new-task
learning while preserving the semantic structure of the embedding space.</p>
      <p>The superior performance of EWC compared to replay methods suggests that principled parameter
protection is more efective than simply revisiting old examples when working memory (replay bufer)
is limited. This aligns with neuroscience findings that synaptic consolidation, rather than replay alone,
enables long-term memory retention.</p>
      <p>
        Our EWC forgetting rate (6.85%) on relation-based partitioned tasks demonstrates efective continual
learning for KG link prediction. While direct comparison with prior work is challenging due to
diferences in datasets and evaluation protocols, our results are consistent with EWC’s performance on
image classification tasks [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ] and suggest that regularization-based continual learning generalizes to
structured knowledge representations.
      </p>
      <p>The task partitioning efect we observe (a 9.8 percentage-point diference) highlights an important
consideration for continual learning evaluation: reported forgetting rates depend on how tasks are
constructed. This suggests that future work should explicitly consider task construction methodology
when designing experiments and reporting results.</p>
    </sec>
    <sec id="sec-7">
      <title>7. Limitations and Future Work</title>
      <p>Our study has several limitations that define the scope and future directions.</p>
      <p>We evaluate only TransE on FB15k-237 across four tasks, a deliberate choice given consumer-grade
GPU constraints (NVIDIA RTX 3070 Ti, 8GB) that enabled rigorous multi-seed experimentation within
feasible time scales. Results may not generalize to: (1) complex embedding methods (RotatE, ComplEx,
TuckER); (2) other datasets (WN18RR, YAGO, Wikidata); (3) more extended task sequences (10+ tasks).
We compare EWC with basic replay baselines, but do not exhaustively benchmark all continual learning
methods. Our relation-based partitioning uses round-robin frequency assignment; alternative strategies
(entity-based, domain-based) warrant investigation.</p>
      <p>Experiments were constrained to academic resources without HPC clusters or large-scale GPU
infrastructure. While this motivated our focused design, a comprehensive multi-model evaluation
would require substantially more resources. We view this as proof-of-concept evidence for future
large-scale studies.</p>
      <p>Future work should evaluate EWC across multiple embedding methods and datasets to assess
generalizability. Scaling studies with 10+ tasks would reveal the long-term dynamics of continual learning.
Systematic investigation of task construction strategies could formalize the relationship between task
partitioning and forgetting. Additionally, combining EWC with more sophisticated replay strategies or
architectural approaches may yield further improvements.</p>
      <p>
        A particularly promising direction is to evaluate continual learning methods on educational KGs,
where knowledge evolves as curriculum content updates and student learning data accumulates. Recent
work has highlighted the potential of open-source LLMs to transform educational contexts through
adaptive and personalized learning [
        <xref ref-type="bibr" rid="ref25">25</xref>
        ]. However, the challenge of maintaining such systems as they
continually learn from new educational content remains underexplored. Building on recent work
in curriculum modeling and adaptive learning [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ], we plan to investigate how EWC performs when
domain-specific educational KGs are continually updated with new learning resources, pedagogical
relations, and student interaction patterns. The integration of open-source LLMs (e.g., Llama [
        <xref ref-type="bibr" rid="ref26">26</xref>
        ],
Mistral [
        <xref ref-type="bibr" rid="ref27">27</xref>
        ], Qwen [
        <xref ref-type="bibr" rid="ref28">28</xref>
        ]) with educational KGs for personalized learning support presents unique
continual learning challenges: both the symbolic KG structure and LLM parameters must adapt to new
content while preserving existing pedagogical knowledge. Our findings on task partitioning efects may
inform how educational content updates should be structured to minimize interference with previously
learned material.
      </p>
      <p>
        Neuromorphic approaches using spiking neural networks with spike-timing-dependent plasticity
ofer promising directions for extending this work. Building on prior work demonstrating SNN-based
relational inference and knowledge representation for knowledge graphs [
        <xref ref-type="bibr" rid="ref29">29</xref>
        ], we plan to
investigate whether the biological learning mechanisms inherent in STDP can provide natural solutions to
catastrophic forgetting. The structured, relational nature of KGs may align particularly well with
neuromorphic computation, where synaptic plasticity mechanisms could enable task-consolidation
without explicit regularization.
      </p>
      <p>While our preliminary experiments on consumer-grade GPUs were inconclusive, a comprehensive
evaluation using specialized neuromorphic hardware (Intel Loihi, IBM TrueNorth, SpiNNaker) could
reveal whether biological learning mechanisms provide advantages for continual learning in knowledge
graphs. The structured, relational nature of KGs may align well with neuromorphic computation,
particularly for consolidating knowledge across tasks. Given EWC’s success through
parameterwise importance weighting, bio-inspired plasticity mechanisms that selectively strengthen or weaken
synaptic connections based on usage patterns warrant thorough exploration with appropriate hardware
infrastructure.</p>
    </sec>
    <sec id="sec-8">
      <title>8. Conclusion</title>
      <p>We evaluated EWC for KG continual learning using TransE embeddings on FB15k-237. Across multiple
experiments with five random seeds, we found that EWC reduces catastrophic forgetting from 12.62% to
6.85%, a 45.7% reduction compared to naive sequential training. This demonstrates that
regularizationbased continual learning efectively preserves KG embeddings across sequential tasks.</p>
      <p>We also observed that the task partitioning strategy significantly afects measured forgetting:
relationbased partitioning exhibits 9.8 percentage points higher forgetting than randomly partitioned tasks for
naive training. This suggests that the evaluation protocol’s design, particularly the task-construction
methodology, influences continual learning measurements and should be carefully considered in
experimental design.</p>
      <p>While our study focuses on a single embedding model and dataset, it provides rigorous evidence
for EWC’s efectiveness and raises essential questions about evaluation methodology in KG continual
learning.</p>
    </sec>
    <sec id="sec-9">
      <title>Acknowledgments</title>
      <p>We acknowledge the support of the Natural Sciences and Engineering Research Council of Canada
(NSERC), Alberta Innovates, Alberta Advanced Education, and Athabasca University, Canada. We would
also like to thank the reviewers for their suggestions on how to improve this work.</p>
    </sec>
    <sec id="sec-10">
      <title>Declaration on Generative AI</title>
      <p>During the preparation of this work, the author(s) used Grammarly and Claude (Anthropic) for Grammar
and spelling checks.</p>
    </sec>
    <sec id="sec-11">
      <title>A. EWC Implementation Details</title>
      <p>We compute the Fisher Information Matrix diagonal after training each task using all triples from that
task, processed in mini-batches of 256. For each batch , we compute the empirical Fisher approximation:
squared gradients across all training examples from the previous task.</p>
      <p>where  is the total number of batches in the task and ℒ is the loss on batch . This accumulates
During training on task , we apply the EWC penalty for all previous tasks:
 ≈  =1</p>
    </sec>
    <sec id="sec-12">
      <title>B. Random Partitioning Results</title>
      <p>vs 2.81% naive). We hypothesize that random partitioning naturally distributes relation types
across tasks, reducing interference. Strong regularization may over-constrain parameters, preventing
necessary adaptation. This suggests that optimal regularization strength depends on task construction.</p>
    </sec>
    <sec id="sec-13">
      <title>C. Hyperparameter Sensitivity</title>
      <p>EWC performance across tested regularization strengths.</p>
      <p>For relation-based partitioning,  = 10 achieves the lowest forgetting. For random partitioning,
weaker regularization ( = 0.1 ) performs best. This suggests that optimal regularization strength
depends on task construction: relation-grouped tasks require stronger protection of essential parameters,
while randomly distributed tasks benefit from more flexibility.</p>
    </sec>
    <sec id="sec-14">
      <title>D. Performance-Forgetting Trade-of</title>
      <p>EWC with  = 10 occupies the optimal region (low forgetting, competitive performance). Replay
methods cluster in the high-forgetting, low-performance region. This visualization confirms that EWC
achieves superior trade-ofs compared to alternative approaches.</p>
    </sec>
    <sec id="sec-15">
      <title>E. Implementation Details</title>
      <p>Code Structure. Experiments implemented in PyTorch 1.13.</p>
      <p>Negative Sampling. We use uniform random tail corruption with a 1:1 ratio of positive to negative
samples. For each positive triple (ℎ, , ), we generate one negative sample (ℎ, , ′) by randomly
replacing the tail entity ′ sampled uniformly from ℰ . We use unfiltered negative sampling during
training (negative samples may coincidentally be true triples), but employ filtered evaluation for link
prediction metrics.</p>
      <p>Hardware. Experiments ran on NVIDIA RTX 3070 Ti (8GB) with approximately 18-20 hours total
computation for all experimental runs (8 methods × 5 seeds × 2 partitioning strategies).</p>
      <p>Task Retention Matrix: EWC (λ=10)</p>
      <p>Task 2 Task 3</p>
      <p>Evaluation Task
Task 4
0.40
0.00
0.20 RR</p>
      <p>M</p>
      <p>The heatmap shows that EWC maintains relatively stable performance across tasks, with limited
degradation on earlier tasks as new tasks are learned. This visualization confirms that EWC efectively
protects previous task performance.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>A.</given-names>
            <surname>Sheth</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Padhee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Gyrard</surname>
          </string-name>
          ,
          <article-title>Knowledge graphs and knowledge networks: the story in brief</article-title>
          ,
          <source>IEEE Internet Computing</source>
          <volume>23</volume>
          (
          <year>2019</year>
          )
          <fpage>67</fpage>
          -
          <lpage>75</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>G.</given-names>
            <surname>Jhajj</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Nomura</surname>
          </string-name>
          ,
          <article-title>Jack and the beansTALK: Towards question answering in plant biology</article-title>
          , in: Eighth Widening NLP
          <string-name>
            <surname>Workshop (WiNLP 2024) Phase</surname>
            <given-names>II</given-names>
          </string-name>
          ,
          <year>2024</year>
          . URL: https://openreview.net/forum? id=0DlJEPHHKe.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>Q.</given-names>
            <surname>Guo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Zhuang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Qin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Zhu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Xie</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Xiong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>He</surname>
          </string-name>
          ,
          <article-title>A survey on knowledge graph-based recommender systems</article-title>
          ,
          <year>2020</year>
          . URL: https://arxiv.org/abs/
          <year>2003</year>
          .00911. arXiv:
          <year>2003</year>
          .00911.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>G.</given-names>
            <surname>Jhajj</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. R.</given-names>
            <surname>Gustafson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Lin</surname>
          </string-name>
          ,
          <string-name>
            <surname>M. P.-C. Lin</surname>
          </string-name>
          ,
          <article-title>Educational knowledge graph creation and augmentation via LLMs</article-title>
          , in: International
          <source>Conference on Intelligent Tutoring Systems</source>
          , Springer,
          <year>2024</year>
          , pp.
          <fpage>292</fpage>
          -
          <lpage>304</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>M. R.</given-names>
            <surname>Kabir</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Lin</surname>
          </string-name>
          ,
          <article-title>An LLM-Powered Adaptive Practicing System</article-title>
          .,
          <source>in: LLM@ AIED</source>
          ,
          <year>2023</year>
          , pp.
          <fpage>43</fpage>
          -
          <lpage>52</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>F.</given-names>
            <surname>Lin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Morland</surname>
          </string-name>
          ,
          <source>Curriculum Modeling for Adaptive Learning</source>
          , Springer Nature Switzerland,
          <year>2025</year>
          , p.
          <fpage>57</fpage>
          -
          <lpage>72</lpage>
          . URL: http://dx.doi.org/10.1007/978-3-
          <fpage>031</fpage>
          -92970-
          <issue>0</issue>
          _5. doi:
          <volume>10</volume>
          .1007/ 978-3-
          <fpage>031</fpage>
          -92970-
          <issue>0</issue>
          _
          <fpage>5</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>G.</given-names>
            <surname>Jhajj</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Lin</surname>
          </string-name>
          ,
          <article-title>Augmenting japanese language acquisition via llms and asr</article-title>
          ,
          <source>in: IEEE Smart World Congress 2025 (IEEE SWC'25)</source>
          , Calgary, Canada,
          <year>2025</year>
          , p.
          <fpage>3</fpage>
          .
          <fpage>84</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>J. R. D.</given-names>
            <surname>Gustafson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , G. Jhajj,
          <string-name>
            <given-names>F.</given-names>
            <surname>Lin</surname>
          </string-name>
          ,
          <article-title>Representing and tracing students' cognitive processes in Project-Based learning through the Function-Behavior-Structure framework and knowledge graphs</article-title>
          ,
          <source>in: IEEE Smart World Congress 2025 (IEEE SWC'25)</source>
          , Calgary, Canada,
          <year>2025</year>
          , p.
          <fpage>6</fpage>
          .
          <fpage>56</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>J. R. D.</given-names>
            <surname>Gustafson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Jhajj</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F. O.</given-names>
            <surname>Lin</surname>
          </string-name>
          ,
          <article-title>Enhancing Project-Based Learning With a GenAI Tool Based on Retrieval: Augmented Generation and Knowledge Graphs</article-title>
          ,
          <source>IGI Global</source>
          ,
          <year>2024</year>
          , p.
          <fpage>161</fpage>
          -
          <lpage>194</lpage>
          . URL: http://dx.doi.org/10.4018/979-8-
          <fpage>3693</fpage>
          -5443-8.ch006.
          <source>doi:10.4018/ 979-8-3693-5443-8</source>
          .
          <year>ch006</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>A.</given-names>
            <surname>Bordes</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Usunier</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Garcia-Duran</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Weston</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Yakhnenko</surname>
          </string-name>
          ,
          <article-title>Translating embeddings for modeling multi-relational data</article-title>
          ,
          <source>Advances in neural information processing systems</source>
          <volume>26</volume>
          (
          <year>2013</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <surname>M. McCloskey</surname>
            ,
            <given-names>N. J.</given-names>
          </string-name>
          <string-name>
            <surname>Cohen</surname>
          </string-name>
          ,
          <article-title>Catastrophic Interference in Connectionist Networks: The Sequential Learning Problem, in: Psychology of Learning and Motivation</article-title>
          , volume
          <volume>24</volume>
          ,
          <string-name>
            <surname>Elsevier</surname>
          </string-name>
          ,
          <year>1989</year>
          , pp.
          <fpage>109</fpage>
          -
          <lpage>165</lpage>
          . URL: https://linkinghub.elsevier.com/retrieve/pii/S0079742108605368. doi:
          <volume>10</volume>
          .1016/S0079-
          <volume>7421</volume>
          (
          <issue>08</issue>
          )
          <fpage>60536</fpage>
          -
          <lpage>8</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>J.</given-names>
            <surname>Kirkpatrick</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Pascanu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Rabinowitz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Veness</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Desjardins</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. A.</given-names>
            <surname>Rusu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Milan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Quan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Ramalho</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Grabska-Barwinska</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Hassabis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Clopath</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Kumaran</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Hadsell</surname>
          </string-name>
          ,
          <article-title>Overcoming catastrophic forgetting in neural networks</article-title>
          ,
          <source>Proceedings of the National Academy of Sciences</source>
          <volume>114</volume>
          (
          <year>2017</year>
          )
          <fpage>3521</fpage>
          -
          <lpage>3526</lpage>
          . URL: https://pnas.org/doi/full/10.1073/pnas.1611835114. doi:
          <volume>10</volume>
          .1073/pnas. 1611835114.
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Feng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <article-title>Knowledge Graph Embedding by Translating on Hyperplanes</article-title>
          ,
          <source>Proceedings of the AAAI Conference on Artificial Intelligence</source>
          <volume>28</volume>
          (
          <year>2014</year>
          ). URL: https://ojs.aaai.org/ index.php/AAAI/article/view/8870. doi:
          <volume>10</volume>
          .1609/aaai.v28i1.
          <fpage>8870</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Sun</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.-H.</given-names>
            <surname>Deng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.-Y.</given-names>
            <surname>Nie</surname>
          </string-name>
          ,
          <string-name>
            <surname>J. Tang,</surname>
          </string-name>
          <article-title>RotatE: Knowledge Graph Embedding by Relational Rotation in Complex Space</article-title>
          , in: International Conference on Learning Representations,
          <year>2019</year>
          . URL: https: //openreview.net/forum?id=HkgEQnRqYQ.
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>T.</given-names>
            <surname>Trouillon</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Welbl</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Riedel</surname>
          </string-name>
          , E. Gaussier, G. Bouchard,
          <article-title>Complex Embeddings for Simple Link Prediction</article-title>
          , in: M. F. Balcan,
          <string-name>
            <given-names>K. Q.</given-names>
            <surname>Weinberger</surname>
          </string-name>
          (Eds.),
          <source>Proceedings of The 33rd International Conference on Machine Learning</source>
          , volume
          <volume>48</volume>
          <source>of Proceedings of Machine Learning Research</source>
          , PMLR, New York, New York, USA,
          <year>2016</year>
          , pp.
          <fpage>2071</fpage>
          -
          <lpage>2080</lpage>
          . URL: https://proceedings.mlr.press/v48/trouillon16. html.
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Hoiem</surname>
          </string-name>
          ,
          <article-title>Learning without forgetting</article-title>
          ,
          <source>IEEE transactions on pattern analysis and machine intelligence</source>
          <volume>40</volume>
          (
          <year>2017</year>
          )
          <fpage>2935</fpage>
          -
          <lpage>2947</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>D.</given-names>
            <surname>Rolnick</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Ahuja</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Schwarz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Lillicrap</surname>
          </string-name>
          , G. Wayne,
          <article-title>Experience Replay for Continual Learning</article-title>
          , in: H.
          <string-name>
            <surname>Wallach</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          <string-name>
            <surname>Larochelle</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Beygelzimer</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>d.</surname>
            Alché-Buc,
            <given-names>E.</given-names>
          </string-name>
          <string-name>
            <surname>Fox</surname>
          </string-name>
          , R. Garnett (Eds.),
          <source>Advances in Neural Information Processing Systems</source>
          , volume
          <volume>32</volume>
          ,
          <string-name>
            <surname>Curran</surname>
            <given-names>Associates</given-names>
          </string-name>
          , Inc.,
          <year>2019</year>
          . URL: https://proceedings.neurips.cc/paper_files/paper/2019/file/ fa7cdfad1a5aaf8370ebeda47a1f1c3-Paper.pdf.
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>A. A.</given-names>
            <surname>Rusu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N. C.</given-names>
            <surname>Rabinowitz</surname>
          </string-name>
          , G. Desjardins,
          <string-name>
            <given-names>H.</given-names>
            <surname>Soyer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Kirkpatrick</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Kavukcuoglu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Pascanu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Hadsell</surname>
          </string-name>
          ,
          <source>Progressive Neural Networks</source>
          ,
          <year>2022</year>
          . URL: https://arxiv.org/abs/1606.04671.
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>H.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Zhao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Xie</surname>
          </string-name>
          ,
          <string-name>
            <surname>M.</surname>
          </string-name>
          <article-title>Guo, Multi-task feature learning for knowledge graph enhanced recommendation</article-title>
          ,
          <source>in: The world wide web conference</source>
          ,
          <year>2019</year>
          , pp.
          <fpage>2000</fpage>
          -
          <lpage>2010</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>A.</given-names>
            <surname>Daruna</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Gupta</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Sridharan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Chernova</surname>
          </string-name>
          ,
          <article-title>Continual learning of knowledge graph embeddings</article-title>
          ,
          <source>IEEE Robotics and Automation Letters</source>
          <volume>6</volume>
          (
          <year>2021</year>
          )
          <fpage>1128</fpage>
          -
          <lpage>1135</lpage>
          . doi:
          <volume>10</volume>
          .1109/LRA.
          <year>2021</year>
          .
          <volume>3056071</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>T.</given-names>
            <surname>Zhao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Ru</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Lin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Geng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Zhu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Pan</surname>
          </string-name>
          , J. Liu,
          <article-title>Rethinking continual knowledge graph embedding: Benchmarks and analysis</article-title>
          ,
          <source>in: Proceedings of the 48th International ACM SIGIR Conference on Research and Development in Information Retrieval</source>
          , SIGIR '25,
          <string-name>
            <surname>Association</surname>
          </string-name>
          for Computing Machinery, New York, NY, USA,
          <year>2025</year>
          , p.
          <fpage>138</fpage>
          -
          <lpage>147</lpage>
          . URL: https://doi.org/10.1145/ 3726302.3730073. doi:
          <volume>10</volume>
          .1145/3726302.3730073.
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>M.</given-names>
            <surname>Delange</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Aljundi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Masana</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Parisot</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Jia</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Leonardis</surname>
          </string-name>
          , G. Slabaugh,
          <string-name>
            <given-names>T.</given-names>
            <surname>Tuytelaars</surname>
          </string-name>
          ,
          <article-title>A continual learning survey: Defying forgetting in classification tasks</article-title>
          ,
          <source>IEEE Transactions on Pattern Analysis and Machine Intelligence</source>
          (
          <year>2021</year>
          )
          <fpage>1</fpage>
          -
          <lpage>1</lpage>
          . URL: https://ieeexplore.ieee.org/document/9349197/. doi:
          <volume>10</volume>
          .1109/TPAMI.
          <year>2021</year>
          .
          <volume>3057446</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <given-names>K.</given-names>
            <surname>Toutanova</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <article-title>Observed versus latent features for knowledge base and text inference</article-title>
          , in: A.
          <string-name>
            <surname>Allauzen</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          <string-name>
            <surname>Grefenstette</surname>
          </string-name>
          ,
          <string-name>
            <surname>K. M. Hermann</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          <string-name>
            <surname>Larochelle</surname>
          </string-name>
          , S. W.-t. Yih (Eds.),
          <source>Proceedings of the 3rd Workshop on Continuous Vector Space Models and their Compositionality</source>
          ,
          <source>Association for Computational Linguistics</source>
          , Beijing, China,
          <year>2015</year>
          , pp.
          <fpage>57</fpage>
          -
          <lpage>66</lpage>
          . URL: https://aclanthology.org/ W15-4007/. doi:
          <volume>10</volume>
          .18653/v1/
          <fpage>W15</fpage>
          -4007.
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [24]
          <string-name>
            <given-names>D. P.</given-names>
            <surname>Kingma</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Ba</surname>
          </string-name>
          ,
          <article-title>Adam: A method for stochastic optimization</article-title>
          ,
          <year>2017</year>
          . URL: https://arxiv.org/abs/ 1412.6980. arXiv:
          <volume>1412</volume>
          .
          <fpage>6980</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          [25]
          <string-name>
            <surname>M. P.-C. Lin</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Chang</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Hall</surname>
          </string-name>
          , G. Jhajj, Preliminary Systematic Review of Open-Source
          <source>Large Language Models in Education</source>
          , Springer Nature Switzerland,
          <year>2024</year>
          , p.
          <fpage>68</fpage>
          -
          <lpage>77</lpage>
          . URL: http://dx.doi. org/10.1007/978-3-
          <fpage>031</fpage>
          -63028-
          <issue>6</issue>
          _6. doi:
          <volume>10</volume>
          .1007/978-3-
          <fpage>031</fpage>
          -63028-
          <issue>6</issue>
          _
          <fpage>6</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          [26]
          <string-name>
            <given-names>H.</given-names>
            <surname>Touvron</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Lavril</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Izacard</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Martinet</surname>
          </string-name>
          , M.
          <article-title>-</article-title>
          <string-name>
            <surname>A. Lachaux</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Lacroix</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Rozière</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          <string-name>
            <surname>Goyal</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          <string-name>
            <surname>Hambro</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Azhar</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Rodriguez</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Joulin</surname>
          </string-name>
          , E. Grave, G. Lample,
          <article-title>Llama: Open and eficient foundation language models</article-title>
          ,
          <year>2023</year>
          . URL: https://arxiv.org/abs/2302.13971. arXiv:
          <volume>2302</volume>
          .
          <fpage>13971</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          [27]
          <string-name>
            <given-names>A. Q.</given-names>
            <surname>Jiang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Sablayrolles</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Mensch</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Bamford</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. S.</given-names>
            <surname>Chaplot</surname>
          </string-name>
          , D. de las Casas,
          <string-name>
            <given-names>F.</given-names>
            <surname>Bressand</surname>
          </string-name>
          , G. Lengyel,
          <string-name>
            <given-names>G.</given-names>
            <surname>Lample</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Saulnier</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L. R.</given-names>
            <surname>Lavaud</surname>
          </string-name>
          , M.
          <article-title>-</article-title>
          <string-name>
            <surname>A. Lachaux</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Stock</surname>
            ,
            <given-names>T. L.</given-names>
          </string-name>
          <string-name>
            <surname>Scao</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Lavril</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Lacroix</surname>
            ,
            <given-names>W. E.</given-names>
          </string-name>
          <string-name>
            <surname>Sayed</surname>
          </string-name>
          , Mistral 7b,
          <year>2023</year>
          . URL: https://arxiv.org/abs/2310.06825. arXiv:
          <volume>2310</volume>
          .
          <fpage>06825</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          [28]
          <string-name>
            <surname>Qwen</surname>
            , :,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Yang</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Yang</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Zhang</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Hui</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Zheng</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Yu</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Li</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Liu</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Huang</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          <string-name>
            <surname>Wei</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          <string-name>
            <surname>Lin</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Yang</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Tu</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Zhang</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Yang</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Yang</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Zhou</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Lin</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          <string-name>
            <surname>Dang</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          <string-name>
            <surname>Lu</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          <string-name>
            <surname>Bao</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          <string-name>
            <surname>Yang</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <string-name>
            <surname>Yu</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Li</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Xue</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Zhang</surname>
            ,
            <given-names>Q.</given-names>
          </string-name>
          <string-name>
            <surname>Zhu</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Men</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Lin</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Li</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Tang</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Xia</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          <string-name>
            <surname>Ren</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          <string-name>
            <surname>Ren</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          <string-name>
            <surname>Fan</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          <string-name>
            <surname>Su</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          <string-name>
            <surname>Zhang</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          <string-name>
            <surname>Wan</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          <string-name>
            <surname>Liu</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          <string-name>
            <surname>Cui</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          <string-name>
            <surname>Zhang</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          <string-name>
            <surname>Qiu</surname>
          </string-name>
          ,
          <source>Qwen2.5 technical report</source>
          ,
          <year>2025</year>
          . URL: https://arxiv.org/abs/2412.15115. arXiv:
          <volume>2412</volume>
          .
          <fpage>15115</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          [29]
          <string-name>
            <given-names>G.</given-names>
            <surname>Jhajj</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. R. D.</given-names>
            <surname>Gustafson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Morland</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C. E.</given-names>
            <surname>Gutierrez</surname>
          </string-name>
          ,
          <string-name>
            <surname>M. P.-C. Lin</surname>
            ,
            <given-names>M. A. A.</given-names>
          </string-name>
          <string-name>
            <surname>Dewan</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Lin</surname>
          </string-name>
          ,
          <article-title>Neuromorphic Knowledge Representation: SNN-Based Relational Inference and Explainability in Knowledge Graphs</article-title>
          ,
          <source>in: International Conference on Intelligent Tutoring Systems</source>
          , Springer,
          <year>2025</year>
          , pp.
          <fpage>159</fpage>
          -
          <lpage>165</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>