<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Validation-Gated Hebbian Learning for Adaptive Agent Memory</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Pragya Singh</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Stanley Yu</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>University of Pennsylvania, Department of Computer and Information Science</institution>
          ,
          <country country="US">USA</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2026</year>
      </pub-date>
      <abstract>
        <p>LLM-based agents struggle with catastrophic forgetting, context limitations, and reasoning drift. Knowledge graphs (KGs) ofer structured memory, but current implementations remain static or do not adapt based on reasoning efectiveness. We introduce Kairos, a multi-agent reasoning system implementing Hebbian plasticity mechanisms for adaptive knowledge graphs. Kairos proposes the formalization of three neuroplasticity-inspired operations: edge strengthening (LTP analog), temporal decay (LTD analog), and emergent connection formation. A key innovation is validation-gated learning, where graph consolidation only occurs when reasoning passes multidimensional quality assessment (logical, grounding, novelty, alignment), preventing hallucination reinforcement. Our proof-of-concept demonstrates that validation-gated Hebbian learning is mechanically sound and shows promising initial results, with adaptive graphs outperforming static baselines. More broadly, these results establish the feasibility of neuro-inspired adaptive agent memory where knowledge structures evolve through validated reasoning efectiveness.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Knowledge Graphs</kwd>
        <kwd>Hebbian Learning</kwd>
        <kwd>Memory</kwd>
        <kwd>Large Language Models</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Large language model (LLM)-based agents face persistent challenges with long-term memory and
reasoning stability, hindered by catastrophic forgetting [
        <xref ref-type="bibr" rid="ref1 ref2">1, 2</xref>
        ], context window limitations [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ], and
reasoning drift [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. Expanded context windows introduce quadratic costs and the "lost in the middle"
efect where information is ignored [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ].
      </p>
      <p>
        Knowledge graphs (KGs) provide structured representations for complex, multi-hop reasoning [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ],
which can ofer a more robust foundation than raw context accumulation. However, important gaps
remain: current systems largely treat graphs as static databases. While graphs may grow through
data ingestion and agents may adapt navigation strategies, the underlying structure rarely learns from
reasoning outcomes.
      </p>
      <p>
        Biological memory ofers inspiration. Synaptic connections strengthen through repeated co-activation
through the Hebbian principle of "neurons that fire together wire together" [
        <xref ref-type="bibr" rid="ref6 ref7">6, 7</xref>
        ]. This suggests KGs
could evolve based on reasoning utility, with structure optimized through validation-based learning
rather than generative expansion alone [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ].
      </p>
      <p>We present Kairos, a multi-agent reasoning system implementing Hebbian plasticity mechanisms for
knowledge graphs. Successful reasoning paths are strengthened (long-term potentiation), unused edges
weaken (long-term depression), and frequently co-activated concepts form emergent connections.
Critically, this adaptation is validation-gated: only reasoning passing multi-dimensional quality assessment
triggers consolidation, analogous to how humans selectively consolidate thoughts judged to be valid
rather than reinforcing all neural activity indiscriminately.</p>
      <p>Our contributions are: 1) A validation-gated learning architecture where graph consolidation is
conditioned on multi-dimensional quality assessment, preventing hallucination reinforcement while enabling
adaptive memory. 2) Formalization of three neuroplasticity-inspired mechanisms (edge
strengthening, temporal decay, emergent connections) for symbolic KG structures in multi-agent systems. 3) A
multi-agent validation framework with specialized agents assessing complementary quality dimensions
(logical consistency, factual grounding, novelty, alignment). 4) A proof-of-concept demonstration on
minimal graphs that these design choices yield mechanical correctness and show initial promise, though
validation at scale remains essential future work.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related Work</title>
      <sec id="sec-2-1">
        <title>2.1. Graph-Based Retrieval and Reasoning Systems</title>
        <p>
          Graph-based retrieval augmentation has evolved significantly from early semantic networks. Microsoft’s
GraphRAG [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ] introduced hierarchical summarization for query-focused retrieval, inspiring eficiency
optimizations in LightRAG [
          <xref ref-type="bibr" rid="ref10">10</xref>
          ] and neurobiologically-inspired indexing in HippoRAG [
          <xref ref-type="bibr" rid="ref11">11</xref>
          ].
Reasoningover-graphs approaches [
          <xref ref-type="bibr" rid="ref12">12</xref>
          ] enable multi-hop inference through traversal, while hybrid systems [
          <xref ref-type="bibr" rid="ref13">13</xref>
          ]
combine semantic and structured search. Plan-on-Graph [
          <xref ref-type="bibr" rid="ref14">14</xref>
          ] introduced adaptive query planning.
These advances focus on retrieval optimization and navigation rather than structural adaptation from
reasoning feedback.
        </p>
      </sec>
      <sec id="sec-2-2">
        <title>2.2. Agent Memory Architectures</title>
        <p>
          Agent memory systems balance storage capacity with retrieval eficiency. MemGPT [
          <xref ref-type="bibr" rid="ref3">3</xref>
          ] pioneered
hierarchical memory management, extended by A-Mem [
          <xref ref-type="bibr" rid="ref15">15</xref>
          ] with atomic linkable units. Generative
Agents [
          <xref ref-type="bibr" rid="ref16">16</xref>
          ] demonstrated importance-weighted retrieval combining recency and relevance. Classical
cognitive architectures, such as ACT-R’s activation-based consolidation [
          <xref ref-type="bibr" rid="ref17">17</xref>
          ] and Soar’s chunking
mechanisms [
          <xref ref-type="bibr" rid="ref18">18</xref>
          ], implement usage-driven adaptation through declarative chunk activation spreading
and procedural rule compilation. However, these operate on predefined symbolic structures without
our validation-gated feedback: ACT-R’s base-level learning strengthens chunks through frequency
and recency, but cannot prune incorrect relations or validate reasoning utility before consolidation.
Multi-agent frameworks [
          <xref ref-type="bibr" rid="ref19 ref20">19, 20</xref>
          ] leverage KGs primarily for coordination rather than adaptive memory.
        </p>
      </sec>
      <sec id="sec-2-3">
        <title>2.3. Dynamic Knowledge Graphs and Continual Learning</title>
        <p>
          Knowledge graph dynamics typically respond to external data rather than internal reasoning. Temporal
approaches like Know-Evolve [
          <xref ref-type="bibr" rid="ref21">21</xref>
          ] model event-driven updates through point processes, while
LLMDA [
          <xref ref-type="bibr" rid="ref22">22</xref>
          ] adapts temporal rules from language understanding. Emergent graph expansion [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ] generates
new structures during reasoning. Continual learning methods [
          <xref ref-type="bibr" rid="ref23">23</xref>
          ] accommodate new entities without
catastrophic forgetting. These methods adapt content but not connection strength based on reasoning
utility.
        </p>
      </sec>
      <sec id="sec-2-4">
        <title>2.4. Neuroscience-Inspired Learning Mechanisms</title>
        <p>
          Biological memory consolidation provides computational metaphors for adaptive systems. Hebbian
plasticity—strengthening co-activated synapses and weakening unused connections [
          <xref ref-type="bibr" rid="ref6 ref7">6, 7</xref>
          ]—has inspired
GNN architectures with activity-dependent weight updates [24, 25]. However, pure Hebbian learning
in graphs lacks task-specific gating: edges strengthen through co-activation regardless of reasoning
correctness. Kairos difers fundamentally through its validation gate—edges strengthen only after LLM
confirmation of reasoning utility, implementing a selectivity absent in unsupervised Hebbian GNNs.
Complementary biological mechanisms include synaptic scaling [26] and systems consolidation [27].
Neural-symbolic AI research [28] identifies dynamic rule learning as an open challenge.
        </p>
      </sec>
      <sec id="sec-2-5">
        <title>2.5. Positioning Kairos</title>
        <p>Kairos uniquely contributes validation-gated structural adaptation: knowledge graphs evolving through
confirmed reasoning efectiveness rather than co-activation patterns (Hebbian GNNs), predefined
symbolic rules (ACT-R), or data ingestion (temporal KGs). Our symbolic structural adaptation—selective
edge strengthening, pruning, and emergent connection formation—operates on discrete graph topology
guided by natural language validation, bridging neural learning principles with interpretable symbolic
updates. This proof-of-concept establishes technical feasibility and identifies critical design principles
for multi-dimensional validation, particularly the orthogonality of novelty and quality assessment.</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. System Architecture</title>
      <sec id="sec-3-1">
        <title>3.1. Core Components</title>
        <p>Query Orchestrator. The orchestrator serves as the system’s entry point, receiving user queries
and routing them to appropriate reasoning modules. Module selection employs semantic similarity
computed via sentence transformers (all-MiniLM-L6-v2), matching query embeddings against module
descriptions. For complex queries requiring multiple perspectives, the orchestrator can invoke several
modules sequentially or in parallel.</p>
        <p>Specialized Reasoning Modules. Kairos employs a modular architecture with domain-specific
agents. Throughout this, we will use the case study of crypto analysis, which has specialized agents
for tasks like security auditing and financial analysis. This demonstrates flexibility across rule-based,
data-driven, and LLM-augmented reasoning paradigms.</p>
        <p>Each module queries the knowledge graph to retrieve relevant entities and relationships, constructs a
reasoning path explaining its inference process, and produces a structured output containing: (1) a
stepby-step reasoning path with data sources and logical inferences, (2) a conclusion with confidence score,
(3) the specific KG triples used (source_triples), and (4) relevant metrics. This structured output enables
downstream validation and Hebbian learning by making explicit which graph elements contributed to
the reasoning.</p>
        <p>Dynamic Knowledge Graph. The knowledge graph represents information as entity-relation
triples with rich metadata. Each relation stores not only subject-predicate-object structure but also
a confidence score (0.0-1.0), source provenance, temporal versioning, and Hebbian-specific metadata
including activation count and cycles since last activation. This metadata enables the system to track
usage patterns over usage. The graph supports both static triples extracted during document ingestion
and emergent relations formed through co-activation patterns during reasoning.</p>
        <p>
          Multi-Agent Validation Layer. Four specialized validation agents assess reasoning quality from
complementary perspectives before any graph adaptation occurs. This diversity of validation helps
prevent premature consolidation of flawed reasoning patterns [
          <xref ref-type="bibr" rid="ref15">15</xref>
          ]. The validation layer outputs
numerical scores (0-1) and textual feedback for each dimension, which gate the Hebbian learning
process.
        </p>
        <p>Aggregator and Results. The aggregator synthesizes outputs from multiple reasoning modules and
validation agents, computing aggregate trust scores and presenting explainable results with provenance
tracking to the user.</p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Validation-Gated Learning Feedback Loop</title>
        <p>
          As show in Figure 1, the critical architectural feature distinguishing Kairos from prior work is the
validation gate between reasoning and learning. When a reasoning module produces output, validation
agents assess quality across four dimensions (detailed in Section D). Only when all validators indicate
acceptable quality (scores above threshold) does the system trigger Hebbian updates to strengthen
the edges traversed during reasoning. This gate serves two purposes: (1) it prevents consolidation of
hallucinated facts or logically incoherent reasoning paths, and (2) it aligns with neuroscientific findings
that successful task completion, not mere neural activity, drives synaptic strengthening [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ]. Failed
reasoning attempts do not trigger strengthening; they simply fade through temporal decay, mirroring
biological learning.
        </p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Hebbian Plasticity for Knowledge Graphs</title>
      <sec id="sec-4-1">
        <title>4.1. Motivation: Beyond Static Knowledge Organization</title>
        <p>Kairos formalizes neuroplasticity mechanisms for KG-based agent memory through three operations:
edges traversed during validated reasoning strengthen (LTP analog), unused edges weaken over time
(LTD analog), and frequently co-activated entities form emergent connections. Our evaluation confirms
these mechanisms operate as specified and identifies critical design principles for multi-dimensional
validation systems.</p>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. Mechanisms</title>
        <sec id="sec-4-2-1">
          <title>4.2.1. Edge Strengthening: Long-Term Potentiation Analog</title>
          <p>When a reasoning module’s output passes validation, the KG edges that it used as a source receive a
strengthening signal. We implement asymptotic strengthening with diminishing returns:
Δstrength =  × (max_strength − current_strength)
(1)
new_strength = min(max_strength, current_strength + Δstrength)
(2)
where  is the learning rate (see Appendix A.4) and max_strength = 1.0. This formulation allows
for noticeable adaptation within 10-20 episodes while preventing single-trial over-consolidation. For
multi-hop paths, each edge receives proportional strengthening.</p>
        </sec>
        <sec id="sec-4-2-2">
          <title>4.2.2. Temporal Decay: Long-Term Depression Analog</title>
          <p>
            Edges not traversed during reasoning gradually weaken via temporal decay, analogous to synaptic
depression [
            <xref ref-type="bibr" rid="ref7">7</xref>
            ]. We implement exponential decay:
          </p>
          <p>where  is the decay rate,  = 5 is the half-life in reasoning cycles, and min_strength = 0.1 is a
pruning threshold (see Appendix A.4). This mechanism balances knowledge retention against removal
of irrelevant associations, adapting to actual usage patterns rather than wall-clock time while preventing
catastrophic forgetting.</p>
        </sec>
        <sec id="sec-4-2-3">
          <title>4.2.3. Emergent Connection Formation</title>
          <p>Kairos also forms emergent relationships by tracking entity co-activations. When two entities appear
together in reasoning contexts past a certain threshold  , a new co_occurs_with edge is created:
initial_strength = min(0.5, count × 0.1)</p>
          <p>This discovers implicit, cross-domain connections. We experiment with a threshold  = 3 to reduce
noise, and the initial strength is capped at 0.5 to distinguish empirical edges from source-derived facts.</p>
          <p>Example: If entities Security-Audit and High-Risk are co-activated three times across diferent
queries in our blockchain analysis domain, Kairos creates an emergent edge:
decay =  ×
︂(
1 − exp
︂(
−
cycles_inactive )︂</p>
          <p>new_strength = max(min_strength, current_strength − decay)
(3)
(4)
(5)
(6)</p>
          <p>Security-Audit –[co_occurs_with, strength=0.3]–&gt; High-Risk</p>
          <p>This new connection captures an empirical pattern that can accelerate future reasoning within the
domain.</p>
          <p>trigger_hebbian ←
{︃True if all validators pass: .valid = True</p>
          <p>False otherwise
where each validator  produces a binary validity decision. Edge strengthening and entity
coactivation tracking occur unconditionally during reasoning, but consolidation operations (emergent
edge formation and temporal decay) only execute when all validators indicate successful reasoning.
This implements consolidation inspired by reward-modulated plasticity.</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Results</title>
      <p>We evaluate Kairos as a proof-of-concept system through three complementary experiments: (1)
Mechanical Validation: Confirming Hebbian mechanisms operate as designed by tracking edge strength
evolution over repeated reasoning cycles, (2) Utility Validation: Demonstrating that adaptive graphs
outperform static baselines through direct comparison; and (3) Architectural Analysis: Examining
the contribution of each system component through ablation study. Our evaluation uses a minimal
knowledge graph representing a blockchain security audit scenario (ApolloContract smart contract
with known vulnerabilities).</p>
      <sec id="sec-5-1">
        <title>5.1. Experimental Setup</title>
        <p>Evaluation Dataset. We constructed a 60-question evaluation dataset covering diverse query types:
security audits, risk analysis, multi-hop reasoning, counterfactual scenarios, and meta-reasoning tasks.
Questions vary in complexity from simple entity lookups ("Has ApolloContract been audited?") to
complex synthesis tasks ("What is the holistic assessment of deploying smart contracts in the current
environment?"). This minimal setup serves as a controlled proof-of-concept for validating that the
Cycle
Hebbian mechanisms operate as specified. However, it is insuficient to demonstrate scalability or
practical value on real-world reasoning tasks.</p>
        <p>Metrics. We measure: (1) Trust score: average of four validation dimensions (0-1 scale), serving
as an aggregate quality metric; (2) Individual validation scores: logical coherence, factual grounding,
novelty, and alignment (each 0-1); (3) Hebbian metrics: edges strengthened, entities activated, emergent
connections formed; (4) Edge strength: confidence values of frequently-traversed graph edges (0-1).</p>
      </sec>
      <sec id="sec-5-2">
        <title>5.2. Mechanical Validation: Hebbian Plasticity Over Time</title>
        <p>To evaluate whether Hebbian mechanisms operate as designed, we run 5 reasoning cycles with the
same set of 3 queries repeated in each cycle. We track edge strength evolution for frequently-traversed
paths. Results are shown in Table 1.</p>
        <p>Edge Strengthening Confirmation. The average strength of frequently-used edges increases
monotonically from 0.919 (cycle 1) to 0.977 (cycle 5), a 6.3% gain. This confirms the Hebbian mechanism
operates according to the asymptotic strengthening formula (Eq. 1): with learning rate  = 0.1 , edges
approach maximum strength (1.0) gradually through repeated activations. The mechanical behavior
matches the designed specification.</p>
        <p>Temporal Decay Confirmation. While emergent connections did not form due to minimal graph
structure, temporal decay operated as designed. Edges not traversed in later cycles showed strength
reduction consistent with the exponential decay formula (Eq. 3-4). With  = 5 cycles, unused edges
exhibited decay behavior confirming the mechanism functions according to specification.</p>
        <p>Emergent Connections. No emergent connections formed during this evaluation. Given the
minimal graph structure and the co-activation threshold ( = 3), the limited entity diversity constrains
emergent edge formation. Emergent connection formation would require either a larger knowledge
graph with more entities or diverse queries that repeatedly co-activate entity pairs not directly connected
in the initial graph structure.</p>
      </sec>
      <sec id="sec-5-3">
        <title>5.3. Utility Validation: Adaptive vs Static Knowledge Graphs</title>
        <p>To demonstrate that Hebbian adaptation provides practical value, we compare adaptive graphs (with
Hebbian updates enabled) against static baselines (Hebbian updates disabled) across 5 reasoning cycles.
Each cycle processes the same 3 queries, allowing us to observe cumulative adaptation efects. Table 2
presents results.</p>
        <p>Adaptive Graphs Outperform Static Baselines. Across all 5 cycles, adaptive graphs consistently
achieve higher trust scores than static baselines, with an average improvement of 4.0%. The advantage
grows slightly over cycles (3.7% in cycle 1 → 4.6% in cycle 5), suggesting cumulative benefits from
edge consolidation. While the minimal graph structure limits the magnitude of observable efects, the
consistent directionality demonstrates that Hebbian adaptation provides measurable value. A paired
t-test comparing adaptive vs static scores across cycles shows statistical significance ( (4) = 8.45,
 = 0.001, Cohen’s  = 1.52), confirming that the observed improvement is not due to random
variation.</p>
        <p>Edge Strengthening Correlates with Performance. The performance advantage emerges as
frequently-used edges strengthen through repeated validation-gated updates. In the adaptive condition,
edge weights increase from 0.919 (cycle 1) to 0.977 (cycle 5), while static graphs maintain constant
weights. This demonstrates that the Hebbian mechanism not only operates mechanically but translates
into improved reasoning outcomes, even on minimal graph structures.</p>
      </sec>
      <sec id="sec-5-4">
        <title>5.4. Architectural Analysis: Component Contributions</title>
        <p>To assess each component’s contribution, we evaluate six system configurations across 15 diverse
questions (90 total reasoning episodes). Table 3 presents results.</p>
        <p>Hebbian Plasticity Removal. Removing Hebbian learning shows no measurable impact within
single-episode evaluation (−0.9% ,  = 0.90). This is expected: plasticity mechanisms strengthen edges
over repeated reasoning episodes, but have minimal efect on individual queries. Section 5.3 examines
the cumulative benefits of adaptation.</p>
        <p>Validation Architecture Insights. Individual validator ablations reveal important design
considerations:
• Novelty removal produces the study’s only statistically significant efect: trust scores increase by
21.5% ( &lt; 0.001) when novelty validation is excluded. This reveals a fundamental mismatch
between novelty and quality assessment. Novelty validators penalize straightforward factual
retrieval (low novelty = low score), while other validators reward accurate factual responses
(correct = high score). Averaging these conflicting signals degrades the aggregate metric. The
data suggests novelty detection and quality validation serve diferent purposes and should be
treated as separate dimensions rather than averaged into a single trust score.
• Grounding removal shows the largest degradation trend (−13.8% ,  = 0.13), though limited
sample size ( = 15) prevents statistical significance. The direction suggests factual verification
may help maintain reasoning quality, but stronger conclusions require larger evaluation.
• Logical removal shows minimal impact (−6.4% ,  = 0.44), suggesting reasoning modules produce
generally coherent outputs in this domain.
• Alignment removal shows minimal impact (+5.7%,  = 0.35). This likely reflects the evaluation
queries rather than alignment’s general importance. Our test set focuses on factual retrieval
rather than preference-sensitive or ethically complex reasoning.</p>
        <p>Note on "No Validation" Condition. The zero trust score for this condition is a measurement
artifact, as the score is derived from the validators themselves; the reasoning modules still produce
output, but the metric is undefined without the validation layer.</p>
      </sec>
      <sec id="sec-5-5">
        <title>5.5. Qualitative Analysis</title>
        <p>The system’s structured output format, which requires each reasoning path to be paired with its specific
source triples, is a critical architectural choice. This explicit provenance directly enables the
validationgated learning loop. The Grounding VN module uses the source triple list to verify each claim against
the KG, allowing for precise error checking. Following successful validation, the Hebbian module uses
the same list to reinforce the exact edges that contributed to the output. This design provides the
essential mechanism for the feedback loop between reasoning quality and knowledge graph adaptation.</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>6. Discussion and Future Work</title>
      <sec id="sec-6-1">
        <title>6.1. Discussion</title>
        <p>Our work demonstrates the viability of neuroplasticity-inspired mechanisms for symbolic knowledge
graphs in multi-agent systems. The Hebbian plasticity evaluation confirms these mechanisms operate
as designed, with edge strengthening following the specified asymptotic formula and adaptive graphs
outperforming static baselines by 4.0% (p = 0.001). This proof-of-concept establishes that biological
memory principles can be formalized for symbolic reasoning architectures, opening a research direction
where knowledge graphs function as adaptive cognitive substrates rather than static databases. We
also identify a critical design principle: novelty and quality are orthogonal dimensions that degrade
when averaged. Our ablation study shows performance increases by 21.5% (p &lt; 0.001, Cohen’s d = 1.39)
when novelty is removed from aggregate trust scores. This finding generalizes beyond our system, with
immediate practical implications for practitioners building multi-dimensional assessment systems.</p>
      </sec>
      <sec id="sec-6-2">
        <title>6.2. Limitations</title>
        <p>As a proof of concept, this work has significant limitations driven primarily by computational constraints.
Lack of compute, including API rate limiting and smaller models, constrained our evaluation to a
minimal knowledge graph and small sample sizes. This prevented comprehensive evaluation on
standard benchmarks, comparison against established baselines at scale, and extensive hyperparameter
optimization. While our results demonstrate feasibility and identify design principles, questions about
scalability, optimal hyperparameter configurations, and performance on complex multi-hop reasoning
tasks remain empirical questions requiring greater computational resources to address conclusively.</p>
      </sec>
      <sec id="sec-6-3">
        <title>6.3. Future Directions</title>
        <p>Future work must first address robustness and scalability: (1) benchmarking on standard datasets
(HotpotQA, MetaQA) with larger graphs (100+ entities); (2) hyperparameter optimization across diverse
domains; (3) comparison against established systems (GraphRAG, MemGPT). Beyond validation at scale,
promising directions include: (4) longitudinal studies analyzing adaptation dynamics and failure modes
over extended episodes; (5) enhancing validation through ensemble methods and user feedback; and
(6) exploring GNN-based reasoning over adaptive graphs and dynamic module selection leveraging
emergent connections.</p>
      </sec>
    </sec>
    <sec id="sec-7">
      <title>7. Conclusion</title>
      <p>We presented Kairos, a system implementing Hebbian plasticity mechanisms for knowledge graphs
in multi-agent reasoning. Our controlled proof-of-concept, while limited in scale by computational
constraints, demonstrates the mechanical feasibility of neuroplasticity-inspired mechanisms and
identiifes critical design principles: adaptive graphs outperformed static baselines by 4.0% (  = 0.001), and
novelty and quality are orthogonal dimensions (21.5% improvement when separated,  &lt; 0.001).</p>
      <p>However, we emphasize that substantial future work is essential before practical deployment.
Our evaluation on minimal graphs and small sample sizes establishes feasibility but cannot validate
scalability, robustness, or performance on real-world tasks. Comprehensive benchmarking on standard
datasets, comparison against established baselines, and evaluation at production scale are critical next
steps to determine whether these mechanisms provide meaningful value beyond controlled settings.</p>
      <p>Despite these limitations, Kairos demonstrates that knowledge graphs can function as adaptive
cognitive substrates rather than static databases. If validated at scale, such systems could enable agents
with episodic-to-semantic memory consolidation, where repeated reasoning patterns automatically
strengthen into semantic knowledge. The validation-gated learning pattern and novelty-quality
separation ofer actionable guidance for building multi-dimensional assessment systems, though their
efectiveness across domains and scales remains an open empirical question.</p>
    </sec>
    <sec id="sec-8">
      <title>Declaration on Generative AI</title>
      <p>During the preparation of this work, the authors used LLMs in order to do review the writing style and
to assist with code generation. After using these tools, the authors reviewed and edited the content as
needed and take full responsibility for the publication’s content.
dings, IEEE Robotics and Automation Letters 6 (2021) 1128–1135.
[24] Y. Liu, W. Zhang, X. Chen, H. Wang, Multi-view incremental learning with structured hebbian
plasticity for enhanced fusion eficiency, arXiv preprint arXiv:2412.12801 (2024).
[25] E. Najarro, S. Sudhakaran, S. Risi, Meta-learning through hebbian plasticity in random networks,</p>
      <p>Neural Networks 158 (2023) 1–13.
[26] G. G. Turrigiano, The self-tuning neuron: Synaptic scaling of excitatory synapses, Cell 135 (2008)
422–435.
[27] Y. Dudai, The neurobiology of consolidations, or, how stable is the engram?, Annual Review of</p>
      <p>Psychology 55 (2004) 51–86.
[28] J. Yang, C. Li, W. Jiang, Ai reasoning in deep learning era: From symbolic ai to neural-symbolic ai,</p>
      <p>Mathematics 13 (2025) 1707.</p>
    </sec>
    <sec id="sec-9">
      <title>A. Implementation Details</title>
      <sec id="sec-9-1">
        <title>A.1. Technical Stack</title>
        <p>Backend:
• Language: Python 3.8+
• Web Framework: FastAPI 0.104.0, Flask 3.0.0
• LLM API: Anthropic Claude-3 Haiku (claude-3-haiku-20240307)
• Embeddings: Sentence Transformers 2.2.0 (all-MiniLM-L6-v2 model)
• NLP: spaCy 3.7.0, Transformers 4.35.0
• Knowledge Graph Storage: JSON-based with in-memory processing
Frontend:
• Framework: Next.js 14.0.0, React 18.2.0
• UI Components: Radix UI, Tailwind CSS 3.3.0
• Language: TypeScript 5.2.0</p>
      </sec>
      <sec id="sec-9-2">
        <title>A.2. Computational Resources</title>
        <p>Development and demonstration runs were conducted on standard CPU infrastructure without
specialized hardware acceleration. The system does not require GPU resources for core functionality,
as reasoning and validation leverage cloud-based LLM APIs (Anthropic Claude-3 Haiku via the
Anthropic API). Sentence transformer embeddings (all-MiniLM-L6-v2) run eficiently on CPU for the scales
demonstrated (knowledge graphs with hundreds to thousands of entities).</p>
        <p>For production deployment at larger scales, GPU acceleration would benefit embedding computation
and enable local LLM inference, though the current cloud API approach was chosen for accessibility
and reproducibility.</p>
      </sec>
      <sec id="sec-9-3">
        <title>A.3. Hyperparameters</title>
        <p>Key hyperparameters for Hebbian learning mechanisms:
• Learning rate (): 0.1
• Maximum edge strength: 1.0
• Decay rate (): 0.05
• Temporal decay half-life (): 5 reasoning cycles
• Minimum edge strength (pruning threshold): 0.1
• Co-activation threshold ( ): 3
• Validation pass thresholds: 0.7 (logical, grounding), 0.5 (novelty, alignment)</p>
      </sec>
      <sec id="sec-9-4">
        <title>A.4. Hyperparameter Optimization</title>
        <p>For strengthening, We chose values around  = 0.1 to prevent single-trial over-consolidation
while allowing noticeable adaptation within 10-20 reasoning episodes. With this learning rate,
an edge at 0.5 strength requires approximately 7 activations to reach 0.95 strength, striking a
balance between rapid adaptation and stability. Higher learning rates ( &gt; 0.3 ) risk premature
convergence to maximum strength, while lower rates ( &lt; 0.05 ) require impractically many
episodes to observe meaningful strengthening.</p>
        <p>We chose decay parameters around  = 0.05,  = 5 . This creates a forgetting curve where edges
retain approximately 63% strength after 5 unused cycles and 95% after 1 cycle. This gradual
decay prevents catastrophic forgetting of temporarily unused knowledge while allowing obsolete
connections to eventually fade. Additional work remains to do more robust hyperparameter
optimization for strengthening and decay to achieve maximal performance.</p>
      </sec>
    </sec>
    <sec id="sec-10">
      <title>B. Specialized Reasoning Modules</title>
      <p>Kairos employs a modular architecture supporting domain-specific reasoning agents. For our
proofof-concept, we implement four reasoning modules: (1) SecurityAuditRM: A rule-based module that
applies predefined security rules (loaded from JSON) to detect vulnerabilities in smart contracts;
(2) MacroAnalysisRM: A data-driven module analyzing macroeconomic trends from local CSV data
(interest rates, inflation); (3) CorporateCommRM: A sentiment analysis module processing corporate
announcements from JSON files; and (4) FinancialAnalysisRM: An LLM-augmented module using
Claude-3 Haiku for dynamic financial risk assessment over KG facts. This heterogeneous module design
demonstrates the architecture’s flexibility across rule-based, data-driven, and LLM-based reasoning
paradigms.</p>
    </sec>
    <sec id="sec-11">
      <title>C. Code Availability</title>
    </sec>
    <sec id="sec-12">
      <title>D. Validator Nodes</title>
      <p>Complete source code, documentation, usage examples, and demonstration scenarios are available in
the project repository.</p>
      <p>
        Logical Validation (LogicalVN) Analyzes the coherence of reasoning paths using an LLM (Claude-3
Haiku) to check for contradictions and logical fallacies (e.g., circular reasoning). It outputs a 0-1 score
and textual feedback. While LLM-based logical assessment has limitations, it provides a practical proxy
for coherence [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ].
      </p>
      <p>Grounding Validation (GroundingVN) Verifies that reasoning claims are anchored in KG facts.
The validator parses claimed triples from the reasoning path, queries the KG, and computes a grounding
ratio:
grounding_score =</p>
      <p>verified_triples
total_claimed_triples
(7)</p>
      <p>
        A 1.0 score indicates all claims are grounded. This component aims to detect when reasoning modules
generate logically coherent but factually unsupported claims [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ].
      </p>
      <p>Novelty Validation (NoveltyVN) Assesses whether a conclusion represents emergent insight or
straightforward fact retrieval, using Claude-3 Haiku to compare reasoning outputs against KG facts.
Unlike other validators that assess quality, this component identifies creative synthesis. As our ablation
study reveals, novelty and quality assessment serve diferent purposes and can conflict when averaged
together as novelty validators penalize accurate factual retrieval, which quality validators reward.</p>
      <p>
        Alignment Validation (AlignmentVN) Checks whether reasoning respects user-defined
preferences, goals, and ethical constraints (e.g., "prioritize risk mitigation") using Claude-3 Haiku to assess
reasoning against these constraints. While comprehensive alignment specification remains an open
challenge [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ], this component provides an architectural placeholder for preference-aware validation.
      </p>
      <p>Trust Score Aggregation After all four validators produce scores, Kairos computes an aggregate
trust score via simple averaging:
4
trust_score = 1 ∑︁ 
4
=1
(8)</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>H.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Ding</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Fang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Tao</surname>
          </string-name>
          ,
          <article-title>Revisiting catastrophic forgetting in large language model tuning, in: Findings of the Association for Computational Linguistics: EMNLP 2024, Association for Computational Linguistics</article-title>
          , Miami, Florida, USA,
          <year>2024</year>
          , pp.
          <fpage>4297</fpage>
          -
          <lpage>4308</lpage>
          . doi:
          <volume>10</volume>
          .18653/v1/
          <year>2024</year>
          . findings-emnlp.
          <volume>249</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Luo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Meng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <string-name>
            <surname>Y. Zhang,</surname>
          </string-name>
          <article-title>An empirical study of catastrophic forgetting in large language models during continual fine-tuning</article-title>
          ,
          <source>arXiv preprint arXiv:2308.08747</source>
          (
          <year>2023</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>C.</given-names>
            <surname>Packer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Wooders</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Lin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Fang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. G.</given-names>
            <surname>Patil</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Stoica</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. E.</given-names>
            <surname>Gonzalez</surname>
          </string-name>
          , Memgpt:
          <article-title>Towards llms as operating systems</article-title>
          ,
          <source>arXiv preprint arXiv:2310.08560</source>
          (
          <year>2023</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>L.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Zaharia</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Zou</surname>
          </string-name>
          ,
          <article-title>How is chatgpt's behavior changing over time?</article-title>
          ,
          <source>arXiv preprint arXiv:2307.09009</source>
          (
          <year>2023</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>S.</given-names>
            <surname>Pan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Luo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <article-title>Unifying large language models and knowledge graphs: A roadmap</article-title>
          ,
          <source>IEEE Transactions on Knowledge and Data Engineering</source>
          <volume>36</volume>
          (
          <year>2024</year>
          )
          <fpage>3580</fpage>
          -
          <lpage>3599</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>D. O.</given-names>
            <surname>Hebb</surname>
          </string-name>
          ,
          <article-title>The Organization of Behavior: A Neuropsychological Theory</article-title>
          , John Wiley &amp; Sons,
          <year>1949</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>L. R.</given-names>
            <surname>Squire</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Genzel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. T.</given-names>
            <surname>Wixted</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. G.</given-names>
            <surname>Morris</surname>
          </string-name>
          , Memory consolidation,
          <source>Cold Spring Harbor Perspectives in Biology</source>
          <volume>7</volume>
          (
          <year>2015</year>
          )
          <article-title>a021766</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>M. J.</given-names>
            <surname>Buehler</surname>
          </string-name>
          ,
          <article-title>Agentic deep graph reasoning yields self-organizing knowledge networks</article-title>
          ,
          <source>arXiv preprint arXiv:2502.13025</source>
          (
          <year>2025</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>D.</given-names>
            <surname>Edge</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Trinh</surname>
          </string-name>
          , N. Cheng, J.
          <string-name>
            <surname>Bradley</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Chao</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Mody</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Truitt</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Larson</surname>
          </string-name>
          ,
          <article-title>From local to global: A graph rag approach to query-focused summarization</article-title>
          ,
          <source>arXiv preprint arXiv:2404.16130</source>
          (
          <year>2024</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Guo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Lam</surname>
          </string-name>
          , Z. Cheng,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Shen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <article-title>Lightrag: Simple and fast retrieval-augmented generation</article-title>
          ,
          <source>arXiv preprint arXiv:2410.05779</source>
          (
          <year>2024</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>B. J.</given-names>
            <surname>Bae</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Ye</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Kim</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Hwang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Heo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Kim</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Zhai</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Feng</surname>
          </string-name>
          , Hipporag:
          <article-title>Neurobiologically inspired long-term memory for large language models</article-title>
          ,
          <source>arXiv preprint arXiv:2405.14831</source>
          (
          <year>2024</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>L.</given-names>
            <surname>Luo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.-F.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Hafari</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Pan</surname>
          </string-name>
          ,
          <article-title>Reasoning on graphs: Faithful and interpretable large language model reasoning</article-title>
          ,
          <source>in: International Conference on Learning Representations (ICLR)</source>
          ,
          <year>2024</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>B.</given-names>
            <surname>Sarmah</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Mehta</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Pasquali</surname>
          </string-name>
          , T. Zhu,
          <article-title>Hybridrag: Integrating knowledge graphs and vector retrieval augmented generation for eficient information extraction</article-title>
          ,
          <source>in: Proceedings of the 5th ACM International Conference on AI in Finance</source>
          ,
          <year>2024</year>
          , pp.
          <fpage>691</fpage>
          -
          <lpage>699</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>L.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Tong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Jin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Sun</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Ye</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Xiong</surname>
          </string-name>
          ,
          <article-title>Plan-on-graph: Self-correcting adaptive planning of large language model on knowledge graphs</article-title>
          ,
          <source>in: Proceedings of the 38th Conference on Neural Information Processing Systems (NeurIPS)</source>
          ,
          <year>2024</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>W.</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Liang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Mei</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Gao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Tan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , A-mem:
          <article-title>Agentic memory for llm agents</article-title>
          ,
          <source>arXiv preprint arXiv:2502.12110</source>
          (
          <year>2025</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>J. S.</given-names>
            <surname>Park</surname>
          </string-name>
          ,
          <string-name>
            <surname>J. C. O'Brien</surname>
            ,
            <given-names>C. J.</given-names>
          </string-name>
          <string-name>
            <surname>Cai</surname>
            ,
            <given-names>M. R.</given-names>
          </string-name>
          <string-name>
            <surname>Morris</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Liang</surname>
            ,
            <given-names>M. S.</given-names>
          </string-name>
          <string-name>
            <surname>Bernstein</surname>
          </string-name>
          ,
          <article-title>Generative agents: Interactive simulacra of human behavior</article-title>
          ,
          <source>in: Proceedings of the 36th Annual ACM Symposium on User Interface Software and Technology</source>
          ,
          <year>2023</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>22</lpage>
          . doi:
          <volume>10</volume>
          .1145/3586183.3606763.
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <surname>J. R. Anderson</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Bothell</surname>
            ,
            <given-names>M. D.</given-names>
          </string-name>
          <string-name>
            <surname>Byrne</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Douglass</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Lebiere</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          <string-name>
            <surname>Qin</surname>
          </string-name>
          ,
          <article-title>An Integrated Theory of the Mind</article-title>
          , Psychological Review,
          <year>2004</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>J. E.</given-names>
            <surname>Laird</surname>
          </string-name>
          ,
          <source>The Soar Cognitive Architecture</source>
          , MIT Press,
          <year>2012</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>J.</given-names>
            <surname>Liang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Sun</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Bai</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Xiao</surname>
          </string-name>
          ,
          <article-title>Beyond isolation: Multi-agent synergy for improving knowledge graph construction</article-title>
          ,
          <source>arXiv preprint arXiv:2312.03022</source>
          (
          <year>2023</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <article-title>Graphs meet ai agents: Taxonomy, progress, and future opportunities</article-title>
          ,
          <source>arXiv preprint arXiv:2506.18019</source>
          (
          <year>2024</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>R.</given-names>
            <surname>Trivedi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Dai</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Song</surname>
          </string-name>
          , Know-evolve:
          <article-title>Deep temporal reasoning for dynamic knowledge graphs</article-title>
          ,
          <source>in: International Conference on Machine Learning, PMLR</source>
          ,
          <year>2017</year>
          , pp.
          <fpage>3462</fpage>
          -
          <lpage>3471</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>J.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Pan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Luo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <article-title>Large language models-guided dynamic adaptation for temporal knowledge graph reasoning</article-title>
          ,
          <source>in: Proceedings of the 38th Conference on Neural Information Processing Systems (NeurIPS)</source>
          ,
          <year>2024</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <given-names>A.</given-names>
            <surname>Daruna</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Gupta</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Mahoudi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Chernova</surname>
          </string-name>
          ,
          <article-title>Continual learning of knowledge graph embed-</article-title>
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>