<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>F. Spadea);</journal-title>
      </journal-title-group>
      <issn pub-type="ppub">1613-0073</issn>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Graph Adaptation for LLM Recom mendations</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Fernando Spadea</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Oshani Seneviratne</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Workshop</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="editor">
          <string-name>Personalized Knowledge Graphs, Large Language Models, Recommendation Systems, Over-Personalization, Filter</string-name>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Bubbles</institution>
          ,
          <addr-line>Neuro-Symbolic AI, Knowledge Graph Adaptation, User-Centric AI, Recommendation Diversity</addr-line>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Rensselaer Polytechnic Institute</institution>
          ,
          <addr-line>Troy, NY</addr-line>
          ,
          <country country="US">USA</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2025</year>
      </pub-date>
      <volume>000</volume>
      <fpage>0</fpage>
      <lpage>0001</lpage>
      <abstract>
        <p>We present a lightweight neuro-symbolic framework to mitigate over-personalization in LLM-based recommender systems by adapting user-side Knowledge Graphs (KGs) at inference time. Instead of retraining models or relying on opaque heuristics, our method restructures a user's Personalized Knowledge Graph (PKG) to suppress feature co-occurrence patterns that reinforce Personalized Information Environments (PIEs), i.e., algorithmically induced filter bubbles that constrain content diversity. These adapted PKGs are used to construct structured prompts that steer the language model toward more diverse, Out-PIE recommendations while preserving topical relevance. We introduce a family of symbolic adaptation strategies, including soft reweighting, hard inversion, and targeted removal of biased triples, and a client-side learning algorithm that optimizes their application per user. Experiments on a recipe recommendation benchmark show that personalized PKG adaptations significantly increase content novelty while maintaining recommendation quality, outperforming global adaptation and naive prompt-based methods.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>Overly tailored recommendations lead to Personalized Information Environments (PIEs), which are
algorithmically reinforced content silos, akin to filter bubbles, where new information is repeatedly
ifltered through prior user preferences [ 1]. Although PIEs can initially enhance relevance, they often
narrow exposure to diverse content, reduce user agency, and inhibit discovery [2, 3]. Without
transparency or control mechanisms, users struggle to diversify their content landscape [4, 5, 6]. Many early
approaches to filter bubble mitigation rely on KG embeddings [ 7] or influence-based retraining [ 8],
which can be efective but often require retraining, complex domain modeling, or hardcoded logic.</p>
      <p>Building on our prior work [9], we introduce a lightweight inference-time method for escaping PIEs
without retraining the underlying language model. By adapting a user’s Personalized Knowledge Graph
(PKG), which is a structured, editable representation of preferences, we enable interpretable, symbolic
control over the recommendation process. Our rule-based PKG adaptation selectively down-weights
overrepresented features in the user’s profile, steering large language models (LLMs) to generate more
novel recommendations while preserving relevance.</p>
      <p>Consider a user whose preferences are captured by a PKG. Over time, their
interactions with a recommender system create biased associations, forming a PIE. For instance, as shown
in Figure 1 Base PKG, a user may consistently give high ratings to Italian dishes containing tomatoes,
such as “Tomato Pasta,” “Lasagna,” and “Margherita Pizza.” Consequently, the user’s PKG becomes
dominated by a strong association between “Italian” cuisine and the feature “Tomato.” Although initially
beneficial for relevance, this over-personalization ultimately limits novelty. Ideally, when seeking new
Italian recipes, the user would appreciate recommendations that introduce variety, such as “Pesto Pasta,”</p>
      <p>CEUR</p>
      <p>ceur-ws.org
Margherita</p>
      <p>Pizza</p>
      <p>5 3
Cannoli 4 User 4</p>
      <p>5 3 2
which align with their preference for Italian dishes but break the entrenched link with tomato-based
recipes.</p>
      <p>Yes</p>
      <p>Knowledge Graph Update</p>
      <p>Rule Adaptation</p>
      <p>Derive KG Completion
2. Methodology
(raFeei.grcgeuo.c,rmoe“Smm2umpegnrgeodenvasdittdaieaotsninoaInnptaiqopliuveaeelnirrnvydeiie.isswhiW”os)sf,huoteheunder PKG APvNIoEoi?d AdapPtKedGPKG SysUtseemr PPrroommppInttput Response
soqyvuseetrre-ymperrisfirsosktnsaclrhiezeianctkifoosnrc,wionhrgePtIahEe.krInftoshowe,na User Request PPIKEG==PPeersrsoonnaallIinzLfeeodgrmeKnandtoiownleEdngveiroGnrampehnt PKG FLinLeM-Tuned
soft/hard/removal adaptation is
applied to adapt the PKG, selectively Figure 2: System overview.
modifying it to reduce the influence of overrepresented feature pairs (such as “Italian + tomato”). The
adapted PKG is used to construct a structured prompt that guides the LLM. This prompt reflects the
user’s preferences but intentionally avoids PIE-aligned content. The LLM then generates novel, relevant
recommendations that satisfy the user’s query.</p>
      <sec id="sec-1-1">
        <title>2.1. Prompt Construction</title>
        <p>The prompt consists of two parts: a system message and a user message. The system message instructs
the model to perform KG completion and provides the PKG. It is formatted as follows:</p>
        <sec id="sec-1-1-1">
          <title>System Message</title>
          <p>You perform Knowledge Graph Completion. You will recommend a new triple to add to
the user's knowledge graph with a tail entity that isn't already in their knowledge
graph. The user's entity is represented by {User ID}. Use this knowledge graph when
responding to their queries: {Knowledge Graph}</p>
          <p>The user message requests a recommendation with a specific trait. For our train and test dataset, it is
formatted as:</p>
        </sec>
        <sec id="sec-1-1-2">
          <title>User Message</title>
          <p>Recommend a recipe with trait of {Relation Type} -&gt; {Trait Value}.</p>
          <p>In the user message, Relation Type is either hasIngredient (for ingredients) or hasTag (for
Food.com tags), and Trait Value corresponds to a specific ingredient or tag.</p>
        </sec>
      </sec>
      <sec id="sec-1-2">
        <title>2.2. Detecting PIEs</title>
        <p>Our system addresses over-personalization by adapting the user’s PKG that contains preferences derived
from prior interactions (e.g., ratings of recipes annotated with ingredients or tags).</p>
        <p>We define a PIE as a user-specific bias toward certain co-occurring pairs of features. Given a pair
( given,  bias), we detect a PIE if a user consistently assigns significantly higher or lower ratings to items
containing both features compared to items containing only  given. For example, as illustrated by the
Base PKG in Figure 1, the pair ( given = Italian,  bias = tomato) forms a positively biased feature
association, creating a “tomato-centric” PIE. Our objective is to detect these biases and adapt the PKG
accordingly to generate recommendations that remain relevant but introduce greater diversity. We
quantify the strength of a PIE using a feature-pair bias score,  bias:
 bias( given,  bias) =
1</p>
        <p>
          ∑
 neutral ⋅ |  given| ∈  given, bias
(  () −  neutral)
(
          <xref ref-type="bibr" rid="ref1">1</xref>
          )
        </p>
        <p>Here,  neutral serves as a baseline for interpreting user preferences (e.g., 2.5 on a 0–5 scale), and   ()
is the rating that user  assigns to item  . The set   given includes all items in the user’s PKG that contain
the feature  given, while   given, bias further restricts this to items that also include  bias.</p>
        <p>The bias score  bias captures how much the user’s ratings for items with both features deviate
from the neutral point, relative to the size of the user’s PKG. A high positive score indicates user
preference amplification for feature pairs (e.g., Italian dishes with tomato), while a strong negative score
indicates underrating. When  bias exceeds a set threshold (e.g., ±0.5), we consider the feature pair to be
PIE-inducing and subject to adaptation.</p>
      </sec>
      <sec id="sec-1-3">
        <title>2.3. Adapting PKGs</title>
        <p>Upon detecting a PIE, we apply one of three symbolic adaptation strategies to selectively modify the
PKG before passing it to the LLM, as illustrated in Figure 1:
• Soft Adaptation: Adjusts ratings of PIE-aligned items by symmetrically inverting their strength
around a neutral midpoint, preserving relative preference order. For instance, highly rated
tomato-based Italian dishes like “Margherita Pizza” (rating 5) become somewhat lightly rated
(rating 2), while “Tomato Sauce Pasta” (rating 4) gets a harsher rating (rating 1), gently nudging
recommendations away from tomato dishes while maintaining the user’s cuisine preference.
• Hard Adaptation: Aggressively assigns the extreme opposite ratings to PIE-aligned items. For
example, dishes previously highly favored by the user, such as “Margherita Pizza” (rating 5) and
“Tomato Sauce Pasta” (rating 4), receive the lowest possible rating (rating 0), strongly discouraging
their recommendation.
• Removal Adaptation: Completely eliminates PIE-aligned triples from the PKG. For example,
recipes like “Margherita Pizza,” which explicitly links Italian cuisine and tomato, are entirely
removed, forcing the recommender system to explore alternative items.</p>
        <p>PIE Characterization:
• Out-PIE: contains  given but not  bias; relevant to the user’s stated interest but breaks the learned
over-personalization. Example: Cannoli (Italian, no tomato) [preferred outcome]
• In-PIE: contains both  given and  bias; reinforces the over-personalized association the user is
trying to avoid. Example: Margherita pizza (Italian, tomato)
• Invalid: does not contain  given; fails to satisfy the user’s original intent or query (e.g.,
recommending a non-Italian dish when the user asked for Italian). Example: Ratatouille (French,
tomato)
Tuning the PKG Adaptation Proportion: To control the extent of symbolic intervention, we
introduce a user-specific parameter called adaptProportion, which determines the fraction of
PIEaligned triples in the PKG that should be adapted before inference. A low adaptProportion results in
minimal intervention, while a high adaptProportion aggressively steers the recommendation away
from the PIE.</p>
        <p>We learn a personalized adaptProportion for each user using a feedback-driven tuning
algorithm, which simulates PIE-avoidance scenarios using synthetic data points, and incrementally adjusts
adaptProportion based on their outcomes. If the adapted PKG still yields In-PIE recommendations,
the proportion is increased. If it produces Invalid results, the proportion is decreased. Successful
Out-PIE results leave the parameter unchanged. This iterative procedure converges on a personalized
adaptation strength tailored to each user’s PKG structure and feature biases.</p>
      </sec>
      <sec id="sec-1-4">
        <title>2.4. Model Fine-Tuning</title>
        <p>We fine-tune Qwen3-0.6B [ 10] using Hugging Face’s KTOTrainer [11], which implements
KahnemanTversky Optimization (KTO) [12]. Each training data point consists of a (prompt, completion, label)
triplet: the prompt includes user context and a query, the completion is a potential response to the
prompt, and the label is a binary indicator of the completion’s quality (positive or negative).</p>
        <p>Training data is derived from a customized version of the Food.com Recipes and Interactions dataset [13],
which includes user ratings (on a 0–5 star scale), ingredients, and categorical tags for recipes. From this
corpus, we construct the PKG capturing individual user preferences as a set of rated recipes.</p>
      </sec>
    </sec>
    <sec id="sec-2">
      <title>3. Experimental Results</title>
      <p>We evaluate our PIE avoidance
framework using PKGs derived
from the Food.com Recipes and In- values are highlighted in red .
teractions dataset [13]. We ran- Strategy Out-PIE ↑ In-PIE ↓ Invalid ↓
domly select 20 user PKGs and, for
each, sample 50 PIE-inducing fea- Soft (personalized) 0.3237 0.2158 0.4604
ture pairs. These are split 80/20 Soft (global) 0.2517 0.2583 0.4901
into 40 training and 10 evalua- Hard (personalized) 0.2848 0.2152 0.5000
tion PIEs per user. The training Hard (global) 0.2768 0.1977 0.5254
set is used to learn a personalized Removal (personalized) 0.3020 0.2416 0.4564
adaptProportion for each user via Removal (global) 0.3277 0.2203 0.4520
the tuning algorithm in Section 2.3.</p>
      <p>For comparison, we also compute Prompt-Based Adaptation 0.1925 0.1863 0.6211
a single global adaptProportion No Adaptation 0.2517 0.2583 0.4901
across all 800 training PIEs. A
learning rate of 0.05 is used in both cases.</p>
      <p>During evaluation, we generate 10 test queries per user (200 total) and categorize each
modelgenerated recommendation into Out-PIE, In-PIE, or Invalid classes. We aggregate and normalize
the counts to obtain proportions summing to one.</p>
      <p>
        We benchmark three PKG adaptation strategies (soft, hard, removal), each with both personalized
and global adaptProportion settings, against two baselines: (
        <xref ref-type="bibr" rid="ref1">1</xref>
        ) a prompt-based method using natural
language instructions to avoid PIEs, and (
        <xref ref-type="bibr" rid="ref2">2</xref>
        ) a no-adaptation baseline.
      </p>
      <p>As shown in Table 1, soft adaptation with personalized tuning achieves the best overall performance.
It increases Out-PIE recommendations from 25.2% (global adaptProportion baseline) to 32.4%, while
simultaneously reducing the Invalid rate from 49.0% to 46.0%, avoiding over-personalization without
decreasing recommendation quality. In contrast, the prompt-based approach, which attempts PIE
avoidance via plain-text instructions, performs poorly, with the lowest Out-PIE rate (19.3%) and the
highest Invalid rate (62.1%). This underscores the limitations of relying solely on natural language
prompting and demonstrates the efectiveness of symbolic PKG adaptations.</p>
    </sec>
    <sec id="sec-3">
      <title>4. Conclusion</title>
      <p>We introduced a novel, neuro-symbolic framework for enhancing personalization and user agency in
recommender systems by adapting PKGs rather than modifying model internals. Unlike conventional
approaches that require model retraining or rely on brittle heuristics, our method operates entirely at
the user-side knowledge representation layer, enabling eficient, interpretable, and privacy-preserving
control over recommendation behavior.</p>
      <p>
        Our key contributions include: (
        <xref ref-type="bibr" rid="ref1">1</xref>
        ) a formalization of PIEs as measurable feature-pair biases within
PKGs, (
        <xref ref-type="bibr" rid="ref2">2</xref>
        ) a suite of symbolic PKG adaptation strategies (Soft, Hard, and Removal) that steer LLMs toward
more diverse, Out-PIE content, and (
        <xref ref-type="bibr" rid="ref3">3</xref>
        ) a client-side learning algorithm for optimizing user-specific
adaptation policies. Through an evaluation on a real-world recipe dataset, we show that personalized
PKG adaptation consistently outperforms both global adaptation and natural-language prompting in
reducing over-personalization without compromising recommendation quality.
      </p>
      <p>These findings point toward a broader paradigm shift: from tuning black-box models to shaping the
symbolic structures that guide them. Our work demonstrates that adapting structured user knowledge
ofers a powerful, generalizable mechanism for embedding user intent and improving controllability in
LLM-powered systems, laying the groundwork for safer, more personalized, and user-aligned AI for
diverse recommendations.</p>
    </sec>
    <sec id="sec-4">
      <title>Declaration on Generative AI</title>
      <p>During the preparation of this work, the authors used ChatGPT, Gemini and Grammarly in order to
rephrase some of the sentences and also to fix grammar and spelling issues. After using these tools
and services, the authors reviewed and edited the content as needed and take full responsibility for the
publication’s content.</p>
    </sec>
    <sec id="sec-5">
      <title>Supplemental Materials</title>
      <p>All research artifacts, including source code, dataset construction scripts, and result generation pipelines,
are available in our GitHub repository. All external datasets and software dependencies used in this
work are documented and linked in the repository’s README.
https://github.com/brains-group/KGAdaptation.
[6] Y. Zhang, X. Chen, et al., Explainable recommendation: A survey and new perspectives,
Foundations and Trends® in Information Retrieval 14 (2020) 1–101.
[7] T. Donkers, J. Ziegler, The dual echo chamber: Modeling social media polarization for
interventional recommending, in: Proceedings of the 15th ACM conference on recommender systems,
2021, pp. 12–22.
[8] V. Anand, M. Yang, Z. Zhao, Mitigating filter bubbles within deep recommender systems, arXiv
preprint arXiv:2209.08180 (2022).
[9] F. Spadea, O. Seneviratne, Bursting the Filter Bubble with Knowledge Graph Inversion, in:
Companion Proceedings of the ACM on Web Science Conference 2025, Association for Computing
Machinery, New York, NY, USA, 2025, pp. 39–43. URL: https://doi.org/10.1145/3720554.3736182.
doi:10.1145/3720554.3736182.
[10] Hugging Face, Qwen3-0.6b, https://huggingface.co/Qwen/Qwen3-0.6B, 2024.
[11] Hugging Face, KTOTrainer,
https://huggingface.co/docs/trl/main/en/kto_trainer#trl.KTO</p>
      <p>Trainer.ref_model, 2024.
[12] K. Ethayarajh, W. Xu, N. Muennighof, D. Jurafsky, D. Kiela, Kto: Model alignment as prospect
theoretic optimization, 2024.
[13] S. Li, Food.com recipes and interactions, 2019. URL: https://www.kaggle.com/dsv/783630. doi:10.
34740/KAGGLE/DSV/783630.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Xi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Weng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Yi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Chen</surname>
          </string-name>
          , G. Guo,
          <string-name>
            <given-names>M.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , J. Wu,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Jiang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Liu</surname>
          </string-name>
          , et al.,
          <article-title>Bursting iflter bubble: Enhancing serendipity recommendations with aligned large language models</article-title>
          ,
          <source>arXiv preprint arXiv:2502.13539</source>
          (
          <year>2025</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <surname>E. Pariser,</surname>
          </string-name>
          <article-title>The filter bubble: What the Internet is hiding from you</article-title>
          ,
          <source>penguin UK</source>
          ,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>J.</given-names>
            <surname>Burrell</surname>
          </string-name>
          ,
          <article-title>How the machine 'thinks': Understanding opacity in machine learning algorithms</article-title>
          ,
          <source>Big data &amp; society 3</source>
          (
          <year>2016</year>
          )
          <fpage>2053951715622512</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>T. T.</given-names>
            <surname>Nguyen</surname>
          </string-name>
          ,
          <string-name>
            <surname>P.-M. Hui</surname>
            ,
            <given-names>F. M.</given-names>
          </string-name>
          <string-name>
            <surname>Harper</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <string-name>
            <surname>Terveen</surname>
            ,
            <given-names>J. A.</given-names>
          </string-name>
          <string-name>
            <surname>Konstan</surname>
          </string-name>
          ,
          <article-title>Exploring the filter bubble: the efect of using recommender systems on content diversity</article-title>
          ,
          <source>in: Proceedings of the 23rd international conference on World wide web</source>
          ,
          <year>2014</year>
          , pp.
          <fpage>677</fpage>
          -
          <lpage>686</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>S.</given-names>
            <surname>Milano</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Taddeo</surname>
          </string-name>
          , L. Floridi,
          <article-title>Recommender systems and their ethical challenges</article-title>
          ,
          <source>Ai &amp; Society</source>
          <volume>35</volume>
          (
          <year>2020</year>
          )
          <fpage>957</fpage>
          -
          <lpage>967</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>