<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <issn pub-type="ppub">1613-0073</issn>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Hierarchical Knowledge Editing for LLMs: Instance-to-Concept Relationship Perspective</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Zhaoyuan Zhang</string-name>
          <email>zhaoyuanzhang@tju.edu.cn</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Tao Luo</string-name>
          <email>luo_tao@tju.edu.cn</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Xiaowang Zhang</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Sai Zhang</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="editor">
          <string-name>Knowledge Editing, Large Language Model, Hierarchical Knowledge, Layered Distillation</string-name>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>College of Intelligence and Computing, Tianjin University</institution>
          ,
          <addr-line>Tianjin, 300350</addr-line>
          ,
          <country country="CN">China</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Knowledge editing has emerged as a promising way to update knowledge in large-scale language models (LLMs) eficiently. However, current knowledge editing methods focus on undiferentiated factual knowledge, neglecting the significance of hierarchical structured knowledge editing. Moreover, cognitive science has revealed the importance of hierarchical knowledge for human learning. This poster introduces hierarchical knowledge editing for instance-to-concept relationship. Through Layered Distillation strategy, we perform knowledge distillation between the original and edited models, thereby preserving instance-to-concept relationship hierarchical knowledge in the original model. Experimental results demonstrate that integrating our strategy with existing knowledge editing methods enhances the performance of hierarchical knowledge editing.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>CEUR
ceur-ws.org</p>
    </sec>
    <sec id="sec-2">
      <title>1. Introduction</title>
      <p>
        With the performance improvement and wide application of large language models, the
problems of LLMs, such as providing outdated, erroneous, or toxic information, have become the
focus of criticism. Retraining LLMs to address these issues takes time and exertion. In contrast,
knowledge editing ofers a low-cost way to update trained models. This has made the
development of eficient and reliable knowledge editing methods for LLMs a key area of research[
        <xref ref-type="bibr" rid="ref1">1</xref>
        ].
However, most existing LLM knowledge editing methods do not diferentiate between types
of knowledge and primarily concentrate on editing factual knowledge individually, which is
ineficient for LLMs with a massive number of parameters that store vast amounts of knowledge.
Moreover, there is a lack of efective retention of structured hierarchical information within
LLMs. Based on this, this poster proposes hierarchical knowledge editing.
      </p>
      <p>
        LLMs memorize various hierarchical knowledge, hierarchical knowledge editing task
includes editing both instance-to-concept and other inter-conceptual relationships hierarchical
knowledge. Focusing on instance-to-concept relationship, we want to retain the integrity of this
relationship in the original model after editing concept. Consider a simple instance-to-concept
relationship hierarchical knowledge in LLMs: instance “tiger” belongs to concept “feline,” which
is defined as a carnivorous mammal known for flexible body and sharp claws. In the case where
the definition of “feline” is edited to winged animal, since “tiger” is wingless, human cognition
will naturally assume that “tiger” no longer belongs to “feline,” which means that we do not want
this editing to modify the hierarchical knowledge of “tiger” belongs to “feline” in the original
LLMs[
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. Thus, it is vital to maintain the original instance-to-concept relationship hierarchical
knowledge when editing LLMs. While a recent work proposes editing conceptual knowledge,
which focuses on the efects of modifying concept definitions within LLMs[
        <xref ref-type="bibr" rid="ref2">2</xref>
        ], hierarchical
knowledge editing concentrates on evolving editing approaches that preserve hierarchical
knowledge within LLMs.
      </p>
      <p>In this poster, we define hierarchical Knowledge Editing task with corresponding metric,
design Layered Distillation strategy for preserving instance-to-concept relationship hierarchical
knowledge, and present experimental evidence of its meaningful efectiveness in enhancing
hierarchical knowledge editing.</p>
    </sec>
    <sec id="sec-3">
      <title>2. Approach</title>
      <p>LLMs memorize various hierarchical knowledge, Hierarchical Knowledge Editing task includes
managing both instance-to-concept and other inter-conceptual relationships. Focusing on
instance-to-concept relationship, our goal is to design an editing approach that retains the
integrity of the relationship when concept definitions are updated.</p>
      <sec id="sec-3-1">
        <title>2.1. Task Definition</title>
        <p>Hierarchical knowledge editing task is formally defined as: given a concept  = (, ) , where
 is the concept name and  is the concept definition, and a set of instances  , all instances 
in  belong to concept  , denoted  ∈  . When the definition  is edited to  ∗, resulting in the
modified concept  ∗ = (,  ∗), a great hierarchical knowledge editing approach must ensure
that the modification of concept does not result in  ∈  ∗, maintain the integrity of  , minimize
instance migration, and avoid unacceptable changes to the original model.</p>
      </sec>
      <sec id="sec-3-2">
        <title>2.2. Knowledge Editing with Layered Distillation</title>
        <p>
          We introduce knowledge distillation to ensure the edited model inherits the original model’s
hierarchical knowledge[
          <xref ref-type="bibr" rid="ref3">3</xref>
          ]. We first use Location-Then-Edit as the base editing method[
          <xref ref-type="bibr" rid="ref1 ref4 ref5">1, 4, 5</xref>
          ].
This method targets specific model layers for editing, enabling focused distillation of the original
and edited model layers, rather than the entire model. Layered Distillation is updating the
model layers again using these original and edited model layers as input, with Mean Square
Error as the distillation loss. The specific process is shown in Figure 1. The base method edits
the concept definition in the original model, resulting in instance migration. Updating the
edited model again through the Layered Distillation strategy retains the instance-to-concept
relationship in the original model.
        </p>
      </sec>
      <sec id="sec-3-3">
        <title>2.3. Metrics</title>
        <p>To more efectively assess the impact of our editing approach on hierarchical knowledge, we
devise Instance Retention(IR) metric. Its definition is as follows:</p>
        <p>1
 ∈ 
  = 1 −
∑[(,  ∗) + | (,  ) −  (,  ∗)|]
(1)</p>
        <p>
          Instance Retention measures the proportion of instance-to-concept relationship knowledge
preserved in the edited model. It is determined by two functions: (,  ), which indicates
whether instance  belongs to concept  ( = 1 if it does,  = 0 if it does not).  (,  ), used
when instance  cannot determine whether it belongs to concept  ( = 1 in this case,  = 0
otherwise). These functions are applied using the reasoning capability of the LLM[
          <xref ref-type="bibr" rid="ref2">2</xref>
          ]. We also
use Reliability(Rel.), Generalization(Gen.), and Locality(Loc.) as comprehensive metrics to
evaluate the success, scope, and impact of knowledge editing[
          <xref ref-type="bibr" rid="ref1">1</xref>
          ].
        </p>
        <p>Main results of the Hierarchical Knowledge Editing experiment. “+LD” stands for combined with our
strategy. Bold results denote optimal performance.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>3. Experiments</title>
      <p>
        Utilizing the ConceptEdit[
        <xref ref-type="bibr" rid="ref2">2</xref>
        ], derived from the DBpedia ontology dataset, we conducted
experiments on the open-source LLMs GPT2-XL (1.5B) [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]and TinyLlama (1.1B)[
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] using the
ROME[
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] and MEMIT[
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] methods combined with Layered Distillation(LD) on GeForce RTX
4090. ConceptEdit’s Intra and Inter modules represent modifications to concepts within and
between superclasses.
      </p>
      <p>Table 1 results indicate that both editing methods struggle with hierarchical knowledge
editing in both models, but the IR of ROME method is significantly higher. Meanwhile, Layered
Distillation strategy with these methods enhances IR without notably altering other metrics,
particularly for ROME editing TinyLlama. This enhancement validates the eficacy of our
strategy for editing instance-to-concept relationship hierarchical knowledge. Furthermore, the
relatively modest enhancement observed in the MEMIT+LD configuration can be attributed to
the editing across multiple MLP layers, which is responsible for the suboptimal performance of
the GPT2-XL model, but concurrently ofers a higher degree of Locality.</p>
    </sec>
    <sec id="sec-5">
      <title>4. Conclusion and Future Work</title>
      <p>This poster proposes Hierarchical Knowledge Editing for LLMs and Layered Distillation strategy
to enhance existing knowledge editing methods for editing instance-to-concept relationship
hierarchical knowledge. Our experiments initially demonstrate the eficacy of Layered
Distillation. Future research will broaden validation to additional methods and larger LLMs. Editing
other hierarchical relationship knowledge will also be further explored.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>N.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Yao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Tian</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Deng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Xi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Mao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Ni</surname>
          </string-name>
          , et al.,
          <article-title>A comprehensive study of knowledge editing for large language models</article-title>
          ,
          <source>arXiv preprint arXiv:2401.01286</source>
          (
          <year>2024</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>X.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Mao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , S. Deng,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Yao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Shen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Liang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Gu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <article-title>Editing conceptual knowledge for large language models</article-title>
          ,
          <year>2024</year>
          . arXiv:
          <volume>2403</volume>
          .
          <fpage>06259</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>J.</given-names>
            <surname>Gou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Yu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. J.</given-names>
            <surname>Maybank</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Tao</surname>
          </string-name>
          ,
          <article-title>Knowledge distillation: A survey</article-title>
          ,
          <source>International Journal of Computer Vision</source>
          <volume>129</volume>
          (
          <year>2021</year>
          )
          <fpage>1789</fpage>
          -
          <lpage>1819</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>K.</given-names>
            <surname>Meng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Bau</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Andonian</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Belinkov</surname>
          </string-name>
          ,
          <article-title>Locating and editing factual associations in gpt</article-title>
          ,
          <source>in: Proceedings of the 36th International Conference on Neural Information Processing Systems</source>
          ,
          <year>2022</year>
          , pp.
          <fpage>17359</fpage>
          -
          <lpage>17372</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>K.</given-names>
            <surname>Meng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. S.</given-names>
            <surname>Sharma</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. J.</given-names>
            <surname>Andonian</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Belinkov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Bau</surname>
          </string-name>
          ,
          <article-title>Mass-editing memory in a transformer</article-title>
          ,
          <source>in: The Eleventh International Conference on Learning Representations, ICLR</source>
          <year>2023</year>
          , Kigali, Rwanda, May 1-
          <issue>5</issue>
          ,
          <year>2023</year>
          , OpenReview.net,
          <year>2023</year>
          . URL: https://openreview.net/ pdf?id=MkbcAHIYgyS.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>A.</given-names>
            <surname>Radford</surname>
          </string-name>
          ,
          <string-name>
            <surname>J. Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Child</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Luan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Amodei</surname>
          </string-name>
          ,
          <string-name>
            <given-names>I.</given-names>
            <surname>Sutskever</surname>
          </string-name>
          , et al.,
          <article-title>Language models are unsupervised multitask learners</article-title>
          ,
          <source>OpenAI blog 1</source>
          (
          <year>2019</year>
          )
          <article-title>9</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>P.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , G. Zeng,
          <string-name>
            <given-names>T.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Lu</surname>
          </string-name>
          ,
          <string-name>
            <surname>Tinyllama:</surname>
          </string-name>
          <article-title>An open-source small language model</article-title>
          ,
          <source>arXiv preprint arXiv:2401.02385</source>
          (
          <year>2024</year>
          ).
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>