<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>SLM-as-a-Judge with Attention Steering for Detailed Topic Extraction from Academic Literature</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Takahiro Kawamura</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Junichiro Mori</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Information Technology Center, The University of Tokyo</institution>
        </aff>
      </contrib-group>
      <abstract>
        <p>This study introduces domain-specific Small Language Models (SLMs) designed to fact-check the outputs of general-purpose Large Language Models (LLMs). The goal is to accurately and automatically extract detailed technical elements from academic papers to build knowledge graphs. Two SLM types were developed: one through continued pre-training and the other using attention steering. Experiments on information extraction in 'Fake News Detection' showed that SLMs improved the sufficiency, accuracy, and stability of extracted information. Future work will involve larger datasets and further knowledge graph construction.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Knowledge Graph</kwd>
        <kwd>LLM</kwd>
        <kwd>Information Extraction</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
    </sec>
    <sec id="sec-2">
      <title>2. Related Work</title>
      <p>
        Science Knowledge Graph. Recently, there has been growing interest in building knowledge
graphs to structure scholarly information. Most existing knowledge graphs focus on
bibliographic data such as publications, authors, affiliations, research topics and their relationships
(citations, authorship, collaborations, topic similarity). Notable initiatives include OpenAlex1,
as well as projects like Semantic Scholar2, AMiner3, and OpenCitations4. These graphs
typically define entities such as papers, authors, institutions, research fields, and keywords as nodes,
with relationships like authorship and citations as edges. Some newer approaches [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] use
systems like Babelfy to identify topics and link them to BabelNet entities, constructing graphs
from co-occurrence, while others [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] utilize AI-in-the-loop methods (e.g., DeepShovel) to build
topic trees and analyze the influence among papers to support idea flow analysis and research
forecasting. Recent trends [
        <xref ref-type="bibr" rid="ref4 ref5 ref6 ref7">4, 5, 6, 7</xref>
        ] also include the extraction of semantic relationships
between technical elements (e.g., research objectives, methods, evaluation metrics) and
representing them in graph-based formats allowing for a richer understanding of research content
compared to simply listing topics. Techniques such as manual annotation and natural language
processing tools are used to extract objectives, methods, results, and their interrelations.
      </p>
      <p>However, these existing approaches face limitations: many employ NLP techniques that are
not necessarily up to date, cover only a small number of papers, are infrequently updated, or
are not publicly available.</p>
      <p>
        LLM for academic information. The development of LLMs tailored for scholarly information
is active [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ], with examples such as Meta’s Galactica, domain-specific models like BioMedLM,
and language-specific models (e.g., Chinese and English) such as ChatGLM3, all of which have
been trained on large scientific datasets. A variety of services, including SciSpace, Elicit, and
Consensus, use LLMs to summarize papers, extract key points, and facilitate literature searches,
though not all support knowledge graph construction or rely directly on LLMs. However, these
LLMs and services were not used in the present research due to diferences in purpose, unresolved
general LLM issues, and a lack of modifiability.
      </p>
      <p>
        Instead, this research applies an underexplored method: attention steering, which involves
focusing or suppressing attention on specific words during LLM inference to guide output. While
similar techniques such as PASTA [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] have shown theoretical and basic task improvements,
realworld applications remain rare. Recent studies [
        <xref ref-type="bibr" rid="ref10 ref9">9, 10</xref>
        ] have explored manipulating attention
and activations within LLMs for better content control and long-form processing. This study
uniquely applies and evaluates attention steering in developing a specialized SLM for academic
information extraction.
      </p>
    </sec>
    <sec id="sec-3">
      <title>3. Proposed Approach</title>
      <p>This project aims to build a detailed science graph where research objectives, methods, and
subjects are nodes. As an example, extraction should capture the main theme, data (e.g.,
Slovak language), techniques (deep learning), specific methods (like CNN or LSTM), and even
implementation details (such as ReLU or TensorFlow). So, we developed a model with continual
pre-training on technical domain data and another using attention steering to focus on technical
terms during inference. These outputs were compared to those from a general LLM to check
for coverage and accuracy. For attention steering, attention weights in the transformer model
were manually biased toward target tokens, intentionally guiding output content. Implemented
in frameworks like PyTorch, this involved adjusting attention scores via log-ratio biases during
the forward pass as below. Although these methods slow inference, real-time responses are not
required for this study.</p>
      <p>= + log(ratio) 1=</p>
      <p>exp()
=</p>
      <p>exp()
: the position of the specific token to which the bias is to be applied. 1= : an indicator function
that returns 1 if equals , and 0 otherwise.</p>
      <sec id="sec-3-1">
        <title>3.1. Experimental Setting</title>
        <p>For continuous pre-training, we collected 48 survey papers on Fake News Detection (FND),
totaling 4.32 million characters, all of which were used for training. The base model was
llama3-8B-Instruct, trained with a learning rate of 5e-6, batch size 2048 tokens, and 1.5 epochs using 8
batches 4 GPUs. Instead of heavy pre-training, we also steered attention toward specific tokens
identified by spaCy’s NER tool (such as PERSON, LOC, ORG, GPE, PRODUCT, NORP,
WORK_OF_ART, LANGUAGE, DATE, TIME, PERCENT, QUANTITY, ORDINAL, and
CARDINAL) by adding a bias to their attention scores across all layers. Full-parameter training
was used rather than LoRA, since we aimed to apply a moderate bias across all layers. Technical
element extraction experiments were done in two ways:
1. Prompting with a paper title and having the model respond based on its internal
knowledge about the paper’s topics. It has been verified that the paper was included in the
training data of the model.
2. Prompting with a 1,000–2,000 character excerpt, selected to contain around 10 topics, and
asking the model to extract the topics appearing in the given text.</p>
        <p>Four models were compared: continuous pre-training (CP), CP plus steered attention (SA),
original llama-3-8B-Instruct, and GPT4.1. Evaluation for (1) knowledge-based extraction used
precision, recall, and F1-score, while (2) prompt-based extraction used Normalized Discounted
Cumulative Gain (NDCG) to check the order of the extracted topics. The ground-truth data was
manually created by project members specializing in computer science, who visually extracted
information from 10 papers on FND. The average number of topics to be extracted per paper
was 41.3. Temperature was set to 1.0.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Experimental Results</title>
      <sec id="sec-4-1">
        <title>4.1. Extraction from Internal Knowledge</title>
        <p>In Table 1, both CP and SA achieved higher F1 scores compared to GPT and the base llama
model, primarily because they extracted a greater number of technical elements, thus increasing
recall. Notably, the SA model also attained the highest precision, demonstrating that the
intended efect of attention steering was realized. Although GPT had high precision, its recall
was low likely due to more conservative outputs, resulting in the lowest F1 score. The number of
elements to be extracted was not specified, since the appropriate number of technical elements
difers for each paper and is unknown in practical use. While CP had the highest F1 score overall,
its lower precision and larger standard deviation suggested instability. In direct comparison, CP
and SA each outperformed the other in five out of ten cases, but SA’s marginally lower F1 score
was ofset by its higher precision and greater stability, making it the more desirable model.
However, both CP and SA often generated structurally awkward sentences and substantially
underperformed GPT in language fluency. As a result, it is recommended that SA be used in
tandem with GPT, relying on GPT for language generation tasks.</p>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. Extraction from Prompts</title>
        <p>Since most LLMs cannot process the full text of a paper in a single prompt (with the exception
of GPT-4.1), the experiment extracted technical elements using only 1,000–2,000 character
excerpts as prompts. This approach matches the current ’Deep Research’ trend, where papers
are segmented and summarized before being fed to LLMs, and also reflects practices used in
Retrieval-Augmented Generation (RAG) systems. Extraction accuracy for technical elements
from such excerpt lengths was nearly 100% across all models, leading to minimal performance
diference among them. Consequently, NDCG@5 was adopted to evaluate the ranking order of
the extracted elements, with relevance scores manually assigned to the ground truths. As shown
in Table 1, the SA model produced the best results, with low variance across runs. In contrast,
the CP model sometimes hallucinated by outputting information not present in the excerpt.
For the SA model, the attention bias was set to a very small value (log(1 + 1 16)), unlike the
larger bias (log(1.5)) used in prior experiments, because a larger bias degrades linguistic quality
when input tokens are short.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Discussion and Future Work</title>
      <p>In this paper, the authors proposed a fact-checking Small Language Model (SLM) designed
specifically for extracting domain-specific technical elements from academic literature, with the
ultimate aim of constructing a detailed science graph. Consequently, GPT rarely makes outright
errors in responses to single queries, but its answers are often overly simple and lack sufficient
detail. Repeating questions can draw out more information, but this eventually leads to
hallucinations (speculative or inaccurate content) which makes it hard to distinguish between factual
and fabricated information. Continuous pre-training enables models to produce more detailed
responses, particularly regarding numerical data (historically difficult for LLMs), yet it also
introduces greater variance between experimental runs and sometimes produces seriously flawed
outputs. Such models also tend to have reduced comprehension of prompts and decreased
fluency in generated language. This is consistent with the well-known phenomenon that acquiring
highly specialized knowledge can diminish a model’s general language abilities and performance
on unrelated tasks. Attention steering, when used appropriately, may enhance the detail and
stability of the information generated. However, it is important to note that these findings are
based on experiments conducted with a relatively small dataset, and the subjectivity of the
ground-truth data remains a concern.</p>
      <p>For future work, the authors’ top priority is to expand their dataset, given that no
goldstandard dataset fully satisfies their requirements. They plan to validate their approach on
a larger scale using general-purpose technical term extraction datasets. It is also essential to
define the scope and size of the target domain — for instance, while current experiments focused
only on FND within Computer Science, expanding the domain is expected to make the model’s
behavior more similar to general-purpose LLMs. This makes it important to carefully define
the domain in line with the intended application. From an algorithmic perspective, unresolved
challenges include developing techniques for introducing bias into the attention mechanism,
quantitatively assessing how varying the degree of bias afects output, and evaluating potential
thresholds as well as the risk of model collapse from excessive bias. The authors plan to use
the extracted information to build science graphs and proceed with assessments in domains
where analysis is demonstrably needed. Toward the goal of science and technology analysis,
we are working with other major Japanese research institutions on science graph sharing and
joint analysis, and also plan to share their methodologies and data internationally. We hope to
promote the development of new, internationally recognized science and technology assessment
methods by collaborating with reputable overseas institutions.</p>
    </sec>
    <sec id="sec-6">
      <title>6. Declaration on Generative AI</title>
      <p>During the preparation of this work, the authors used ChatGPT in order to: Text Translation.
After using this service, the authors reviewed and edited the content as needed and take full
responsibility for the publication’s content.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>X.</given-names>
            <surname>Gu</surname>
          </string-name>
          ,
          <string-name>
            <surname>M.</surname>
          </string-name>
          <article-title>Krenn, Forecasting high-impact research topics via machine learning on evolving knowledge graphs</article-title>
          ,
          <source>Machine Learning: Science and Technology</source>
          <volume>6</volume>
          (
          <year>2025</year>
          ). doi:
          <volume>10</volume>
          .1088/
          <fpage>2632</fpage>
          -2153/add6ef.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <surname>M. D. L. Tosi</surname>
            ,
            <given-names>J. C.</given-names>
          </string-name>
          dos
          <string-name>
            <surname>Reis</surname>
          </string-name>
          ,
          <article-title>Scikgraph: A knowledge graph approach to structure a scientific field</article-title>
          ,
          <source>Journal of Informetrics</source>
          <volume>15</volume>
          (
          <year>2021</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>X.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Fu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Gan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Wen</surname>
          </string-name>
          , G. Zheng,
          <string-name>
            <given-names>J.</given-names>
            <surname>Ding</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Xiang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Ye</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Jin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Liang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Lu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Deng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Kang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Guo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Qi</surname>
          </string-name>
          , P. Liu,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Ren</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <article-title>Acemap: Knowledge discovery through academic graph</article-title>
          ,
          <year>2024</year>
          . doi:
          <volume>10</volume>
          .48550/arXiv.2403.02576.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>S.</given-names>
            <surname>Fathalla</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Vahdati</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Auer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Lange</surname>
          </string-name>
          ,
          <article-title>Towards a knowledge graph representing research findings by semantifying survey articles</article-title>
          „
          <source>in: International Conference on Theory and Practice of Digital Libraries</source>
          ,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>D.</given-names>
            <surname>Dessì</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Osborne</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. R.</given-names>
            <surname>Recupero</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Buscaldi</surname>
          </string-name>
          , E. Motta,
          <article-title>Generating knowledge graphs by employing natural language processing and machine learning techniques within the scholarly domain</article-title>
          ,
          <source>Future Generation Computer Systems</source>
          <volume>116</volume>
          (
          <year>2021</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>S.</given-names>
            <surname>Auer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Ilangovan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Stocker</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Tiwari</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Vogt</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Hussein</surname>
          </string-name>
          , Open Research Knowledge Graph,
          <year>2024</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>D.</given-names>
            <surname>Dessí</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Osborne</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Buscaldi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. R.</given-names>
            <surname>Recupero</surname>
          </string-name>
          , E. Motta,
          <article-title>Cs-kg 2.0: A large-scale knowledge graph of computer science</article-title>
          ,
          <source>Sci Data 12</source>
          (
          <year>2025</year>
          ).
          <source>doi:10.1038/ s41597-025-05200-8.</source>
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>Q.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Singh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Yu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Gao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Zhao</surname>
          </string-name>
          ,
          <article-title>Tell your model where to attend: Post-hoc attention steering for llms</article-title>
          ,
          <source>in: 12th International Conference on Learning Representations</source>
          ,
          <year>2024</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>W.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Peng</surname>
          </string-name>
          ,
          <article-title>Semantics-adaptive activation intervention for llms via dynamic steering vectors</article-title>
          ,
          <source>in: 13th International Conference on Learning Representations (poster)</source>
          ,
          <year>2025</year>
          . doi:
          <volume>10</volume>
          .48550/arXiv.2410.12299.
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Gu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Yao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Du</surname>
          </string-name>
          ,
          <article-title>Llmsteer: Improving long-context llm inference by steering attention on reused contexts</article-title>
          ,
          <source>in: Machine Learning for Systems Workshop at the 38th Annual Conference on Neural Information ProcessingSystems</source>
          ,
          <year>2024</year>
          . doi:
          <volume>10</volume>
          .48550/arXiv.2411. 13009.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>