<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>URL:
http://arxiv.org/abs/</journal-title>
      </journal-title-group>
      <issn pub-type="ppub">1613-0073</issn>
    </journal-meta>
    <article-meta>
      <article-id pub-id-type="doi">10.48550/arXiv.2501.12948</article-id>
      <title-group>
        <article-title>Extracting problem-solving knowledge from LLMs with reasoning abilities</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Maxime Haurel</string-name>
          <email>maxime.haurel@loria.fr</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Workshop</string-name>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Datanello</institution>
          ,
          <addr-line>Nancy</addr-line>
          ,
          <country country="FR">France</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Université de Lorraine</institution>
          ,
          <addr-line>CNRS, LORIA, F-54000 Nancy</addr-line>
          ,
          <country country="FR">France</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2501</year>
      </pub-date>
      <volume>12948</volume>
      <issue>300</issue>
      <fpage>2</fpage>
      <lpage>6</lpage>
      <abstract>
        <p>Automatic Knowledge Acquisition (AKA) aims to automate the process between domain experts and knowledge engineers, that is to create a domain Knowledge Base (KB). Such automation is necessitated because the collaboration between domain experts and knowledge engineers is costly. Modern approaches use Large Language Models (LLMs) simulating domain experts to create ontologies and knowledge graphs. Recently, LLMs with reasoning abilities received attention due to their great performances on several benchmarks. These LLMs output reasoning traces that lead to the answer. This early stage PhD thesis (started 7 months ago) focuses on the use of those reasoning traces in the automatic construction of a Knowledge Base (KB), under the assumption that they express the knowledge necessary to solve the problem prompted to the LLM. To achieve this, a first step consists in obtaining from the LLM reasoning traces that are expressed in a well-defined formalism. The results of our initial experiments show that, while some models are able to generate reasoning traces backed by formally expressed knowledge, there is still room for improvement. The remainder of this PhD will therefore involve improving LLMs with reasoning abilities to increase their capability to express the knowledge they used for problem solving in a way that is exploitable for AKA, coherent, and valid.</p>
      </abstract>
      <kwd-group>
        <kwd>automatic knowledge acquisition</kwd>
        <kwd>large language models</kwd>
        <kwd>reasoning models</kwd>
        <kwd>prolog</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>CEUR</p>
      <p>ceur-ws.org
more interpretable). In the next sections, we outline several approaches to verify our hypotheses. In
the end of this document, we enumerate the next steps of this work.</p>
      <p>&lt;think&gt; Okay, so I’m trying to figure out the possible heart condition based on the symptoms
described. The user mentioned sudden weakness in the left arm and leg, which makes me think of
a stroke or possibly a TIA (transient ischemic attack). But then there’s also recent long-distance
travel and a swollen, tender right lower leg.</p>
      <p>First, the left-sided weakness points towards neurological issues, likely related to blood flow. A
stroke or TIA is a common cause for sudden weakness in one side of the body. However, the user
also provided other symptoms that I need to consider together.</p>
      <p>The long-distance travel makes me think about prolonged sitting or physical activity during the
trip. This could contribute to deep vein thrombosis (DVT), especially if there’s a history of blood
clots or risk factors like obesity, age over 40, or being sedentary for long periods.</p>
      <p>Now, the right lower leg symptoms—swollen and tender—are classic signs of DVT. DVT occurs
when blood clots form in the legs and can travel to the lungs (causing a pulmonary embolism) or
other organs. If someone experiences sudden weakness on one side along with DVT, it might be
related to an embolism blocking a blood vessel in the brain, leading to a stroke.</p>
      <p>So, putting this together, the person might have experienced a pulmonary embolism from DVT,
which then caused a stroke-like syndrome. This would explain both sets of symptoms: the left-sided
weakness from the stroke and the right leg issues from DVT. &lt;/think&gt;
&lt;answer&gt; The most likely cardiac abnormality is a pulmonary embolism secondary to deep vein
thrombosis (DVT). The recent long-distance travel increases the risk of DVT, which can lead to
a pulmonary embolism. This condition can cause sudden neurological symptoms like left-sided
weakness due to an embolism blocking a cerebral artery, mimicking a stroke or TIA.&lt;/answer&gt;</p>
    </sec>
    <sec id="sec-2">
      <title>2. Importance</title>
      <p>AKA is a crucial question in the artificial intelligence community focusing on knowledge-based systems.
Indeed, representing unstructured domain knowledge into structured information is a process that
generates both human and time costs. This bottleneck limits our ability to build KBs that are both
large-scale [7] and reliable [7]. We make the hypothesis that using LLMs to automatically acquire
knowledge and manage a KB’s lifecycle will reduce this issue.</p>
      <p>
        Our approach aims to automatically build a KB, using the reasoning traces produced by LLMs with
reasoning abilities as the expression of knowledge. We want to obtain a formalized version of the
reasoning traces in the objective of building a KB with a formalism that has proven to be eficient for
this task in the knowledge representation and reasoning community. In addition, LLMs with reasoning
abilities have recently emerged, in particular with DeepSeek-R1 in January 2025 [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. While these models
are able to show a form of explanation compared to LLMs without reasoning abilities, LLMs in general
(i.e. with or without reasoning abilities) are still operating as black boxes [8, 9, 10]. Our approach will
help us better understand these models, contributing to interpretability. Indeed, our approach could
explain a prediction through logical reasoning, or even validate/invalidate the results obtained based
on the coherence of the logical reasoning.
      </p>
      <p>Our research is motivated by a problem encountered in the HR domain. The issue is to identify
indicators of soft skills that are expressed in a text that summarizes the decisions made by an individual
in the role of a manager facing a challenging situation (e.g. dealing with an angry customer due to a
problem of shipping). This task is highly knowledge-intensive as it requires solid psychology-related
knowledge from an expert. Also, no labeled data is available for this task of predicting soft skills
indicators from a text. Moreover, we face the issue of the availability of domain experts qualified on
the subject. This is why we require a method that is both automatic and explainable. In addition, the
HR domain is a high-stakes domain as listed in the EU AI Act [11]. Specifically, the document states
that: “AI systems intended to be used to make decisions afecting terms of work-related relationships, the
promotion or termination of work-related contractual relationships, to allocate tasks based on individual
behaviour or personal traits or characteristics or to monitor and evaluate the performance and behaviour of
persons in such relationships” are classified as high risk. We then need a method that provides answers
given a complex problem (e.g. predicting soft skills indicators from a text) using expert knowledge
formally defined in a KB while providing explanations regarding the reasoning process used by the
model to reach the answer. Thus, building a problem-solving knowledge focused KB automatically with
an LLM to then use this KB to reason over complex problems is promising. While our objective is to
use LLMs with reasoning abilities as domain experts to manage a KB’s lifecycle, we are aware of the
ethical considerations [12] and we will deal with these considerations when developing the framework.
While other approaches exist, our approach is intended to reduce the barriers of the lack of domain
experts and the lack of annotated data. In the study described in this paper, we take examples from
the medical domain from a public dataset. This kind of knowledge requires solid reasoning to provide
reliable answers [13], making it close to the HR domain. Furthermore, this dataset has proven to be a
good support to exert reasoning from LLMs [13].</p>
    </sec>
    <sec id="sec-3">
      <title>3. Related work</title>
      <p>Our work is placed in AKA, but with a set of constraints. As said earlier, our objective is to capture
knowledge useful in the resolution of a specific (naturally knowledge-intensive) problem, such as
the prediction of soft skill indicators. Additionally, the knowledge that we need is not stored in any
document. We do not have access to large quantities of past cases of such problems solved, and the
solution is not automatically verifiable. Our interest in LLMs therefore comes from the assumption that,
as an efect of their training, they have captured both common sense and domain-specific knowledge,
which means that they could, at least partly, take the role of a domain expert. Explainability is a crucial
consideration, given the high level of risk stated by the EU AI act in the HR domain [11]. Relying on
explicitly formalized knowledge in a KB is therefore favored here. Hence, here, we review existing
work in AKA, especially with LLMs. We also investigate the literature to review the formalisms used
by the knowledge representation community.</p>
      <p>Automatic Knowledge Acquisition Several works have focused on building KBs automatically.
We identified two approaches in the AKA community: document (i.e. natural language) processing
approach and learning (i.e. statistical) approach. As an example in document processing, [14] proposes
to automatically construct a KB through document parsing, which therefore requires access to a corpus
of relevant documents. As an example in a learning approach, [15] compares an inductive learning
system to a naive bayesian learning system. The former generates knowledge as a decision tree, while
the latter generates a table of conditional and prior probabilities. Both systems (inductive learning and
naive bayesian learning) are based on a collection of concrete problems, solved by experts in the past.
Here too, the applicability of the method is limited to cases where a base of previously solved problems
is available, which is not the scenario on which we focus.</p>
      <p>Automatic Knowledge Acquisition with LLMs LLMs are trained on a vast corpus of data, making
them able to leverage a great amount of cross-domain knowledge. They even surpass humans in some
ifelds such as coding [ 16], generating stories [17], and answering genetic-related questions [18].</p>
      <p>
        The following related works rely on the idea that LLMs can be exploited as a source of knowledge,
similarly to domain experts. Recently, [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] proposed to use LLMs to semi-automatically (i.e. with a
human involved in the loop) construct an ontology and a Knowledge Graph (KG). Their approach relies
on the generation of competency questions (CQs) to create an ontology and the answering of CQs
to construct the KG. Other works have used LLMs to automatically construct a KG [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ], where LLMs
play the role of domain experts. Their solution is a framework made up of 3 levels: Generator, Verifier ,
and Pruner. For a specified entity (e.g. a domain name), Generator retrieves the most relevant triples
from an open-source encyclopedic KG. Then, Verifier identifies and filters out the erroneous triples
generated by the LLM at the previous level (i.e. Generator). Finally, Pruner, a binary classifier, predicts
for each tail entity (object) from each correct triple, whether the entity should proceed to generate
the next-level KG (i.e. continue the loop) or cease the generation. This work shows that LLMs can
produce better KGs than the previous state-of-the-art methods and makes a great step towards AKA
as no human is involved. Thus, it supports our research directions. However, as expressed above, our
objective concerns the acquisition of knowledge that is useful for the resolution of specific problems,
while most of the previous work on using LLMs for AKA has focused on ontologies and KGs, which are
more general and abstract.
      </p>
      <p>
        Automatic Knowledge Acquisition with reasoning abilities in LLMs Recent approaches utilize
RL to influence the behavior of an LLM. The Group Relative Policy Optimization (GRPO) algorithm [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]
provides an eficient way of training the LLM to maximize a reward with reinforcement learning.
This algorithm is the basis behind the DeepSeek-R1 model [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] that achieved great scores on several
benchmarks and opened the way to LLMs with reasoning abilities. The GRPO algorithm plays a key
role as it maximizes the reward designed by the DeepSeek team. This reward aims to influence the
behavior of the LLM so that the reasoning is outputted within &lt;think&gt;&lt;/think&gt; tags. Released in
April 2025, the Cogito LLM [19] has received attention thanks to great performance on benchmarks.
Since such models with reasoning abilities are quite recent, few works in the literature used them for
AKA. [
        <xref ref-type="bibr" rid="ref4">4, 20</xref>
        ] show the use of LLMs with reasoning abilities to generate ontologies. They represent the
knowledge in the OWL language after giving CQs as input to the LLM.
      </p>
      <p>Knowledge Representation and Reasoning While many knowledge representation formalisms
exist, Prolog is a good candidate for modeling symbolic representations of reasoning problems [21,
22, 23, 24, 25, 26, 27]. For example, [26] improves the logical reasoning of LLMs through a framework
notably including a Translator module that aims to parse given premises and a question statement with
an LLM into a symbolic format adopting Prolog’s grammar. Another approach [27] uses Vadalog [28],
a language derived from Datalog [29]. In the study, the Vadalog language is used to describe the
knowledge graph and act as an intermediate layer between the natural language question formulated
by the user and the actual knowledge represented as a graph.</p>
      <p>In the scientific literature, Prolog has been widely used. For the application we are tackling throughout
(i.e. human resources) this thesis, we believe that Prolog is the best candidate to represent knowledge for
solving problems. In addition, Prolog is a logical formalism that can be expressed as text, which is crucial
for asking an LLM to output it. Also, Prolog being widely represented among logical formalisms, we
think that an LRM will be able to output it properly. While Prolog has many advantages, a limitation is
that representing probabilities is not possible. To overcome this limitation and if we need a probabilistic
representation of knowledge, we will consider using ProbLog [30], an extension of Prolog. For a Prolog
clause, ProbLog allows to add probabilistic information. Finally, we plan to experiment with other types
of formalisms than Prolog in the context of AKA as future work.</p>
    </sec>
    <sec id="sec-4">
      <title>4. Research question(s) and hypotheses</title>
      <p>This thesis aims to contribute to AKA by representing the knowledge in KBs by means of LLMs with
reasoning abilities. More precisely, we aim to answer the following question: How to leverage knowledge
expressed in the reasoning traces of a LLM with reasoning abilities to automatically create and manage a
domain KB?</p>
      <p>Developing the research question, we postulate the hypotheses described below.</p>
      <p>H1: Given a defined reasoning formalism, LLMs with reasoning abilities are able, possibly
with additional training, to generate a reasoning trace in this formalism.</p>
      <p>LLMs have been used to generate Prolog outputs [25] but no works on this task have been published
using LLMs with reasoning abilities, probably due to their recency. In a first step, we aim to assess
whether such LLMs can generate syntactically valid Prolog in the reasoning part. The associated
experiment is detailed in Section 5.</p>
      <p>H2: Reinforcing the ability of the LLM to generate formalized knowledge for reasoning
traces about a problem strengthens its reliability to generate valid answers, and valid
knowledge.</p>
      <p>
        This hypothesis necessitates an interest in both the reasoning traces for a given problem and, more
generally, the KB that is produced by solving multiple problems. Regarding the reasoning traces, we
start by introducing a measure  that allows us to measure how much a code is syntactically valid in a
given language. We initially applied it to the Prolog language with the idea that further work will need
to be done to also assess the semantic validity of the generated Prolog. Assessing the quality of a KB is
much more complicated and will be the focus of further work. This kind of evaluation already exists
and it involves several steps [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]:
• Debugging : detecting inconsistencies (conflict, redundancy, subsumption)
• Checking for completeness : detecting incompleteness (missing rules)
      </p>
      <p>In the next section (Section 5), we begin by working on the first hypothesis. We designed an
experiment to assess if each LLM tested is able to generate the reasoning trace in the syntax of the
Prolog formalism. To achieve that, we use the measure  and, in the rest of the PhD, we will introduce
other measures to evaluate the quality of the created KB.</p>
    </sec>
    <sec id="sec-5">
      <title>5. Preliminary results</title>
      <p>Among the available (i.e. open-weight) LLMs with reasoning abilities, we consider DeepSeek-R1, and
Cogito, both in their distilled versions due to their computational costs. Concretely, we test the 1.5B1
and 8B2 distilled versions of DeepSeek-R1 and the 3B3 and 8B4 distilled versions of Cogito. We did not
consider QwQ [31] or any larger model as they are too large for the resources to which we had access
at this point. As we aim to fine-tune the model in future work, closed-source LLMs are not considered
here.</p>
      <p>
        Prompting A problem (i.e. a question whose answer requires reasoning) is prompted to one of the
four LLMs, instructing it to express its reasoning in Prolog. In these preliminary experiments, this
problem is taken from the medical-o1-dataset-SFT [13]. The problem is inserted in a prompt inspired by
the one used to train DeepSeek [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ], as shown in Table 3. We evaluate the quality of the Prolog generated
in the reasoning part using the reward score  defined below. We repeat the process 10 times for each
LLM for the same problem to get an averaged overview of the quality of the Prolog generated in the
reasoning.
      </p>
      <p>Given a parser5,  is defined as
 = {
1 − (/)
1
if  &gt; 0
else
where  is the number of errors detected by the parser and  is the number of lines of the reasoning
traces.  is equal to 0 if the LLM did not return anything in the reasoning.</p>
      <p>Table 2 showcases the results obtained. The low  values obtained are justified by the fact that LLMs are
not specifically trained for this task. However, we did expect to see better results from the DeepSeek-R1
models. When investigating these results, the reasoning is always blank for DeepSeek-R1:8B. Actually,
no matter the version of DeepSeek-R1, and despite the number of tries with DeepSeek-R1, we did not
achieve to incentivize the model to generate Prolog in the reasoning. We notice that the Cogito models
1https://huggingface.co/deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B
2https://huggingface.co/deepseek-ai/DeepSeek-R1-Distill-Llama-8B
3https://huggingface.co/deepcogito/cogito-v1-preview-llama-3B
4https://huggingface.co/deepcogito/cogito-v1-preview-llama-8B
5Here we use https://github.com/akvnn/prolog-parser as the base of the parser.
are performing great on our measure, Cogito:3B having better performances than Cogito:8B on average.
Even though the measure  is low, it is a measure evaluating the syntax of the Prolog code, it does not
evaluate how much the Prolog code is meaningful. The meaningfulness of the Prolog code generated
will be assessed when we evaluate the KB in the future of this thesis.</p>
      <sec id="sec-5-1">
        <title>Model</title>
        <p>cogito:3b
cogito:8b
deepseek-r1:1.5b
deepseek-r1:8b</p>
      </sec>
      <sec id="sec-5-2">
        <title>Mean</title>
        <p>A conversation between User and Assistant. The user asks a question, and the Assistant solves it.
The assistant first thinks about the reasoning process in the mind and then provides the user with
the answer. The reasoning process and answer are enclosed within &lt;think&gt; &lt;/think&gt;, i.e., &lt;think&gt;
reasoning process here &lt;/think&gt; tags. All reasoning inside &lt;think&gt; must be done in Prolog style:
define facts, rules, and queries like a Prolog program. Use step-by-step logic and inference. Do not
explain in natural language — only use Prolog. User: prompt. Assistant:</p>
        <p>
          LLM Fine-Tuning The results obtained in the prompting phase encourage us to pursue with
finetuning the Cogito LLM. The Prolog code generated in the reasoning of Cogito compared to DeepSeek-R1
makes us think that there is room for improvement. We plan to use the GRPO algorithm [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ] (i.e. the
same algorithm behind DeepSeek-R1) to fine-tune Cogito, using the score  as the reward to maximize.
This fine-tuning is the next step in the thesis.
        </p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>6. Evaluation</title>
      <p>
        To answer our research question, we will focus on the same dataset (i.e. the medical-o1-reasoning-SFT
dataset). However, to go beyond the syntactic evaluation we realized in Section 5, a first step will be to
build a first KB by making an LLM reason over more than one problem. Taking inspiration from the
literature [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], we will then introduce measures to validate the automatically acquired knowledge. In a
second step, we plan to test our method on the use case in the HR domain. To validate the automatically
acquired knowledge, we will survey practitioners of the domain.
      </p>
    </sec>
    <sec id="sec-7">
      <title>7. Reflection and future work</title>
      <p>
        As mentioned in Section 5, we plan to fine-tune the Cogito LLM with the GRPO algorithm, using the
measure  as the reward to train the LLM to more robustly generate syntactically valid Prolog in its
reasoning traces. One objective of our initial experiments was to check whether existing models had
some level of ability on which such fine-tuning could build to achieve this task. As we have shown,
this appears to be true for Cogito, but not for DeepSeek-R1. The next experiment to be carried out
will therefore have for objective to test how much the ability of Cogito to generate reasoning traces as
Prolog can be improved through fine-tuning. This will be the first step into an automatic management
of a KB’s lifecycle. We plan to build on this capability to extract domain knowledge, dealing with
semantic validity and coherence over multiple problems in addition to syntactic validity. In addition, as
mentioned earlier in this paper, we chose to use Prolog as a first formalism with which to test LLMs. An
expected contribution will also include understanding which other formalisms might be more suitable
for expressing the reasoning traces of LLMs. We also anticipate the need for a larger, more robust
framework as showcased in related studies [
        <xref ref-type="bibr" rid="ref3">3, 26, 27</xref>
        ]. Building a larger framework will allow us to
evaluate all steps independently in order to achieve the automatic construction of a robust KB. This
framework, being applied to the HR domain, will also need a formalization, ensuring a high robustness.
      </p>
      <p>As mentioned above, we need to go beyond evaluating the syntactic validity of the generated reasoning
trace, and also address the quality of the KB created from combining such knowledge used for multiple,
related problems. Hence, a bibliographic study on the evaluation of the quality of KBs is one of our
priorities. In the long-term view of this research, we will use the formalism generation capability of
the LLM in the reasoning to validate or invalidate its final prediction. This will act as a safeguard
to flag invalid and/or illogical answers. We therefore anticipate for our method to also contribute to
interpretability, providing a verifiable link between the reasoning trace provided and the generated
answer. Indeed, while natural language in the reasoning, as it is currently expressed, provides a way of
better understanding whether the answer might or might not be correct, our approach could provide a
way to automatize this verification.</p>
    </sec>
    <sec id="sec-8">
      <title>Acknowledgments</title>
      <p>The author acknowledges Pr. Armelle Brun, professor of computer science at Université de Lorraine, Pr.
Mathieu d’Aquin, professor of computer science at Université de Lorraine as supervisors for this PhD,
and Dr. Yacine Abboud, director at Datanello. This research is funded by an ANRT grant through the
CIFRE system with Datanello as the industrial collaborator.</p>
    </sec>
    <sec id="sec-9">
      <title>Declaration on Generative AI</title>
      <p>During the preparation of this work, the author used GPT-4 to generate LATEX formulas and Writefull to
review the syntax of the document. After using these tools, the author reviewed and edited the content
as needed and take(s) full responsibility for the publication’s content.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>M.</given-names>
            <surname>Suwa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. C.</given-names>
            <surname>Scott</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E. H.</given-names>
            <surname>Shortlife</surname>
          </string-name>
          ,
          <article-title>An approach to verifying completeness and consistency in a rule-based expert system</article-title>
          ,
          <source>Ai Magazine</source>
          <volume>3</volume>
          (
          <year>1982</year>
          )
          <fpage>16</fpage>
          -
          <lpage>16</lpage>
          . URL: https://ojs.aaai.org/aimagazine/index. php/aimagazine/article/view/377.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>V. K.</given-names>
            <surname>Kommineni</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>König-Ries</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Samuel</surname>
          </string-name>
          ,
          <article-title>From human experts to machines: An LLM supported approach to ontology and knowledge graph construction, 2024</article-title>
          . URL: http://arxiv.org/abs/2403. 08345. doi:
          <volume>10</volume>
          .48550/arXiv.2403.08345, arXiv:
          <fpage>2403</fpage>
          .08345 [cs].
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>H.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Shen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Lv</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Ni</surname>
          </string-name>
          , J. Ye, SAC-KG:
          <article-title>Exploiting Large Language Models as Skilled Automatic Constructors for Domain Knowledge Graphs</article-title>
          ,
          <year>2024</year>
          . URL: http://arxiv.org/abs/2410.02811. doi:
          <volume>10</volume>
          .48550/arXiv.2410.02811, arXiv:
          <fpage>2410</fpage>
          .02811 [cs].
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>A. S.</given-names>
            <surname>Lippolis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. J.</given-names>
            <surname>Saeedizade</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Keskisärkkä</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Zuppiroli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Ceriani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Gangemi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Blomqvist</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. G.</given-names>
            <surname>Nuzzolese</surname>
          </string-name>
          ,
          <source>Ontology Generation using Large Language Models</source>
          ,
          <year>2025</year>
          . URL: http://arxiv.org/ abs/2503.05388. doi:
          <volume>10</volume>
          .48550/arXiv.2503.05388, arXiv:
          <fpage>2503</fpage>
          .05388 [cs].
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Shao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Zhu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Song</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Bi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y. K.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Wu</surname>
          </string-name>
          , D. Guo,
          <source>DeepSeekMath: Pushing the Limits of Mathematical Reasoning in Open Language Models</source>
          ,
          <year>2024</year>
          . URL: http://arxiv.org/abs/2402.03300. doi:
          <volume>10</volume>
          .48550/arXiv.2402.03300, arXiv:
          <fpage>2402</fpage>
          .03300 [cs].
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <surname>DeepSeek-AI</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Guo</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Yang</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          <string-name>
            <surname>Zhang</surname>
            , J. Song,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Zhang</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Xu</surname>
            ,
            <given-names>Q.</given-names>
          </string-name>
          <string-name>
            <surname>Zhu</surname>
            , S. Ma,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          <string-name>
            <surname>Bi</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          <string-name>
            <surname>Zhang</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          <string-name>
            <surname>Yu</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          <string-name>
            <surname>Wu</surname>
            ,
            <given-names>Z. F.</given-names>
          </string-name>
          <string-name>
            <surname>Wu</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          <string-name>
            <surname>Gou</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          <string-name>
            <surname>Shao</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          <string-name>
            <surname>Li</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          <string-name>
            <surname>Gao</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Liu</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Xue</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Wu</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Feng</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Lu</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Zhao</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Deng</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Zhang</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Ruan</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Dai</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Chen</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Ji</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          <string-name>
            <surname>Li</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Lin</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Dai</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Luo</surname>
            ,
            <given-names>G.</given-names>
            Hao, G. Chen, G.
          </string-name>
          <string-name>
            <surname>Li</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          <string-name>
            <surname>Zhang</surname>
            ,
            <given-names>H.</given-names>
            Bao, H.
          </string-name>
          <string-name>
            <surname>Xu</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          <string-name>
            <surname>Ding</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          <string-name>
            <surname>Xin</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          <string-name>
            <surname>Gao</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          <string-name>
            <surname>Qu</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          <string-name>
            <surname>Li</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Guo</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Li</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Chen</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Yuan</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Qiu</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Li</surname>
            ,
            <given-names>J. L.</given-names>
          </string-name>
          <string-name>
            <surname>Cai</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Ni</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Liang</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Chen</surname>
          </string-name>
          , K. Dong,
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>