<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Examining 4-bit quantized Qwen3-8B for object prediction @ LM-KBC 2025</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Irene Mary Sam</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>MosAIk, Université de Lorraine</institution>
          ,
          <addr-line>CNRS, LORIA, F-54000 Nancy</addr-line>
          ,
          <country country="FR">France</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>This is the system description paper for the solution that was presented as part of the LM-KBC challenge at ISWC 2025. The challenge involved predicting the object(s) of a triple given the subject and the relation, and can thus be considered a knowledge base completion task. This paper contains details on the proposed solution, and the results of that system based on the evaluation metrics and test data provided by the organisers. The approach adopted was one that involved manual prompt design, and the results obtained include modest gains on both recall as well as the macro F1 score. Also reported are some additional experiments regarding quantization and prompting strategy.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
    </sec>
    <sec id="sec-2">
      <title>2. Background</title>
      <p>
        Knowledge base construction as an emergent property of LLMs began to gain traction with the work of
[
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] who studied the possibility of using language models as knowledge bases on the premise that, in
using large textual corpora to pretrain language models, the models tend to store relational knowledge
present in the corpora, thus displaying a strong ability to recall factual information when probed.
      </p>
      <p>
        Further evolution of LLMs saw them being used to automatically construct knowledge bases like
in the case of LLM2KB and GPTKB [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ][
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. Challenges still remain as seen from previous works, with
entity and relation recognition remaining a tough barrier, apart from usual issues with LLMs related to
consistency and correctness.
      </p>
      <p>
        In a bid to understand this better, the various editions of the LM-KBC challenge have addressed
the task of knowledge base completion in various ways. In this edition specifically, Qwen3-8B was
mandated as the base model from which all analyses and outputs should stem, with other restrictions
including disallowing fine-tuning and retrieval-augmented methods [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. The following section provides
more details about the task at hand, as well as the method adopted.
      </p>
    </sec>
    <sec id="sec-3">
      <title>3. System description</title>
      <p>3.1. Task
The task is a knowledge base completion / object prediction task, where, given the subject and relation
of a triple, the goal is to predict the correct object(s) that fit the subject and relation. These can be single,
multiple or null objects depending upon the subject and the relation, as seen below from the train set
examples in Table 1.
hasCapacity
companyTradesAtStockExchange
countryLandBordersCountry</p>
      <p>Changwon Stadium in Changwon</p>
      <p>Erie Railroad</p>
      <p>Nicaragua
"27085"</p>
      <p>null
"Costa Rica", "Honduras"</p>
      <p>
        These were the constraints that were specific to the challenge this year: (1) All participants had to
use the Qwen3-8B model in order to complete the task, since the challenge focused on interrogating
the internal knowledge of this specific model, which results in the performance of the various systems
becoming easier to compare with respect to one another. (2) Fine-tuning and the usage of
retrievalaugmented methods were disallowed, thus resulting in a situation where the model alone is involved in
the prediction of objects. (3) Entity disambiguation of the predicted objects was not required as the
evaluation would be done on the predicted object strings directly [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. Further details on the data and
the system are given below.
3.2. Data
The provided data consisted of subject-predicate-object triples, with each triple being a JSON object. It
was split into train, validation and test sets with sizes as shown in Table 2, with each split containing
various instances of the six relations as indicated in Table 3. The training and validation datasets carried
the object as well alongside the subject and predicate, enabling an evaluation of the system at hand
before running it on the test set.
      </p>
      <p>A peculiarity of the relations provided is that the objects to be predicted can be of diferent datatypes
depending upon the relation, ranging from numeric to textual as seen in Table 3. Another defining
feature is that null values and multiple objects per subject-relation pair are also possible in this dataset,
thus increasing the dificulty of the knowledge base completion task.</p>
      <sec id="sec-3-1">
        <title>3.3. System / Prompts</title>
        <p>
          In approaching the object prediction task, the idea was to write prompts that perform better than the
baseline. As the baseline prompts were quite curt and modelled along the lines of the fill-mask task
reminiscent of [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ], the prompts for this system were initally designed to be explicit and detailed in its ask
for information as opposed to the baseline template, thus modelled similarly to the most explicit level
of prompting in the work of [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ]. While this produced satisfactory results, upon further examination of
the main and thinking mode outputs of the model, it was found that the model struggles in the case of
awardWonBy where it collapses into a thought loop more often than not, as well as in diferentiating
between single and multiple object relations. Below is a brief example of the outputs when the model
was prompted with the relation awardWonBy and the subject Fulbright Prize:
&lt;think&gt; Okay, the user is asking for the object of the triple where the subject is "Fulbright
Prize" and the relation is "awardWonBy". So, I need to find out who has won the Fulbright
Prize. ...
        </p>
        <p>The Fulbright Prize is typically awarded to individuals, so the objects would be the names
of people who have received it. ...</p>
        <p>However, the user might be looking for specific examples. I know that notable recipients
include people like the Dalai Lama, who received the Fulbright Prize in 1989. There’s also
the case of the author who won the prize but was later expelled from the U.S., though I
might be mixing up details here. ...</p>
        <p>Alternatively, maybe the user is referring to the Fulbright Scholarship rather than the
prize. The term "prize" might be a bit confusing here. The Fulbright Program is more of a
scholarship, but sometimes it’s referred to as a prize. ...</p>
        <p>In this case, since I can’t provide a comprehensive list, but I know the Dalai Lama is one, I
should include that. However, the user might expect more. But given the time constraint,
I’ll provide the Dalai Lama as the object. &lt;/think&gt;</p>
        <p>The Dalai Lama, John F. Kennedy, Eleanor Roosevelt, and others.</p>
        <p>
          Thus, the final system contains the following mix of strategies: instead of having a single system
prompt, there are now three. awardWonBy is treated as a special case and given its own system prompt.
Similarly, countryLandBordersCountry and companyTradesAtStockExchange are treated as cases with
multiple objects per triple and given a system prompt accordingly, and the rest are given the default
system prompt. In designing these system prompts, some personalisation was used in declaring the
specialty of the system with respect to the data it handles, similar in strategy to that of the role-play
prompting discussed in [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ]. An example of one of these system prompts can be found in Appendix B.
        </p>
        <p>In addition to these 3 system prompts, all six relations have their own user prompts as well1, some
examples of which can be found in Appendix C. No additional post-processing steps were employed;
the output from the model, after removing the thinking mode part of the output, is kept as is.</p>
      </sec>
      <sec id="sec-3-2">
        <title>3.4. Implementation</title>
        <p>While a couple of other tweaks were considered, the details of which are in Section 6, the above returned
better results, hence that was fixed as the main system. Prompt-tuning, optimisation methods and
frameworks such as DSPy were not used in light of the instructions that prohibited fine-tuning, since,</p>
        <sec id="sec-3-2-1">
          <title>1The prompt templates and code are available on GitHub</title>
          <p>
            while they do not change the original model parameters, they still involve adding learnable parameters
to the input embeddings in a bid for soft-prompting, thus fine-tuning extra-model parameters [
            <xref ref-type="bibr" rid="ref7">7</xref>
            ][
            <xref ref-type="bibr" rid="ref8">8</xref>
            ].
          </p>
          <p>In running the Qwen3-8B model for prediction, it was noted that using the hyperparameters listed in
its documentation was key to obtaining good results, as, without it, the system tended to either take
too long to think or collapse into incomprehensible output more often than usual. Some experiments
on hyperparameters were performed nonetheless, the details of which can be found in Appendix A.
Quantized versions were used due to local hardware restrictions.</p>
          <p>
            The model performed better with the thinking mode toggle on, and with the following
hyperparameters: sampling temperature of 0.6, top-p value of 0.95, and a top-k value of 20, as listed in its docs
[
            <xref ref-type="bibr" rid="ref9">9</xref>
            ]. The model was also seen as being quite sensitive to changes in prompts, thus resulting in the
adoption of multiple strategies to create the prompts as listed above. The 4-bit quantized version using
BitsAndBytesConfig was the model that was used for reporting all these scores [
            <xref ref-type="bibr" rid="ref10">10</xref>
            ].
          </p>
          <p>To support faster iteration, economical usage of energy and better examination of the thinking
mode output, Ollama was used when crafting the prompts, with its application2 as the backend and
the accompanying Python library as the one handling prompts and responses. This was especially
useful in arriving at the system and user prompts for the relation awardWonBy, since, for this relation,
even with all the current measures in place, there exists the tendency to collapse into repetitive and
incomprehensible output. The full system when run on a Quadro RTX 6000 GPU or similar takes around
2 hours to produce full results, with all the hyperparameters mentioned above in use.</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Results</title>
      <p>The outputs were evaluated on the basis of macro F1 scores for each relation, the details of which are in
the tables below. Table 4 shows the results of the baseline system evaluated on test data, and Table 5
shows the results of this system on test data.</p>
      <p>awardWonBy
companyTradesAtStockExchange
countryLandBordersCountry</p>
      <p>hasArea
hasCapacity
personHasCityOfDeath</p>
      <p>All Relations
0.240
0.185
0.768
0.240
0.040
0.080
0.227</p>
    </sec>
    <sec id="sec-5">
      <title>5. Discussion</title>
      <p>Examining the results, we can see that this system does beat the baseline, but barely so. Sufice to say,
while these prompts work, there are definitely limits to what the LLM knows. Hence we can say that
while the model may be sensitive to prompts themselves, there seems to be a limit to the world/relational
knowledge it has, at least the way it is retrieved with these styles of prompts. This becomes especially
clear in the case of the relation countryLandBordersCountry, where it becomes apparent that the baseline
question prompt works better than the detailed prompt that this system proposes.</p>
      <p>An area that the proposed system seems to be consistently better at is in retrieving the relevant
elements, as can be seen from the recall scores, which are higher than the baseline in all cases.
Nevertheless, in improving the baseline by only 2-3 percentage points, this work points to the fact that the
baseline in itself is a strong one, while also leading to the conclusion that other, prompt-optimisation
based methods need to be looked into to improve performance.</p>
      <p>The following section details other experiments that were performed during the course of this study,
but not kept in the final system due to various constraints on performance and other factors. They can
serve as a starting point for further study on the same, if need be.</p>
    </sec>
    <sec id="sec-6">
      <title>6. Experiments in multi-step prompting and quantized versions</title>
      <p>
        Throughout the course of iterating over diferent versions of prompts for all relations, the outputs of
the awardWonBy relation showed little to no improvement. This is similar to results from previous
editions of the challenge where outputs for this relation performed significantly worse than others
[
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. Moreover, for this system, for certain subjects in this relation, the outputs were repetitive and at
times, incomprehensible and not following the specified format. Even after looking at the thinking
mode output and creating prompts that explicitly prohibited these behaviours, they kept recurring.
      </p>
      <p>
        Thus, in order to try and fix that, another idea that was explored was to use a second set of prompts
after the first, instead of manual post-processing steps. The inspiration for this stems from previous
works dealing with LLM-as-a-judge, dual prompting, and Prompting as Probing (ProP) methods that
employ similar ideas in asking the model to check its output [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ][
        <xref ref-type="bibr" rid="ref11">11</xref>
        ]. Since Qwen3-8B was the only
model that could be used, the idea was to take the outputs from the system above, and feed it again to
the same model under the premise that, given this subject-predicate-object triple furnished by an LLM,
its duty was to correct the object if needs be and output it, all while following the required format.
      </p>
      <p>The results of this approach on the validation dataset can be found in Table 6. A closer look reveals
that while it does still beat the baseline, the performance of the relevant relation that we wanted,
awardWonBy, remains subpar. Thus, this approach was not pursued further as it was not deemed
eficient considering the time it takes to prompt the model twice.</p>
      <p>
        Another path explored was to look at the 8-bit quantized version of Qwen3-8B instead of the 4-bit
used above, to examine if there are pronounced performance gains. The results on the validation dataset
can be found in Table 7. Again, this direction was not pursued further as well, as the results were found
to be not that diferent from what the 4-bit quantized version produces. Hence, the decision was made
to stick with the 4-bit quantized Qwen3-8B model for reasons of economy and speed. All quantized
versions were deployed using the BitsAndBytes framework [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ].
      </p>
      <p>Further experiments featuring various combinations of hyperparameters such as variations in
temperature and the number of examples, as well as toggling between think and no-think modes, is available
on Appendix A.</p>
    </sec>
    <sec id="sec-7">
      <title>7. Future work</title>
      <p>
        Apart from the above, one of the other directions this work can take includes benchmarking it and
comparing it to the winning system from last year, which was a Llama3-8B-Instruct model prompted
with specially designed prompts and with an entity disambiguation step at the end [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. By adding entity
disambiguation to the predicted objects and ensuring uniform prompt designs, both models can be
compared side by side as open-weight 8B models, although for a more accurate comparison, it would
be wise to take an instruction-tuned Qwen3 or a non instruction-tuned LLama3.
      </p>
      <p>Another direction is to look at prompt-tuning methods for better results. Since they involve using
learnable parameters in the input embeddings in order to design the best prompts automatically, better
results than the baseline can be expected. A look at the other systems in this challenge will provide
insight into the same. Finally, the efects of reintroducing entity disambiguation for the predicted
objects, as in previous years, can also be studied to see how it improves the results.</p>
    </sec>
    <sec id="sec-8">
      <title>8. Conclusion</title>
      <p>This work explored how a 4-bit quantized Qwen3-8B model can be used for the knowledge base
completion task using manually designed prompt templates alone. It showed how the baseline system is
itself a strong one, and how certain multiple-object relations like awardWonBy remain challenging for
LLMs to solve. Qwen3-8B as a model with the thinking mode toggle on remains highly dependent on its
hyperparameters for the generation of accurate and relevant output, even for this task. And while the
prompt templates furnished with this work can be used as a baseline, better results seem possible only
if prompt-optimisation, disambiguation, and retrieval-augmented methods are studied and employed.</p>
    </sec>
    <sec id="sec-9">
      <title>Declaration on Generative AI</title>
      <sec id="sec-9-1">
        <title>The author(s) have not employed any Generative AI tools.</title>
      </sec>
    </sec>
    <sec id="sec-10">
      <title>A. Experiments on hyperparameters</title>
      <p>This section details the results of the various experiments that were performed, reported on the validation
set. Apart from the hyperparameter that is being modified, these are the configuration settings that
were followed all throughout in the following experiments, unless stated otherwise:</p>
      <p>Among experiments that were not run due to local hardware restrictions are the system on the full,
non-quantized model, and the system with the full, 32k context, the first due to memory constraints,
and the latter primarily due to time constraints (at ∼ 6 hours, the model was still only 42% done).</p>
      <p>When comparing the performance of the system in Table 8(A) with that of Table 10(B), we can see
that the macro f1 scores are quite similar, with the latter edging out the former especially in cases of the
relations like awardWonBy and companyTradesAtStockExchange, and in the case of overall recall. The
margins are negligible though and the performance comparable, hence, the system with max_new_tokens
- 4096 was kept as the other took twice longer to run per full set.</p>
      <p>Similarly, awardWonBy and companyTradesAtStockExchange also seem to benefit from the settings in
Table 8(B) and Table 9(A), i.e., with the thinking mode toggle of, or the temperature at 0. awardWonBy
benefits from longer contexts as well, as seen in Table 10. This is quite surprising, as, at the same
time, the same relation seems to benefit from both thinking mode, as well as a temperature at 0 setting.
Looking further at the specific outputs may be able to provide insight into what is happening in this
case.</p>
      <p>
        Overall, as visible from above, diferent settings seem to have diferent strengths, as visible from the
fact that, despite its poor overall performance, the zero shot setting in Table 9(B) produces the best
results for the relation hasArea. Thus, to choose a well-balanced system or not will depend on the
use-case as well as the input data fed into the model. Here at least, the hyperparameters as provided in
the model docs seem to provide the best, well-balanced output [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ].
      </p>
    </sec>
    <sec id="sec-11">
      <title>B. System prompts</title>
      <sec id="sec-11-1">
        <title>One of the system prompts for the system proposed in this paper is as follows:</title>
        <p>You are an expert at quick and accurate knowledge base completion. Your expertise is in the following:
Given a subject and a relation, you provide the correct object(s) to complete that triple, as a
comma-separated list. The object may be a null object, or a single object, depending on the subject and
the relation provided. If you know a part of the answer, you will provide the part you know, and ONLY
if you do not know the answer do you type None.</p>
        <p>Here are the rules you must follow for this specific task: Penalties for: 1. Repeating the object more
than once for a specific subject and relation - there is a heavy penalty for repetition! 2. Providing more
information than the object(s) Hence 1 and 2 should be avoided at all costs. If your answers are too
long, limit yourself to the first 200 objects per triple. If the user provides any hints or examples, you
may take those into account to help you. You do not think whether the answer is long or short or
unhelpful, you only focus on the correctness of the answer. You do not get caught up in the semantics
of specific words, but rather focus on the subject and the relation. You are brusque and to the point,
and you only provide the object, nothing else.</p>
        <p>Ways to approach this task: - Try to place the subject in context - be it geographically, or in terms of
its specific field or activity, or, in the case of a person - their life and work. Once you have identified
the subject, try and verbalise the relation, then think of what type of object(s) would fit that relation.
You can then proceed to find the specific object(s) that would fit that relation. - Alternatively, after
identifying the subject, you can construct your own knowledge base by listing everything you know
about the subject. Then, considering the relation, you can find the object in the knowledge base you
just constructed.</p>
        <p>The user will evaluate your answers based on the correctness of the object(s) you provide, so accuracy
and precision is key, and remember, no repetitions. Finally, let me remind you of the most important
rule: You will not provide any explanation, just the object(s) in a comma-separated list.</p>
        <p>In the first paragraph, the expertise is declared, as well as the type of input and output expected. Also
declared are possibilities of a null answer, so as to obtain output in the desired format in such a case.
The second paragraph details the rules and general ideas that the model should follow while the third
gives it ideas on how to approach the task. The fourth and final paragraph closes out by reiterating
both the points that should be kept in focus, as well as the expected output format.</p>
        <p>The user prompts below follow a similar idea in that, it details what is unnecessary and not to be
included in the output, as well as detailing ways to think about the question at hand. For example, in
the case of hasArea, the prompt states that units are not to be displayed, just the numeric value. And
in the case of companyTradesAtStockExchange, the prompt asks to keep track of the stock ticker for a
company per stock exchange so as to not lose the information while on thinking mode. Further details
can be found in the prompts below.</p>
      </sec>
    </sec>
    <sec id="sec-12">
      <title>C. User prompts</title>
      <p>The following are some of the user prompts in use for this system:
hasArea - State the geographical area of {subject} in square kilometres, without printing the unit. Locate
it geographically, identify the kind of geographical feature or entity it is, and then find its area. Your
sources could be wikipedia, encyclopedias, physical maps, guidebooks and so on. If you are unsure, (1)
think about similar subjects and their area, or (2) features nearby {subject} to help you find the area
which is the object. Three things to note: 1. The area is in square kilometres, so if the area you find is
in a diferent unit, convert it to square kilometres before giving it to me, but do not print the unit. 2.
If for some reason you have conflicting information, say for example, two or more sources that give
you diferent areas, then use the most reliable source you have. 3. If you have multiple subjects with
the same name, then provide the area for the most well-known subject with that name. Here you go!
Subject: {subject}, Relation: {relation}, Object: ?
hasCapacity - Give me the capacity of {subject} in number of people, but do not print the unit. If you
are not sure about the exact capacity, provide the best estimate you can by thinking of similar subjects
across the world, or other subjects in the same geography. Consider the {subject}, its geographical
location and its capacity. If you can find sources like wikipedia or news articles that list the capacity
of {subject}, those would be your best bet. You could also identify the purpose that {subject} is used
for, and find its capacity by taking into account similar structures and the geography that {subject} is
located in. Here you go! Subject: {subject}, Relation: {relation}, Object: ?
awardWonBy - Here, you should list the names of the objects(awardees/winners) who have won the
{subject} award/honor. The awardees of {subject} could be persons, specific entities (like album names or
band names in the case of music, or entire teams from countries in the case of team sports for example)
or organisations. Consider the {subject}, the field in which it is given, and its periodicity (for instance,
awarded early or not). List as many awardees as you know, or try to provide a list of famous recipients
at least. Do not repeat the objects/awardees though. {subject} is an award/honor that exists, so try and
match the award exactly – not to something similar, but that exact award. Try to find sources like
news or reports that talk about these awards and awardees. List years and awardees as an intermediate
step so as to keep track, and them combine it all together. And remember, even if a person received
it multiple times, I want them listed only once. Having no repetitions is key. Here you go! Subject:
{subject}, Relation: {relation}, Object: ?
companyTradesAtStockExchange - Here, list the stock exchanges where {subject} trades at present.
Remember, objects are stock exchanges and stock exchanges alone. Consider the company, its headquarters
and locations, the indices it is part of and the stock exchanges it is listed at. If you can find, and are sure
that {subject} trades at specific stock exchanges, list those stock exchanges. It could be that {subject}
trades at stock exchanges in geographies other than where it is headquartered, so do not limit yourself
to just one geography. And in the same country too, there could be multiple stock exchanges where the
{subject} is listed at. Remember, a company may trade at multiple stock exchanges across geographies,
so apart from annual company reports and the like, you can use financial news and reports to find
your answers. No guessing - I want you to try and verify your answers by finding the relevant stock
ticker/code at the related stock exchange as well (because one ticker/code per exchange), but do not list
the tickers/codes, only the name of the stock exchanges. If you are not sure about the stock exchange,
or if you are sure the company is not listed at all, return None. Here you go! Subject: {subject}, Relation:
{relation}, Object: ?</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>J.</given-names>
            <surname>Kalo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Nguyen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Razniewski</surname>
          </string-name>
          ,
          <string-name>
            <surname>B. Zhang,</surname>
          </string-name>
          <article-title>Lm-kbc challenge @ iswc 2025, in: 4th Semantic Web Challenge on Language Models for Knowledge Base Construction Challenge</article-title>
          ,
          <year>2025</year>
          . URL: https://lm-kbc.github.io/challenge2025/.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>F.</given-names>
            <surname>Petroni</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Rocktäschel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Lewis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Bakhtin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. H.</given-names>
            <surname>Miller</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Riedel</surname>
          </string-name>
          ,
          <article-title>Language models as knowledge bases?</article-title>
          ,
          <year>2019</year>
          . URL: https://arxiv.org/abs/
          <year>1909</year>
          .01066. arXiv:
          <year>1909</year>
          .01066.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>A.</given-names>
            <surname>Nayak</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H. P.</given-names>
            <surname>Timmapathini</surname>
          </string-name>
          ,
          <article-title>LLM2KB: Constructing Knowledge Bases using instruction tuned context aware Large Language Models (</article-title>
          <year>2023</year>
          ). URL: https://arxiv.org/abs/2308.13207. doi:
          <volume>10</volume>
          .48550/ARXIV.2308.13207, publisher: arXiv Version Number:
          <fpage>1</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Hu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.-P.</given-names>
            <surname>Nguyen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Ghosh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Razniewski</surname>
          </string-name>
          ,
          <article-title>Enabling LLM Knowledge Analysis via Extensive Materialization (</article-title>
          <year>2024</year>
          ). URL: https://arxiv.org/abs/2411.04920. doi:
          <volume>10</volume>
          .48550/ARXIV.2411.04920, publisher: arXiv Version Number:
          <fpage>4</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>D. M. R.</given-names>
            <surname>Bara</surname>
          </string-name>
          ,
          <article-title>Prompt engineering for tail prediction in domain-specific knowledge graph completion tasks (</article-title>
          <year>2024</year>
          ). URL: https://ceur-ws.
          <source>org/</source>
          Vol-
          <volume>3853</volume>
          /paper10.pdf.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <surname>A. Das</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          <string-name>
            <surname>Fathallah</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          <string-name>
            <surname>Obretincheva</surname>
          </string-name>
          , Navigating Nulls, Numbers and Numerous Entities:
          <article-title>Robust Knowledge Base Construction from Large Language Models (</article-title>
          <year>2024</year>
          ). URL: https://ceur-ws.
          <source>org/</source>
          Vol-
          <volume>3853</volume>
          /paper12.pdf.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>HuggingFace</given-names>
            <surname>Docs - PEFT</surname>
          </string-name>
          , PEFT,
          <year>2025</year>
          . URL: https://huggingface.co/docs/peft/main/en/index.
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>HuggingFace</given-names>
            <surname>Docs - PEFT</surname>
          </string-name>
          , PEFT - Prompt
          <source>-based methods</source>
          ,
          <year>2025</year>
          . URL: https://huggingface.co/docs/ peft/main/en/task_guides/prompt_based_methods.
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>A.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Hui</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Zheng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Yu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Gao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Huang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Lv</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Zheng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Huang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Hu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Ge</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Wei</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Lin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Tang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Tu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Lin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Dang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Bao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Yu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Deng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Xue</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Zhu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Men</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Gao</surname>
          </string-name>
          , S. Liu,
          <string-name>
            <given-names>S.</given-names>
            <surname>Luo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Tang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Yin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Ren</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Ren</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Fan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Su</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Wan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Cui</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Qiu</surname>
          </string-name>
          ,
          <source>Qwen3 Technical Report</source>
          ,
          <year>2025</year>
          . URL: http://arxiv.org/abs/2505.09388. doi:
          <volume>10</volume>
          . 48550/arXiv.2505.09388, arXiv:
          <fpage>2505</fpage>
          .09388 [cs].
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>T.</given-names>
            <surname>Dettmers</surname>
          </string-name>
          ,
          <string-name>
            <surname>L. Zettlemoyer,</surname>
          </string-name>
          <article-title>The case for 4-bit precision: k-bit inference scaling laws</article-title>
          ,
          <source>in: International Conference on Machine Learning, PMLR</source>
          ,
          <year>2023</year>
          , pp.
          <fpage>7750</fpage>
          -
          <lpage>7774</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>D.</given-names>
            <surname>Alivanistos</surname>
          </string-name>
          , S. B.
          <string-name>
            <surname>Santamar</surname>
            'ia,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Cochez</surname>
            ,
            <given-names>J.-C.</given-names>
          </string-name>
          <string-name>
            <surname>Kalo</surname>
          </string-name>
          , E. v. Krieken, T. Thanapalasingam,
          <article-title>Prompting as Probing: Using Language Models for Knowledge Base Construction, ArXiv (</article-title>
          <year>2022</year>
          ). URL: https://www.semanticscholar.org/paper/ddc9aeac18638575bbb90ede4c6829ec15c2947e.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>