<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>LM-KBC 2025: 4th Challenge on Knowledge Base Construction from Pre-trained Language Models</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Jan-Christoph Kalo</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Simon Razniewski</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Bohui Zhang</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Tuan-Phong Nguyen</string-name>
          <xref ref-type="aff" rid="aff3">3</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>King's College London</institution>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>ScaDS.AI &amp; TU Dresden</institution>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>University of Amsterdam</institution>
        </aff>
        <aff id="aff3">
          <label>3</label>
          <institution>VNU University of Engineering and Technology</institution>
        </aff>
      </contrib-group>
      <abstract>
        <p>Pretrained language models (LMs) have significantly advanced a variety of semantic tasks and have shown promise as sources of knowledge elicitation. While prior work has studied this ability through probing or prompting, the potential of LMs for large-scale knowledge base construction remains underexplored. The fourth edition of the LM-KBC Challenge invited participants to build knowledge bases directly from LMs, given specific subjects and relations. Unlike existing probing benchmarks, the challenge imposed no simplifying assumptions on relation cardinality-allowing a subject entity to be linked to zero, one, or multiple object entities. To ensure accessibility, the challenge featured a single track based the same LLM to be used by all participants. Five submissions were received, which explored a variety of ideas from self-consistency, self-RAG, reasoning, and prompt optimization.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Large language models (LLMs) such as Qwen [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], Llama [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ], and ChatGPT [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] are optimized for masked
language modeling or text completion and have achieved remarkable success on a wide range of
downstream NLP tasks, including question answering, information retrieval, and machine translation.
More recently, LLMs have attracted attention for their potential to directly produce structured knowledge
from their parameters. This is promising, as current knowledge bases (KBs) such as Wikidata [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] and
ConceptNet [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ], though central to the Semantic Web ecosystem, remain inherently incomplete [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ].
KB construction is particularly challenging due to the optional nature of many relations (e.g., place
of death, cause of death, parent organization) and the presence of multiple correct objects for a single
subject–relation pair (e.g., shares border, employer, speaks language). Moreover, KBs must be materialized
for trustworthy and consistent downstream usage [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ].
      </p>
      <p>
        Traditional approaches to knowledge base construction (KBC) have leveraged unstructured
text [
        <xref ref-type="bibr" rid="ref8 ref9">8, 9</xref>
        ], crowdsourcing [
        <xref ref-type="bibr" rid="ref10 ref11">10, 11</xref>
        ], and semi-structured resources [
        <xref ref-type="bibr" rid="ref12 ref13 ref14">12, 13, 14</xref>
        ]. Automated KBC has
long been a core topic in the Semantic Web community, spanning decades of research on knowledge
extraction, consolidation, and schema matching. The seminal LAMA paper [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ] demonstrated that
language models could rank correct object tokens highly when prompted with a subject–relation query.
While subsequent work has reported both progress [
        <xref ref-type="bibr" rid="ref16 ref17">16, 17</xref>
        ] and criticism [
        <xref ref-type="bibr" rid="ref18">18, 19, 20</xref>
        ], the potential
of LMs for KBC remains underexplored. Importantly, the LAMA benchmark and its variants are not
designed for true KB construction. Although LLMs are increasingly studied in Semantic Web tasks
such as entity recognition, relation extraction, and reasoning, most evaluations of factual knowledge
extraction remain rooted in NLP-style benchmarks with simplified assumptions.
      </p>
      <p>This challenge seizes the opportunity to bridge the gap by exploring how LLMs can contribute
to practical KB construction. Continuing previous eforts [ 21, 22, 23, 24], the 4th edition focuses
on leveraging a single locally runnable LLM to construct KBs without prior knowledge of relation
cardinalities. Specifically, given a subject–relation pair, participants were asked to design an LLM-based
system that generates candidate subject–relation–object triples, and decides whether to accept or reject
each one. The predictions were evaluated using F1-score.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Task Description</title>
      <p>In the LM-KBC Challenge, the knowledge base construction task is defined as follows: given a subject
entity  and a relation , the goal is to generate all correct object entities [1, 2, . . . , ] by probing
language models. For example, given the tuple (Greece, countryLandBorderCountry), a participant
might query the language model with a prompt such as “Greece shares a border with [MASK]”. The
system should then output country entities like [Albania, North Macedonia, Bulgaria, Turkey], in any
order. Similarly, for numeric-answer relations, such as (Wembley Stadium in London, hasCapacity),
the expected output would be [“90000”].</p>
      <p>Participants are required to build LM-based systems that produce entity labels without relying on
external resources (e.g., web search engines, retrieval-augmented generation), i.e., submitted systems
had to be fully self-contained. For comparison, we released a baseline method based on prompt templates,
covering both question-style prompts and fill-in-the-blank templates.</p>
    </sec>
    <sec id="sec-3">
      <title>3. Dataset Construction</title>
      <sec id="sec-3-1">
        <title>Relation</title>
      </sec>
      <sec id="sec-3-2">
        <title>Train</title>
      </sec>
      <sec id="sec-3-3">
        <title>Validation</title>
      </sec>
      <sec id="sec-3-4">
        <title>Test Special features</title>
        <p>awardWonBy
companyTradesAtStockExchange
countryLandBordersCountry</p>
        <p>hasArea
hasCapacity
personHasCityOfDeath</p>
        <p>The dataset was built by querying Wikidata and manually refining the results to reduce errors and
improve quality. Compared with previous editions [21, 22], the 2025 version focuses on six challenging
relations with distinctive characteristics, enabling participants to design approaches tailored to specific
problem types. These relations fall into the following categories:
1. Relations with many missing objects (e.g., a person’s place of death, or the stock exchange
where a company is listed).
2. Relations with long object lists (e.g., the list of award winners in a given field).
3. Standard relations carried over from the previous edition.</p>
        <p>For each relation, up to 100 subject entities are provided for the training, validation, and hidden test
sets used in challenge evaluation. The relations were carefully selected to ensure diversity, with subject
entities spanning diferent types such as persons, countries, and organizations. The subject–object pairs
were automatically sampled from Wikidata under the following constraints:
1. Balanced object list lengths: Longer lists were oversampled to avoid dominance by single-object
examples.
2. Balanced subject popularity: Using proxies such as total Wikidata statements or web hits, we
ensured roughly a 50/50 split between popular and long-tail subject entities for each relation.
3. Balanced object complexity: Both single-token and multi-token object entities were included.
edarsem
JingboHe
acmc
isam
aclay
Relation-Wise Self-consistency</p>
        <p>Self-RAG and DaC</p>
        <p>Soft Thinking</p>
        <p>LLM-as-a-Judge
Prompt optimization</p>
      </sec>
      <sec id="sec-3-5">
        <title>Average</title>
      </sec>
      <sec id="sec-3-6">
        <title>F1-score</title>
        <p>The 2025 dataset is publicly available on GitHub1. Participating systems submit their predictions on the
CodaLab platform2 [25], where final scores are computed on the hidden test set.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Comparison with Previous Editions</title>
      <p>Compared with previous editions of the LM-KBC challenge [21, 22], the 4th edition introduced the
following characteristics:
1. More diverse relations: As in the 3rd edition, we continue to use a smaller, representative set of
relations grouped into topical categories, enabling participants to better tailor their approaches.</p>
      <p>Compared with the 3rd edition, the numeric-answer relations were replaced by two new ones.
2. Single parameter-bounded track: The challenge continues to follow a single-track format.</p>
      <p>
        Building on last year’s fixed limit of 10 billion parameters for LLMs, we further standardized the
setting by requiring participants to use Qwen3-8B [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ].
3. High data quality: As before, substantial manual efort was invested to ensure the dataset meets
the highest quality standards.
      </p>
    </sec>
    <sec id="sec-5">
      <title>5. Results of Submissions</title>
      <p>1https://github.com/lm-kbc/dataset2025
2https://codalab.lisn.upsaclay.fr/competitions/23218</p>
      <p>From the detailed per-relation results in Table 3, we observe that no single method
consistently outperforms the others across all relations. The ReWiSe system proposed by
AlbertRoulhac and Zouaq achieved the highest F1-score in three out of six relations. The
SelfRAG and DaC, Soft Thinking, and LLM-as-a-Judge methods achieved the best performance on
awardWonBy, companyTradesAtStockExchange, and hasCapacity, respectively. Overall, the
relations awardWonBy, hasArea, and hasCapacity remain particularly challenging. Compared with
the baseline, system improvements on these relations were marginal, and their absolute F1-scores stayed
relatively low.</p>
      <sec id="sec-5-1">
        <title>Recall F1-score</title>
        <p>edarsem
JingboHe
acmc
isam
aclay
baseline
edarsem
JingboHe
acmc
isam
aclay
baseline
edarsem
JingboHe
acmc
isam
aclay
baseline
edarsem
JingboHe
acmc
isam
aclay
baseline
edarsem
JingboHe
acmc
isam
aclay
baseline
edarsem
JingboHe
acmc
isam
aclay
baseline</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>Declaration on Generative AI</title>
      <p>The authors have not employed any Generative AI tools.
in Natural Language Inference, in: ACL, 2019. URL: https://aclanthology.org/P19-1334/. doi:10.
18653/v1/P19-1334.
[19] N. Kassner, H. Schütze, Negated and Misprimed Probes for Pretrained Language Models: Birds
Can Talk, But Cannot Fly, in: ACL, 2020. URL: https://aclanthology.org/2020.acl-main.698/.
doi:10.18653/v1/2020.acl-main.698.
[20] B. Cao, H. Lin, X. Han, L. Sun, L. Yan, M. Liao, T. Xue, J. Xu, Knowledgeable or Educated Guess?
Revisiting Language Models as Knowledge Bases, in: C. Zong, F. Xia, W. Li, R. Navigli (Eds.),
ACL, Online, 2021. URL: https://aclanthology.org/2021.acl-long.146/. doi:10.18653/v1/2021.
acl-long.146.
[21] J.-C. Kalo, S. Singhania, S. Razniewski, J. Z. Pan, et al., LM-KBC 2023: 2nd Challenge on Knowledge
Base Construction from Pre-trained Language Models, in: Joint proceedings of 1st workshop
on Knowledge Base Construction from Pre-Trained Language Models (KBC-LM) and the 2nd
challenge on Language Models for Knowledge Base Construction (LM-KBC), 2023. URL: https:
//ceur-ws.org/Vol-3577/paper0.pdf.
[22] J.-C. Kalo, T.-P. Nguyen, S. Razniewski, B. Zhang, Preface: LM-KBC Challenge 2024, in: Joint
Proceedings of the KBC-LM Workshop and the LM-KBC Challenge 2024, 2024. URL: https://
ceur-ws.org/Vol-3853/paper0.pdf.
[23] D. Alivanistos, S. B. Santamaría, M. Cochez, J.-C. Kalo, E. van Krieken, T. Thanapalasingam,
Prompting as Probing: Using Language Models for Knowledge Base Construction, 2023. URL:
https://arxiv.org/abs/2208.11057. arXiv:2208.11057.
[24] B. Zhang, I. Reklos, N. Jain, A. M. Peñuela, E. Simperl, Using Large Language Models for Knowledge
Engineering (LLMKE): A Case Study on Wikidata, 2023. URL: https://arxiv.org/abs/2309.08491.
arXiv:2309.08491.
[25] A. Pavao, I. Guyon, A.-C. Letournel, D.-T. Tran, X. Baro, H. J. Escalante, S. Escalera, T. Thomas,
Z. Xu, CodaLab Competitions: An Open Source Platform to Organize Scientific Challenges, Journal
of Machine Learning Research (2023). URL: http://jmlr.org/papers/v24/21-1436.html.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>A.</given-names>
            <surname>Yang</surname>
          </string-name>
          , et al.,
          <source>Qwen3 Technical Report</source>
          ,
          <year>2025</year>
          . URL: https://arxiv.org/abs/2505.09388. arXiv:
          <volume>2505</volume>
          .
          <fpage>09388</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>H.</given-names>
            <surname>Touvron</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Lavril</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Izacard</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Martinet</surname>
          </string-name>
          , M.
          <article-title>-</article-title>
          <string-name>
            <surname>A. Lachaux</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Lacroix</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Rozière</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          <string-name>
            <surname>Goyal</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          <string-name>
            <surname>Hambro</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Azhar</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Rodriguez</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Joulin</surname>
            , E. Grave, G. Lample, LLaMA: Open and
            <given-names>Eficient</given-names>
          </string-name>
          <string-name>
            <surname>Foundation Language Models</surname>
          </string-name>
          ,
          <year>2023</year>
          . URL: https://arxiv.org/abs/2302.13971. arXiv:
          <volume>2302</volume>
          .
          <fpage>13971</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <surname>O. team</surname>
          </string-name>
          , GPT-4
          <source>Technical Report</source>
          ,
          <year>2024</year>
          . URL: https://arxiv.org/abs/2303.08774. arXiv:
          <volume>2303</volume>
          .
          <fpage>08774</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>D.</given-names>
            <surname>Vrandečić</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Krötzsch</surname>
          </string-name>
          , Wikidata:
          <string-name>
            <given-names>A Free</given-names>
            <surname>Collaborative</surname>
          </string-name>
          <string-name>
            <surname>Knowledgebase</surname>
          </string-name>
          , Commun. ACM (
          <year>2014</year>
          ). URL: https://doi.org/10.1145/2629489. doi:
          <volume>10</volume>
          .1145/2629489.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>R.</given-names>
            <surname>Speer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Chin</surname>
          </string-name>
          , C. Havasi, ConceptNet
          <volume>5</volume>
          .
          <article-title>5: An Open Multilingual Graph of General Knowledge</article-title>
          , in: AAAI,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>S.</given-names>
            <surname>Razniewski</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Arnaout</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Ghosh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Suchanek</surname>
          </string-name>
          , Completeness, recall, and
          <article-title>negation in openworld knowledge bases: A survey, ACM Computing Surveys (</article-title>
          <year>2024</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>B.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , E. Koutsiana,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Meroño-Peñuela</surname>
          </string-name>
          , E. Simperl,
          <article-title>Trustworthy Knowledge Graphs: Practices and Approaches, in: Handbook on Neurosymbolic AI and Knowledge Graphs</article-title>
          , IOS Press,
          <year>2025</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>N.</given-names>
            <surname>Nakashole</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Theobald</surname>
          </string-name>
          , G. Weikum,
          <article-title>Scalable Knowledge Harvesting with High Precision and High Recall</article-title>
          , in: WSDM,
          <year>2011</year>
          . URL: https://doi.org/10.1145/1935826.1935869. doi:
          <volume>10</volume>
          .1145/ 1935826.1935869.
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>X.</given-names>
            <surname>Dong</surname>
          </string-name>
          , E. Gabrilovich, G. Heitz,
          <string-name>
            <given-names>W.</given-names>
            <surname>Horn</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Lao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Murphy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Strohmann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Sun</surname>
          </string-name>
          , W. Zhang,
          <article-title>Knowledge Vault: A Web-Scale Approach to Probabilistic Knowledge Fusion</article-title>
          , in: KDD,
          <year>2014</year>
          . URL: https://doi.org/10.1145/2623330.2623623. doi:
          <volume>10</volume>
          .1145/2623330.2623623.
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>M.</given-names>
            <surname>Acosta</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Zaveri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Simperl</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Kontokostas</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Auer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Lehmann</surname>
          </string-name>
          ,
          <article-title>Crowdsourcing Linked Data Quality Assessment</article-title>
          , in: H.
          <string-name>
            <surname>Alani</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <string-name>
            <surname>Kagal</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Fokoue</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Groth</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Biemann</surname>
            ,
            <given-names>J. X.</given-names>
          </string-name>
          <string-name>
            <surname>Parreira</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <string-name>
            <surname>Aroyo</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          <string-name>
            <surname>Noy</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Welty</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          Janowicz (Eds.), ISWC,
          <year>2013</year>
          . URL: https://link.springer.com/chapter/ 10.1007/978-3-
          <fpage>642</fpage>
          -41338-4_
          <fpage>17</fpage>
          . doi:
          <volume>10</volume>
          .1007/978-3-
          <fpage>642</fpage>
          -41338-4\_
          <fpage>17</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>A.</given-names>
            <surname>Kobren</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Logan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Sampangi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>McCallum</surname>
          </string-name>
          ,
          <article-title>Domain Specific Knowledge Base Construction via Crowdsourcing</article-title>
          , in: AKBC,
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>K.</given-names>
            <surname>Bollacker</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Evans</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Paritosh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Sturge</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Taylor</surname>
          </string-name>
          , Freebase:
          <string-name>
            <given-names>A Collaboratively</given-names>
            <surname>Created</surname>
          </string-name>
          <article-title>Graph Database for Structuring Human Knowledge</article-title>
          , in: SIGMOD, Association for Computing Machinery,
          <year>2008</year>
          . URL: https://doi.org/10.1145/1376616.1376746. doi:
          <volume>10</volume>
          .1145/1376616.1376746.
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>S.</given-names>
            <surname>Auer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Bizer</surname>
          </string-name>
          , G. Kobilarov,
          <string-name>
            <given-names>J.</given-names>
            <surname>Lehmann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Cyganiak</surname>
          </string-name>
          ,
          <string-name>
            <surname>Z. Ives,</surname>
          </string-name>
          <article-title>DBpedia: A Nucleus for a Web of Open Data</article-title>
          , in: ISWC,
          <year>2007</year>
          . URL: https://doi.org/10.1007/978-3-
          <fpage>540</fpage>
          -76298-0_
          <fpage>52</fpage>
          . doi:
          <volume>10</volume>
          .1007/ 978-3-
          <fpage>540</fpage>
          -76298-0_
          <fpage>52</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>J.</given-names>
            <surname>Hofart</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F. M.</given-names>
            <surname>Suchanek</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Berberich</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Lewis-Kelham</surname>
          </string-name>
          , G. de Melo, G. Weikum,
          <article-title>YAGO2: Exploring and Querying World Knowledge in Time, Space, Context, and Many Languages</article-title>
          , in: WWW,
          <year>2011</year>
          . URL: https://doi.org/10.1145/1963192.1963296. doi:
          <volume>10</volume>
          .1145/1963192.1963296.
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>F.</given-names>
            <surname>Petroni</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Rocktäschel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Lewis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Bakhtin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. H.</given-names>
            <surname>Miller</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Riedel</surname>
          </string-name>
          ,
          <article-title>Language Models as Knowledge Bases?</article-title>
          ,
          <year>2019</year>
          . URL: https://arxiv.org/abs/
          <year>1909</year>
          .01066. arXiv:
          <year>1909</year>
          .01066.
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>K.</given-names>
            <surname>Guu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Tung</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Pasupat</surname>
          </string-name>
          , M.-
          <string-name>
            <given-names>W.</given-names>
            <surname>Chang</surname>
          </string-name>
          , REALM:
          <string-name>
            <surname>Retrieval-Augmented Language Model Pre-Training</surname>
          </string-name>
          ,
          <year>2020</year>
          . URL: https://arxiv.org/abs/
          <year>2002</year>
          .08909. arXiv:
          <year>2002</year>
          .08909.
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>A.</given-names>
            <surname>Roberts</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Rafel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Shazeer</surname>
          </string-name>
          ,
          <article-title>How Much Knowledge Can You Pack Into the Parameters of a Language Model?</article-title>
          , in: B.
          <string-name>
            <surname>Webber</surname>
            , T. Cohn,
            <given-names>Y.</given-names>
          </string-name>
          <string-name>
            <surname>He</surname>
          </string-name>
          , Y. Liu (Eds.), EMNLP,
          <year>2020</year>
          . URL: https: //aclanthology.org/
          <year>2020</year>
          .emnlp-main.
          <volume>437</volume>
          /. doi:
          <volume>10</volume>
          .18653/v1/
          <year>2020</year>
          .emnlp-main.
          <volume>437</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>R. T.</given-names>
            <surname>McCoy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Pavlick</surname>
          </string-name>
          , T. Linzen,
          <article-title>Right for the Wrong Reasons: Diagnosing Syntactic Heuristics</article-title>
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>