<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>Corresponding author.
£ mmachado@ibm.com (M. Machado); joao.porto@ibm.com (J. P. P. Campos); guilherme.lima@ibm.com (G. Lima);
vivianet@br.ibm.com (V. T. da Silva)
Ȉ</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <article-id pub-id-type="doi">10.1145/1376616.1376746</article-id>
      <title-group>
        <article-title>KIF-QA: Using Of-the-shelf LLMs to Answer Simple Questions over Heterogeneous Knowledge Bases</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Marcelo Machado</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>João Pedro Porto Campos</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Guilherme Lima</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Viviane Torres da Silva</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>IBM Research</institution>
          ,
          <addr-line>Rio de Janeiro</addr-line>
          ,
          <country country="BR">Brazil</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2025</year>
      </pub-date>
      <volume>000</volume>
      <fpage>0</fpage>
      <lpage>0002</lpage>
      <abstract>
        <p>We present KIF-QA, a semantic parsing-based approach for answering simple questions over heterogeneous knowledge bases. KIF-QA uses of-the-shelf pre-trained large language models (LLMs) and in-context (few-shot) learning to transform questions into interpretable logical forms (queries) without requiring any fine-tuning. Because it uses KIF (the knowledge integration framework) to mediate all access to the underlying knowledge base, KIF-QA can be easily adapted to target any base accessible through KIF (which out-of-the-box includes Wikidata, DBpedia, PubChem, and others). We evaluate KIF-QA over the Wikidata and DBpedia versions of the SimpleQuestions benchmark using Llama 3.3, Llama 4 Maverick, and Mistral Medium 3. The results show competitive performance to comparable state-of-the-art methods. KIF-QA implementation is made available under an open-source license.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;knowledge base question answering</kwd>
        <kwd>large language model</kwd>
        <kwd>in-context learning</kwd>
        <kwd>Wikidata</kwd>
        <kwd>SPARQL</kwd>
        <kwd>KIF</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Knowledge base question answering (KBQA) is a classical task in artificial intelligence. Given a natural
language question, the goal of KBQA is to produce an answer supported by facts extracted from a
knowledge base [
        <xref ref-type="bibr" rid="ref1 ref2">1, 2</xref>
        ]. In the knowledge graph version of KBQA, the base is a knowledge graph, such as
Wikidata [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] or DBpedia [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ], and the answers are nodes (entities) or edges (statements or triples) in the
graph. For example, given the question “Where was James Brown born?”, a KBQA system may retrieve
and present as an answer the entity “Barnwell, SC, USA”, which in Wikidata would be wd:Q586262 and
in DBpedia dbr:Barnwell,_South_Carolina.
      </p>
      <p>
        The appeal of KBQA lies in making knowledge bases accessible to non-expert users. The user can
ask natural language questions directly to the system, without any prior understanding of the
knowledge base schema or its query language (e.g., in the case of knowledge graphs, without having to write
SPARQL [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] or Cypher [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] queries). A similar argument applies to non-human users. For instance, a
retrieval augmented generation (RAG) [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] agent can use KBQA to pose natural language questions
directly to a knowledge base. This way, the obtained answers are grounded and can be used as additional
context to improve the output of pre-trained (large) language models ((L)LMs) [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ].
      </p>
      <p>
        KBQA systems can be broadly classified according to the type of question they are designed to handle
(simple versus complex questions) and the method they use to produce the final answer (semantic
parsing versus information retrieval) [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. Simple questions such as “What is the capital of Brazil?”
involve a single fact (or hop) in the graph, while complex questions involve multiple hops or complex
operations (e.g., aggregation or arithmetic). In semantic parsing-based KBQA systems, answers are
produced by first transforming the natural language question into an logical form (query) and then
executing it over the graph. In contrast, in information retrieval-based systems, answers are obtained
by searching the graph directly, without employing an explicit query.
      </p>
      <p>We present KIF-QA, a semantic parsing-based KBQA approach that uses of-the-shelf LLMs to
answer simple questions over heterogeneous knowledge graphs.</p>
      <p>
        We focus on simple questions because they are easier to work with and, once solved, can be used as
a basis for solving complex questions iteratively [
        <xref ref-type="bibr" rid="ref10 ref11 ref9">9, 10, 11</xref>
        ]. Also, simple questions capture the essence
of the KBQA task, serving as a baseline for testing the applicability of LLMs to semantic parsing in
a training-free setting. Most methods for simple question answering, including the most successful
ones [
        <xref ref-type="bibr" rid="ref12 ref13 ref14 ref15">12, 13, 14, 15</xref>
        ], involve some form of training, which makes them dependent on the availability of
training datasets or couple them to the structure of the underlying graph. Our approach, on the other
hand, requires no prior knowledge of the graph schema and relies entirely on in-context (few-shot)
learning [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ] which means it involves neither training nor fine-tuning. All that is needed is access to
(i) a pre-trained language model; (ii) a standard query interface to the target knowledge graph (e.g.,
a SPARQL endpoint); and (iii) a search interface capable of looking up graph entities by label (e.g., a
keyword-based search service).
      </p>
      <p>
        The reason we chose to work with semantic parsing, i.e., producing an explicit query (logical form),
is twofold. First, compared to information retrieval, semantic parsing is more interpretable as the
produced logical form serves as an explanation for the answers [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. Second, semantic parsing can
be designed in such a way that the final logical form is obtained through successive refinements [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ],
starting with an abstract “draft query” and ending with a concrete query to be evaluated over the target
graph. This design avoids an early commitment with the particular query language and schema of the
target graph. In KIF-QA, we go one step further and delegate the problem of producing a final query
and eficiently evaluating it to a completely separate layer, called KIF.
      </p>
      <p>
        KIF [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ]1 is an open-source knowledge integration framework based on Wikidata. It uses a plugin
architecture to expose a unified interface to the graphs it integrates allowing us, for example, to query
DBpedia [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] or PubChem (RDF) [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ] as if they followed the data model (syntax) of Wikidata. KIF-QA
uses KIF to mediate all access to underlying knowledge graph, which makes our approach somewhat
independent of it. As long as there are KIF plugins to query and search the target graph, KIF-QA can
be instantiated over it to answer simple questions.
      </p>
      <p>Our contributions are as follows:
• An approach for semantic parsing that uses in-context (few-shot) learning over an of-the-shelf</p>
      <p>
        LLM to transform a natural language question into a set of KIF filters (queries to the KIF layer).
• An evaluation of our approach over the Wikidata and DBpedia versions of the
SimpleQuestions dataset [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ], namely, SimpleQuestions-Wikidata [
        <xref ref-type="bibr" rid="ref20">20</xref>
        ] and SimpleDBpediaQA [
        <xref ref-type="bibr" rid="ref21">21</xref>
        ], using
Llama 3.3 (70B Instruct) [
        <xref ref-type="bibr" rid="ref22">22</xref>
        ], Llama 4 Maverick (17Bx128E), and Mistral Medium 3 (20.25). The
results indicate that our approach achieves competitive performance against the state-of-the-art
for simple questions while requiring no training or fine-tuning and being easily adaptable to
work with any knowledge base accessible via KIF.
• An open-source implementation of our approach together with the datasets used to evaluate it,
which should ease the replication of our experiments.2
      </p>
      <p>The rest of the paper is organized as follows. Section 2 presents some background on semantic
parsing-based KBQA and KIF. Sections 3 and 4 present KIF-QA and its experimental evaluation.
Section 5 presents our conclusions and future work.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Background and Related Works</title>
      <sec id="sec-2-1">
        <title>2.1. Semantic Parsing-based KBQA and LLMs</title>
        <p>
          In semantic parsing-based KBQA, the system answers a natural language question by first translating it
into a logical form (query) and then executing the logical form over a knowledge base [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ]. A standard
1https://github.com/IBM/kif
2https://github.com/IBM/kif-llm
semantic parsing system typically performs three tasks: entity linking, relation linking, and query
generation. Entity linking and relation linking are the processes of identifying and linking mentions of
entities and relations in the question to corresponding entities and relations in knowledge base. These
tasks are sometimes referred to as alignment or grounding tasks and may involve sub-tasks such as
named entity recognition, candidate generation, and disambiguation. Query generation is the process
of generating an executable query capturing the intent of the question.
        </p>
        <p>
          Some semantic parsing systems follow an end-to-end design in the sense that they perform the three
tasks at once, using a single neural network model to generate a query directly from the question. A
recent example of such a system is SPARKLE [
          <xref ref-type="bibr" rid="ref15">15</xref>
          ] which, given a question, employs constrained
decoding within a single sequence-to-sequence model (a fine-tuned BART [
          <xref ref-type="bibr" rid="ref23">23</xref>
          ]) to generate a corresponding
SPARQL [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ] query. It is more common for semantic parsing-based KBQA systems to adopt a
multistage design though. In multi-stage systems, entity linking, relation linking, and query generation are
performed in stages, one after the other, and are sometimes delegated to specialized tools, such as the
popular entity linker BLINK [
          <xref ref-type="bibr" rid="ref24">24</xref>
          ].
        </p>
        <p>
          As we discuss in detail in Section 3, our system, KIF-QA, follows a multi-stage design. For entity and
relation linking, KIF-QA first uses KIF [
          <xref ref-type="bibr" rid="ref17">17</xref>
          ] to obtain candidate entities and relations in the namespace
of the target graph. Then it employs an of-the-shelf pre-trained (large) language model ((L)LM) with
in-context learning for disambiguation (selecting the best candidates). KIF-QA also uses in-context
(few-shot) learning to instruct the same LLM to generate a draft logical form from the natural language
question. Finally, this draft logical form together with the top-ranked candidate entities and relations
are used to generate the final queries (in our case, KIF queries).
        </p>
        <p>
          Many multi-stage systems also adopt pre-trained language models for query generation combined
with independent entity and relation linking modules. For instance, STaG-QA [
          <xref ref-type="bibr" rid="ref12">12</xref>
          ] uses a pre-trained
language model to generate a skeleton SPARQL query which is refined by resolving entities using
BLINK [
          <xref ref-type="bibr" rid="ref24">24</xref>
          ]. GETT-QA [
          <xref ref-type="bibr" rid="ref13">13</xref>
          ] uses a fine-tuned text-to-text model (T5 [
          <xref ref-type="bibr" rid="ref25">25</xref>
          ]) to generate a skeleton
SPARQL query which is refined by linking entities via a BM-25-based label search over the graph. Other
multi-stage systems that employ pre-trained language models for generating SPARQL are [
          <xref ref-type="bibr" rid="ref10 ref26 ref27">10, 26, 27</xref>
          ].
One of the few multi-stage systems that uses an LLM in a training-free setting (in-context learning) for
semantic parsing is KB-Binder [
          <xref ref-type="bibr" rid="ref28">28</xref>
          ]. KB-Binder uses examples (few-shot) from a training dataset when
available and disambiguate entities using a lexicon-based similarity search over the whole graph.
        </p>
        <p>
          Three things distinguish KIF-QA from the aforementioned multi-stage systems. First, while KIF-QA
uses of-the-shelf models and in-context (few-shot) learning, with the exception of KB-Binder, all of
the other systems require some form of training or fine-tuning. Second, while KIF-QA can be adapted
to work with any target graph (as long as it can be queried and searched through KIF), most of the
aforementioned systems target a particular graph or depend on direct access to the raw graph data
(e.g., to construct custom indexes). Third, while KIF-QA delegates to KIF the problem of eficiently
executing the generated query (e.g., the handling of pagination, parallelization of disjoint subqueries,
SPARQL dialect-specific optimizations), the aforementioned systems focus on the query generation
task and mostly ignore the pragmatics of SPARQL query execution in the real world.
2.2. KIF
KIF [
          <xref ref-type="bibr" rid="ref17">17</xref>
          ] is an open-source knowledge integration framework based on Wikidata [
          <xref ref-type="bibr" rid="ref3">3</xref>
          ]. It adopts the
data model3 and vocabulary of Wikidata as a lingua franca for integrating heterogeneous knowledge
sources. The integration done by KIF is virtual in the sense that the data model and vocabulary
translations happen at query time, guided by used-provided mappings.
        </p>
        <p>KIF is in essence a Python library (kif-lib4) for evaluating queries over engines and obtaining objects
of the Wikidata data model as a result. KIF engines are implemented following a plugin
architecture. There are various predefined (built-in) engine plugins which can be extended by applications
and loaded dynamically into the library. The two basic types of KIF engines are stores and searchers.</p>
        <sec id="sec-2-1-1">
          <title>3https://www.wikidata.org/wiki/Wikidata:Data_model 4https://pypi.org/project/kif-lib/</title>
          <p>
            Store. A KIF store is an interface to a knowledge source, typically but not necessarily a knowledge
graph. The most common type of KIF store is the SPARQL store, which reads Wikidata-like statements
from a given SPARQL endpoint. (SPARQL [
            <xref ref-type="bibr" rid="ref5">5</xref>
            ] is the query language of RDF [
            <xref ref-type="bibr" rid="ref29">29</xref>
            ], a standard format
for representing graph data.) To create a SPARQL store, one needs specify the address of the target
SPARQL endpoint and the SPARQL mappings to be used to generate a concrete SPARQL query (i.e.,
one adhering to the RDF encoding of the target graph). Besides mappings for Wikidata itself, KIF
comes with SPARQL mappings for DBpedia [
            <xref ref-type="bibr" rid="ref4">4</xref>
            ], PubChem (RDF) [
            <xref ref-type="bibr" rid="ref18">18</xref>
            ], UniProt [
            <xref ref-type="bibr" rid="ref30">30</xref>
            ], among others.
New mappings can be added programmatically.
          </p>
          <p>We can use the KIF command-line interface (CLI), which is distributed together with the library, to
run a simple query over a KIF store. For instance:
The command on line 1 runs a KIF filter (i.e., a statement match query) using the default store
plugin, wdqs, which targets the public Wikidata Query Service (WDQS) and uses KIF’s Wikidata SPARQL
mappings. This particular filter instructs KIF to search for statements whose subject is wd:Q7251 (Alan
Turing) and property is wd:P184 (doctoral advisor). To evaluate a filter, the SPARQL store uses the
provided mappings to compile it into a SPARQL query and then runs the query over the target endpoint.5
The resulting bindings are used to construct Wikidata-like statements. The returned statement in this
case is shown on line 2 and stands for the assertion “Alan Turing’s doctoral advisor is Alonzo Church”.</p>
          <p>In the Wikidata data model, which is adopted by KIF, every statement consists of a subject (which
can be an item or property) and a snak. The subject of the above statement is the item Alan Turing
and its snak is a value snak (i.e., property-value pair) with property equal to doctoral advisor and value
equal to the item Alonzo Church.</p>
          <p>KIF distinguishes between unannotated statements, such as the one above, and annotated statements.
The latter, besides a subject and a snak, also carry a set of qualifiers, a set of reference records, and a
rank. For instance, the filter below matches annotated statements (due to the --annotated flag) with
property wd:P2054 (density) and value equal to 0.88 grams per cubic centimeter:
The annotated statement shown on lines 4–12 asserts that “benzene’s density is 0.88±0.01 g/cm³”. Its
qualifier record (lines 5–7) qualifies this assertion, i.e., says in addition that this is the case when “the
temperature is 20±1 °C” and “the phase of matter is liquid”. Its reference record set (lines 8–11)
contains a single reference record (lines 9–11) which indicates the provenance of the statement, namely,
the entry with the given HSDB ID in the Hazardous Substances Data Bank. Finally, the value
NormalRank for its rank (line 12) is the default one and means “neutral” (neither preferred nor deprecated).
More complex filters. So far, we have used simple values to match the components of statements
in filters. In KIF, we can also specify the desired subject, property, or value indirectly, by giving a
constraint that it must satisfy. For instance:
5KIF handles query pagination and parallelization automatically. It also comes with an asyncio-compliant async API which
can be used to run queries asynchronously (without blocking waiting on their results).
This filter matches statements such that (i) the subject “ shares border with Argentina or Paraguay”
and “is located in or next to the Atlantic Ocean”; and (ii) the property is wd:P37 (oficial language).
The only matching subjects in this case are wd:Q155 (Brazil) and wd:Q77 (Uruguay) and their oficial
languages, as indicated by the statements above, are wd:Q5146 (Portuguese) and wd:Q1321 (Spanish).</p>
          <p>Besides single-hop checks, KIF supports constraints involving  -hop property paths. These can be
combined via conjunction (&amp;) and disjunction (|) operators, as shown above. Simple datatype checks
and projection on the statement components (subject, property, value) are also supported.
Targeting other graphs. To query a graph other than Wikidata, all we need to do is change the
store plugin. For instance:
$ kif filter --store=dbpedia --subject='db.r("Alan_Turing")'
↪ --value='(db.almaMater/db.city)(db.r("Princeton,_New_Jersey"))'
(Statement (Item dbr:Alan_Turing) (ValueSnak (Property dbo:doctoralAdvisor) (Item
↪ dbr:Alonzo_Church)))
(Statement (Item dbr:Alan_Turing) (ValueSnak (Property doctoral advisor) (Item
↪ dbr:Alonzo_Church)))
The --store=dbpedia argument instructs KIF to run the filter using the dbpedia store plugin, which
targets the public DBpedia SPARQL endpoint and uses KIF’s DBpedia SPARQL mappings. The
filter in this case matches statements whose subject is dbr:Alan_Turing and value is “someone whose
dbo:almaMater is located in the dbo:city of dbr:Princeton,_New_Jersey”.</p>
          <p>Despite querying DBpedia, as expected, we get as a result two Wikidata-like statements (lines 18–
19) asserting that “Alan Turing’s doctoral advisor is Alonzo Church”. We say these statements are
“Wikidata-like” because although following Wikidata syntax, their components mostly have DBpedia
URLs. This happens because KIF’s DBpedia SPARQL mappings do not attempt to convert entities in
the DBpedia namespace into that of Wikidata. The exception are properties. If there exists a Wikidata
property which is said to be the same as (owl:sameAs) a given DBpedia property in the DBpedia graph,
then the generated SPARQL query will match it. This is why on line 19 we get a statement whose
property is wd:P184 (doctoral advisor) even though we are querying DBpedia.</p>
          <p>Searchers. Besides stores, the other kind of engine in KIF are searchers. A KIF searcher is an interface
to a similarity search method within a namespace. For example, given a string, the wikidata-wapi
search plugin looks in the Wikidata namespace for entities with a similar label or alias:
$ kif search --search=wikidata-wapi 'pizza' --item
(Item pizza)
(Item Mariagrazia Pizza)
(Item Pizza Hut)
⋮
The --search=wikidata-wapi argument specifies the search plugin to be used, “pizza” is the search
string, and --item indicates that we want to search for items. (KIF, as Wikidata, distinguishes between
three types of entities: items, properties, and lexemes.) The top three items found by the search are
shown on lines 21–23.</p>
          <p>In addition to Wikidata, KIF comes with built-in plugins to search for entities in the namespaces of
DBpedia and PubChem.</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Proposal</title>
      <p>The KIF-QA approach consists of five main steps (see Figure 1).</p>
      <p>Step 1: Draft triple pattern generation.</p>
      <p>LLM to generate a draft triple pattern</p>
      <p>Given a simple question  in natural language, we ask the
(, , )</p>
      <p>using text labels for the known components ( -
or  - ) and the variable “?x” for the unknown component (either  or  ) (see Figure 2 (left)). We
provide examples of “(question, draft triple pattern)” pairs to the LLM when these are available.
Step 2: Entity linking. Given a draft triple pattern
(, , )
, we perform entity-linking on its topic
entity label, i.e., the known  or  . This is done through KIF by searching for entities with a
similar label or alias in the namespace of the target graph. The result is a set of candidate entities
accompanied by their descriptions. We then ask the LLM to disambiguate the topic entity label
against the candidate entities, i.e., select among the candidates those entities whose label and
description best match the topic entity label and the text of the question (see Figure 2 (right)).</p>
      <p>The result of this step is a refined set ℰ of candidate topic entities.</p>
      <p>Step 3: Relation linking. For resolving the predicate component  of the draft triple pattern (, , ) ,
we proceed as follows. For each candidate topic entity  in ℰ , we use KIF to obtain the candidate
relations   and   such that, for some  ,   (, )
or   (, )
in the graph. We then ask the LLM to
disambiguate the predicate label  against the label and description of each candidate relation   ,
the label and description of  , and the text of question  . The result of this step is, for each topic
entity  , a refined set</p>
      <p>ℛ of candidate relations for  . This set is constructed in a way that, for
each   in ℛ , we can always determine whether  occurs as source or target of   in the graph.
Steps 4 and 5: Filter generation and evaluation. For each candidate topic entity  in ℰ and
candidate relation   in ℛ , we generate a KIF filter  (, 
 , _) if  occurs as a source in   and
a KIF filter  ( _,   , )</p>
      <p>if  occurs as a target in   . We then execute all filters generated for all
candidate entities in parallel. The resulting statement set  is the answer set associated to the
original natural language question  .</p>
      <p>Step 1
Draft triple
pattern generation</p>
      <p>Step 2
Entity linking</p>
      <p>Step 3
Relation linking</p>
      <p>ℛ 1 , … , ℛ</p>
      <p>Step 4
Filter generation
 1, … ,</p>
      <p>Step 5
Filter evaluation</p>
      <p>(“James Brown”, “birth location”, “?x”)
ℛ 1 = {P19 (place of birth)}
ℛ 2 = {P19}</p>
      <p>⋮
ℛ  = {P19}
(Q5950, P19, Q586262 (Barnwell))
(Q6130327, P19, Q30 (USA))
⋮
(Q635131, P19, Q3705927 (Desdemona))
 = “Where was James Brown born?”
⋮
 1 = Q5950 (American musician)
 2 = Q6130327 (American biologist)
  = Q635131 (American actor)</p>
      <p>ℰ = { 1, … ,   }
⋮
 1 =  ( Q5950, P19, _)
 2 =  ( Q6130327, P19, _)
  =  ( Q635131, P19, _)</p>
      <p>A few remarks about the steps are in order.</p>
      <p>
        Step 1. KIF-QA assumes that the input question is a simple question whose corresponding draft triple
pattern has a single unknown component, either the subject or the object. This is a standard
assumption in simple-question KBQA [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. (Some systems and benchmarks adopt an even more restrictive
assumption, namely, that the unknown component is always the object [
        <xref ref-type="bibr" rid="ref19 ref31 ref32">19, 31, 32</xref>
        ].)
      </p>
      <p>When asking the LLM to transform the question into a draft triple pattern (see Figure 2 (left)), KIF-QA
can be provided with examples. These are question-draft triple pattern pairs, such as:</p>
      <p>When did World War II begin?
Where was Freddie Mercury born?</p>
      <p>Who painted Mona Lisa?
(“World War II”, “begin”, “?x”)
(“Freddie Mercury”, “born at”, “?x”)
(“?x”, “painted”, “Mona Lisa”)
The set of examples can be fixed (the same for all questions) or can be chosen from a pool of examples
relevant to the question at hand. In the evaluation of KIF-QA, discussed in Section 4, we compare the
performance of using a fixed set of examples versus a custom set chosen from a training dataset.
Steps 2 and 3. The disambiguation of the topic entity ( or  ) and the predicate ( ) of the draft
triple pattern (, , ) involves searching for entities and relations with similar labels or aliases in the
namespace of the target graph. KIF-QA does this through KIF by using an appropriate search plugin.
For instance, for searching in the Wikidata entity namespace (http://www.wikidata.org/entity/)
KIFQA uses the built-in KIF search plugin wikidata-wapi, which calls Wikidata’s public Wikibase API6.
Similarly, for searching in the DBpedia resource (http://dbpedia.org/resource/) and ontology (http://
dbpedia.org/ontology/) namespaces, KIF-QA uses the search plugins dbpedia-lookup and
dbpediaddgs. The former calls DBpedia’s public Lookup API7 to search for DBpedia resources, while the
latter uses the meta-search Python library DDGS8 to search in the DBpedia website for URLs matching
those of DBpedia ontology classes and properties. To make the disambiguation step work with a
diferent graph, all one needs to do is use a diferent KIF search plugin (one that supports searches
in the namespace of the target graph).</p>
      <p>
        After obtaining the candidate entities and relations, KIF-QA asks the LLM to pick the candidates
that best match the question (see Figure 2 (right)). A similar method is used in [
        <xref ref-type="bibr" rid="ref33">33</xref>
        ] to disambiguate
candidate entities within a certain textual context. Note that we could use a diferent method here.
A popular alternative is to use some embedding distance metric, such as [
        <xref ref-type="bibr" rid="ref34">34</xref>
        ], to compare the labels,
aliases, or descriptions of the candidates entities and relations to the text of the question.
      </p>
      <p>In Step 3, the candidate relations for a given topic entity  are obtained by querying the graph for
statements (edges) whose source or target entity is  . This is done through KIF by using an
appropriate store plugin. For querying Wikidata and DBpedia, KIF-QA uses the store plugins wdqs and
dbpedia which evaluate queries over the public Wikidata Query Service (WDQS)9 and the public DBpedia
SPARQL endpoint10, respectively. The query in this case is the union of the KIF filters
kb.filter_p(subject= )
and
kb.filter_p(value= ))
where kb is the target KIF store. The call kb.filter_p() looks for statements in kb matching the given
pattern (i.e., statements with subject  or value  ) and selects (projects) their property component,
hence the sufix “ _p”. By using a KIF here, instead of SPARQL, we delegate to KIF both the problem
of generating a concrete SPARQL query (i.e., one matching the RDF encoding of the target graph) and
the problem of executing it eficiently (see Section 2.2). As before, to target a diferent graph, all one
needs to do is instruct KIF-QA to use a diferent KIF store plugin—the filters remain the same.</p>
      <sec id="sec-3-1">
        <title>6https://www.wikidata.org/w/api.php</title>
        <p>7https://lookup.dbpedia.org/
8https://github.com/deedy5/ddgs
9https://query.wikidata.org/sparql
10https://dbpedia.org/sparql
[SYSTEM]
You are responsible for recognizing
incomplete subject-predicate-object triple
patterns from simple natural language
questions.
• Subjects are entities (e.g., people,
organizations, locations), and objects can be
either entities or literals (e.g., dates,
numbers).
• The predicate describes the relationship</p>
        <p>between the subject and the object.
• Each triple must have exactly one
un</p>
        <p>known element, represented as “?x”.
• Return a Python list of dictionaries,
where each dictionary contains exactly
one triple, represented as three string
values under the keys “subject”, “predicate”,
and “object”.
• Your output must adhere to valid Python
syntax. Do not include any extra
explanations or text.
(Examples go here)
[USER]
Where was James Brown born?
[SYSTEM]
You are a precise entity-linking assistant. Your task is to select the ID of
the candidate that unambiguously matches the target term in the sentence,
ensuring factual accuracy and semantic coherence. Rules:
1. Strict Matching: Only return a candidate ID if the context in the
sen</p>
        <p>tence explicitly aligns with the candidate’s description.
2. No Guessing: Do not infer or assume missing information. If the
sen</p>
        <p>tence lacks suficient context, return nothing.
3. Output Format:
• If there is a match, your response should be a list of comma
sepa</p>
        <p>rated IDs.
• If there is no clear match respond an empty string. Do not include</p>
        <p>any extra explanations.
(Examples go here)</p>
        <p>Steps 4 and 5. The same KIF store kb that was used in Step 3 to obtain the candidate relations is used
in Step 5 to execute (in parallel) the filters generated in Step 4. That is, the answer set  is obtained
by computing, for each  in ℰ and  in ℛ , the union of the following filters:
{
kb.filter(subject= , property= )
kb.filter(property= , value= )
if  is a source of  in the graph
if  is a target of  in the graph
Because we use the call kb.filter above,  will contain KIF statements. These are objects consisting
of a subject (item or property) and snak, i.e., a pair property-value where the value is an entity, shallow
data value (IRI, text, external id, or IRI), or deep data value (quantity or time).</p>
        <p>We could have used kb.filter_annotated in place of kb.filter above to obtain statements
together with their qualifiers, references, and rank (see Section 2.2). (In the case DBpedia, the default
built-in SPARQL mappings always set the rank to “normal” and do not generate qualifiers or
references.) Another possibility is using kb.filter_s (projection on subject) or kb.filter_v (projection
on value) in place of kb.filter to obtain just the unknown entity or value instead of whole statements.</p>
        <p>Note that because we use KIF we avoid the trouble of having to know how “statements” (which are
not necessarily RDF triples) and “entities” or “values” (which are not necessarily single RDF terms)
are represented in the RDF encoding of the target graph. We also do not need to worry about writing
a SPARQL query that will run eficiently on a particular triplestore. Wikidata’s public query service
uses Blazegraph11 while DBpedia’s uses Virtuoso RDF12. Although these two triplestores claim to be
standards compliant, in practice, they accept diferent dialects of SPARQL—subtle changes in the query
can afect the performance in unexpected ways. For instance, when targeting Blazegraph, the KIF’s
SPARQL compiler uses Blazegraph’s named subquery extension to enforce an evaluation order in
subqueries, which greatly speeds up some types of queries.
11https://blazegraph.com/
12https://virtuoso.openlinksw.com/</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Evaluation</title>
      <sec id="sec-4-1">
        <title>4.1. Dataset</title>
        <p>
          To evaluate KIF-QA, we use the Wikidata and DBpedia ports of the SimpleQuestions dataset [
          <xref ref-type="bibr" rid="ref19">19</xref>
          ].
The original SimpleQuestions dataset targeted Freebase [
          <xref ref-type="bibr" rid="ref35">35</xref>
          ] and consisted of simple questions paired
with the single triple pattern that answered them over Freebase. The Wikidata port of
SimpleQuestions is called SimpleQuestions-Wikidata [
          <xref ref-type="bibr" rid="ref20">20</xref>
          ] and the DBpedia port is called SimpleDBpediaQA [
          <xref ref-type="bibr" rid="ref21">21</xref>
          ].
For our evaluation, we selected 1,238 questions from the test subset of the answerable version of
SimpleQuestions-Wikidata and 1,222 from the test subset of SimpleDBpediaQA.13
        </p>
        <p>
          A known issue of the SimpleQuestions dataset is ambiguity, which stems from the method used to
create it [
          <xref ref-type="bibr" rid="ref31">31</xref>
          ]. Annotators were shown a single Freebase triple such as “(James Brown, birth place,
Beaumont)” and asked to write a simple question, e.g., “which city was james brown born”. The problem
here is that the link between the specific entity James Brown mentioned in the triple and the substring
“james brown” in the question is lost. Just from the question, without any other context, it is impossible
to determine the exact James Brown to which it refers. For instance, the subject of the gold triple for
this question in SimpleQuestions-Wikidata is wd:Q6130343 (James Brown), not the famous American
musician, but the American football quarterback born in 1975.
        </p>
        <p>Since our focus is measuring KIF-QA’s ability in providing accurate final answers, rather than solely
creating logical forms, to workaround this ambiguity problem, we extended the set of gold triples in
both SimpleQuestions-Wikidata and SimpleDBpediaQA to include all entities in the corresponding
knowledge graph whose label or alias matched exactly the label of the gold entity. For example, in
the “James Brown” case above, we included as gold entities every item in Wikidata whose label or
alias matched the label of wd:Q6130343. This means that in our curated versions of
SimpleQuestionsWikidata and SimpleDBpediaQA every question is paired not with a single gold triple but with a set of
one or more gold triples.</p>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. Models and Metrics</title>
        <p>We tested three diferent LLMs with varying scales and architectural approaches: (i) Llama 3.3 (70B
Instruct)14, a large instruction-tuned 70B-parameter model with strong general language understanding,
serving as a high-performance baseline for knowledge-intensive tasks; (ii) Llama 4 Maverick (17Bx128E)15,
a 17B-parameter multimodal model featuring a Mixture of Experts (MoE) architecture that activates 128
experts simultaneously; and (iii) Mistral Medium 3 (20.25)16, a smaller, 25B-parameter model optimized
for eficient inference, representing resource-constrained scenarios.</p>
        <p>The KIF-QA architecture allows the use of diferent models in diferent steps (draft triple pattern
generation, entity linking, and relation linking). But in the evaluation discussed here we used the
same model in all steps of a same run.</p>
        <p>We report both micro- and macro-averaged precision (P), recall (R), and F1 scores. Micro-averaging
aggregates results over all instances, while macro-averaging computes the metrics per question and
then averages them, giving equal weight to each question regardless of answer set size.</p>
      </sec>
      <sec id="sec-4-3">
        <title>4.3. Results</title>
        <p>
          Fixed-prompt baseline. The results on SimpleQuestions-Wikidata using a prompt with a fixed set
of examples (manually curated and always the same) are shown in Table 1. As can be seen, all models
achieved a high recall, indicating that the system retrieves many correct answers. The significant drop
in precision observed in the micro-averaged results is explained by a few anomalous questions which
produced a large number of false positives.
13https://huggingface.co/ibm-research/kif-qa
14https://github.com/meta-llama/llama-models/blob/main/models/llama3_3/MODEL_CARD.md
15https://github.com/meta-llama/llama-models/blob/main/models/llama4/MODEL_CARD.md
16https://mistral.ai/news/mistral-medium-3
Retrieval-augmented prompting. We also tested the performance of the models using a 5-shot
retrieval approach. For each question  , before running the KIF-QA pipeline we used a Sentence
Transformers model [
          <xref ref-type="bibr" rid="ref34">34</xref>
          ] to find questions similar to  in the training dataset of
SimpleQuestionsWikidata. We then selected the top-5 most similar questions together with their verbalized gold triple
patterns and used these as examples to the draft triple pattern generation step. To goal here was to
improve relation disambiguation. For instance, consider the question “which city was james brown
born”. One of the top-ranked examples retrieved from the training dataset for this question was:
        </p>
        <p>Where was charles grayson born? (“charles grayson”, “place of birth”, “?x”)
This example should influence the LLM to generate a draft triple pattern in which “james brown” occurs
in the subject and the predicate is “place of birth”, which matches exactly the label of the gold relation.</p>
        <p>Table 2 shows the results on SimpleQuestions-Wikidata using 5-shot retrieval. The dramatic
improvement in the micro-averaged precision values is explained by a significant reduction in the
number of false positives for all models. The improvement in the recall of Llama 3.3 and Mistral Medium 3
suggests consistent gains across diverse relation types.</p>
        <p>Single-answer questions. In an attempt to reduce the influence of ambiguity in the results, we
tested a variation of the 5-shot approach considering only single-answer questions, i.e., questions
whose gold answer set consists of a single gold triple. For example, questions about specific subjects
like “who published neo contra” are usually associated to exactly one triple in the graph. (Contrast this
with the question “Who was born in san diego”.) By focusing on single-answer questions, we can
evaluate the model performance in a context where the influence of disambiguation errors is minimized.</p>
        <p>The results on SimpleQuestions-Wikidata using 5-shot retrieval considering only single-answer
questions are shown in Table 3. As can be seen, all models maintained a high recall, with Llama 4
Maverick achieving the highest micro precision and Llama 3.3 leading in macro F1. These results
suggest that less ambiguous questions tend to generate fewer false positives. In other words, when the
question is associated with a small number of triples in the knowledge base, the models seem to be
better at both identifying the correct answer and avoiding extraneous or incorrect ones. This suggests
that question ambiguity plays a significant role in KBQA performance and that precision gains can be
achieved by narrowing the scope of candidate answers.</p>
        <p>DBpedia. Finally, to demonstrate the generality KIF-QA across knowledge graphs, we tested it on
SimpleDBpediaQA using a 5-shot retrieval configuration. The results are shown in Table 4. As can
be seen, Llama 4 Maverick achieved the highest micro F1 and Llama 3.3 the highest macro F1, while
Mistral Medium 3 showed competitive results. Overall the numbers obtained on SimpleDBpediaQA are
slightly lower than those obtained on SimpleQuestions-Wikidata using a similar 5-shot configuration.</p>
      </sec>
      <sec id="sec-4-4">
        <title>4.4. Classes of Failures</title>
        <p>In light of the above results, we sought to identify classes of failures by analyzing cases in which KIF-QA
failed by either producing wrong answers or no answer at all.</p>
        <p>The first class of failures consists of those caused not by a problem in KIF-QA but by issues in the
SimpleQuestions-Wikidata and SimpleDBpediaQA datasets. For example, the question “where was
what play did born” is agrammatical and its associated gold triple is completely unrelated:
(“Gabrielle Roy”, “place of birth”, “?x”)</p>
        <p>The second class of failures consists of those caused by unconstrained logical forms. These are
usually generated when KIF-QA drops a constraint from the question at the first step of the pipeline
(draft triple pattern generation). For instance, consider the question “who is someone famous with
attention deficit hyperactivity disorder”. Using Llama 3.3 and given this question, KIF-QA generates
the following draft triple pattern:</p>
        <p>(“?x”, “diagnosed with”, “Attention Deficit Hyperactivity Disorder”)
This pattern does not capture the constraint that “?x” must be someone famous. Because of that, when
evaluated over the knowledge base, the corresponding query will produce a large number of false
positives.</p>
        <p>Related to this issue, we observed that in many cases Llama 4 Maverick and Mistral Medium 3 failed
to generate the draft triple pattern precisely because they tried to constrain it using more than one
triple pattern. For instance, for the question “What swedish city was gerard de geer born in?”, Mistral
Medium 3 generated two draft triple patterns:
(“Gerard De Geer”, “place of birth”, “?x”)
and
(“?x”, “country”, “Sweden”)
Note that this a clear indication that in-context (few-shot) learning could be used to instruct these
models to tackle questions requiring more than one hop.</p>
        <p>Failures of the third class are those caused by model hallucinations. For example, Llama 3.3
misrecognized “freekstyle” as “freakstyle” in the question “who published the game freekstyle”, making it
impossible to retrieve the correct entity from the knowledge base.</p>
        <p>Finally, failures of the fourth class are those caused by problems in entity or relation linking. In some
cases, the search returned no candidate for the topic entity label of the question, while in other cases
the returned candidates had labels which were unrelated to that of the topic entity. For instance, for the
question “where is sheila larken from”, the search returned the candidate entity wd:Q3481761, whose
label is “Marteen Huell” and description is “American television actress”. Because of the mismatch of
surface forms (“sheila larken” versus “Marteen Huell”) the LLM failed to disambiguate the topic entity.
One way to mitigate this problem is by using not only the labels and descriptions of the candidate
entities but also their aliases in the disambiguation step.</p>
      </sec>
      <sec id="sec-4-5">
        <title>4.5. Remarks</title>
        <p>The results presented in this section show that even without fine-tuning KIF-QA achieves a competitive
performance on the SimpleQuestions-Wikidata and SimpleDBpediaQA datasets, which demonstrates
the flexibility and robustness of its modular, prompt-based design. This robustness is notable given the
inconsistencies, missing or mislabeled entities, and unconstrained questions that make 100% accuracy
in these datasets unattainable. Many of the failures we discussed above stem from factors external to
the LLMs, which suggests that simple extensions like handling constraints and employing aliases in
disambiguation could yield immediate gains.</p>
        <p>The results also demonstrate that performance is improved when training data is incorporated in
the draft triple pattern generation step. High recall values were obtained consistently in diferent
settings. This suggests that the approach is particularly suitable for applications where coverage is more
important than perfect precision, i.e., applications such as exploratory search, knowledge discovery,
or open-domain question answering. Finally, the competitive performance with smaller models like
Mistral Medium 3 shows that KIF-QA can be employed efectively even in resource-constrained
environments, broadening its applicability to real-world scenarios where eficiency and scalability are
essential.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Conclusion</title>
      <p>In this paper, we presented a semantic parsing-based KBQA approach, called KIF-QA, that uses
ofthe-shelf LLMs to answer simple question over heterogeneous knowledge graphs. Our evaluations
showed that it is possible to obtain competitive performance (comparable to the state-of-the-art) by
employing only in-context (few-shot) learning, without any fine-tuning. Because in the design of
KIFQA we separate the knowledge base-independent part of semantic-parsing, namely, generating the
draft logical form (step 1 of our pipeline), from the entity linking, relation linking, and query generation
parts (steps 2–4), and because we use KIF (the knowledge integration framework) to mediate all access
to the underlying knowledge base, our approach can be easily adapted to work with any knowledge
base accessible through KIF. This flexibility is demonstrated by our evaluations of the Wikidata and
DBpedia versions of SimpleQuestions in which the change from Wikidata to DBpedia involved only
instantiating KIF-QA with a diferent set of KIF plugins.</p>
      <p>In future work, we will explore extensions of the approach to simple questions involving constraints,
such as “What Italian dish includes cheese as an ingredient?”. To answer questions such as this one we
need to be able to constraint the subject to “Italian dishes”. This type of constraint is already supported
by KIF filters and our informal experiments have shown that it is possible to use in-context learning
to instruct LLMs to generate them.</p>
    </sec>
    <sec id="sec-6">
      <title>Declaration on Generative AI</title>
      <p>The authors have not used any generative AI tools to write this paper.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>J.</given-names>
            <surname>Berant</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Chou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Frostig</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Liang</surname>
          </string-name>
          ,
          <article-title>Semantic parsing on Freebase from question-answer pairs</article-title>
          ,
          <source>in: Proc. 2013 Conf. Empirical Methods in Natural Language Processing</source>
          , Seattle, Washington, USA,
          <fpage>18</fpage>
          -
          <lpage>21</lpage>
          October,
          <year>2013</year>
          , ACL,
          <year>2013</year>
          , pp.
          <fpage>1533</fpage>
          -
          <lpage>1544</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Lan</surname>
          </string-name>
          , G. He,
          <string-name>
            <given-names>J.</given-names>
            <surname>Jiang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Jiang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W. X.</given-names>
            <surname>Zhao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.-R.</given-names>
            <surname>Wen</surname>
          </string-name>
          ,
          <article-title>Complex knowledge base question answering: A survey</article-title>
          ,
          <source>IEEE Trans. Know. Data Eng</source>
          .
          <volume>35</volume>
          (
          <year>2023</year>
          )
          <fpage>11196</fpage>
          -
          <lpage>11215</lpage>
          . doi:
          <volume>10</volume>
          .1109/TKDE.
          <year>2022</year>
          .
          <volume>3223858</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>D.</given-names>
            <surname>Vrandečić</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Krötzsch</surname>
          </string-name>
          ,
          <article-title>Wikidata: A free collaborative knowledgebase</article-title>
          ,
          <source>Commun. ACM</source>
          <volume>57</volume>
          (
          <year>2014</year>
          )
          <fpage>78</fpage>
          -
          <lpage>85</lpage>
          . doi:
          <volume>10</volume>
          .1145/2629489.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>J.</given-names>
            <surname>Lehmann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Isele</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Jakob</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Jentzsch</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Kontokostas</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. N.</given-names>
            <surname>Mendes</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Hellmann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Morsey</surname>
          </string-name>
          , P. van Kleef,
          <string-name>
            <given-names>S.</given-names>
            <surname>Auer</surname>
          </string-name>
          , C. Bizer, DBpedia
          <article-title>- a large-scale, multilingual knowledge base extracted from Wikipedia</article-title>
          ,
          <source>Semant. Web</source>
          <volume>6</volume>
          (
          <year>2015</year>
          )
          <fpage>167</fpage>
          -
          <lpage>195</lpage>
          . doi:
          <volume>10</volume>
          .3233/SW- 140134.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>W3C</given-names>
            <surname>SPARQL Working</surname>
          </string-name>
          <string-name>
            <surname>Group</surname>
          </string-name>
          ,
          <source>SPARQL 1.1 Overview</source>
          , W3C Recommendation,
          <year>W3C</year>
          ,
          <year>2013</year>
          . URL: http://www.w3.org/TR/2013/REC-sparql11
          <string-name>
            <surname>-</surname>
          </string-name>
          overview-20130321/.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>N.</given-names>
            <surname>Francis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Green</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Guagliardo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Libkin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Lindaaker</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Marsault</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Plantikow</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Rydberg</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Selmer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Taylor</surname>
          </string-name>
          , Cypher:
          <article-title>An evolving query language for property graphs</article-title>
          ,
          <source>in: Proc. 2018 Int. Conf. Management of Data, SIGMOD '18</source>
          ,
          <string-name>
            <surname>ACM</surname>
          </string-name>
          ,
          <year>2018</year>
          , pp.
          <fpage>1433</fpage>
          -
          <lpage>1445</lpage>
          . doi:
          <volume>10</volume>
          .1145/3183713. 3190657.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>P.</given-names>
            <surname>Lewis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Perez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Piktus</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Petroni</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Karpukhin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Goyal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Küttler</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Lewis</surname>
          </string-name>
          , W. tau Yih, T. Rocktäschel,
          <string-name>
            <given-names>S.</given-names>
            <surname>Riedel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Kiela</surname>
          </string-name>
          ,
          <article-title>Retrieval-augmented generation for knowledge-intensive NLP tasks</article-title>
          ,
          <source>in: Proc. 34th Int. Conf. Neural Information Processing Systems</source>
          , NIPS '20, Curran Associates Inc.,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>W.</given-names>
            <surname>Fan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Ding</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Ning</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Yin</surname>
          </string-name>
          , T.-S. Chua,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <article-title>A survey on RAG meeting LLMs: Towards retrieval-augmented large language models</article-title>
          ,
          <source>in: Proc. 30th ACM SIGKDD Conf. Knowledge Discovery and Data Mining, KDD '24</source>
          ,
          <string-name>
            <surname>ACM</surname>
          </string-name>
          ,
          <year>2024</year>
          , pp.
          <fpage>6491</fpage>
          -
          <lpage>6501</lpage>
          . doi:
          <volume>10</volume>
          .1145/3637528.3671470.
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>W.</given-names>
            <surname>Xiong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X. L.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Iyer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Du</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Lewis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W. Y.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Mehdad</surname>
          </string-name>
          , W. tau Yih,
          <string-name>
            <given-names>S.</given-names>
            <surname>Riedel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Kiela</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Oğuz</surname>
          </string-name>
          ,
          <article-title>Answering complex open-domain questions with multi-hop dense retrieval</article-title>
          ,
          <year>2021</year>
          . arXiv:
          <year>2009</year>
          .12756.
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>X.</given-names>
            <surname>Huang</surname>
          </string-name>
          , S. Cheng, Y. Shu,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Bao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Qu</surname>
          </string-name>
          ,
          <article-title>Question decomposition tree for answering complex questions over knowledge bases</article-title>
          ,
          <source>in: Proc. AAAI Conf. Artificial Intelligence</source>
          , volume
          <volume>37</volume>
          ,
          <year>2023</year>
          , pp.
          <fpage>12924</fpage>
          -
          <lpage>12932</lpage>
          . doi:
          <volume>10</volume>
          .1609/aaai.v37i11.
          <fpage>26519</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>H.</given-names>
            <surname>Rajabzadeh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H. J.</given-names>
            <surname>Kwon</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <article-title>Multimodal multi-hop question answering through a conversation between tools and eficiently finetuned large language models</article-title>
          ,
          <year>2023</year>
          . arXiv:
          <volume>2309</volume>
          .
          <fpage>08922</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>S.</given-names>
            <surname>Ravishankar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Thai</surname>
          </string-name>
          , I. Abdelaziz,
          <string-name>
            <given-names>N.</given-names>
            <surname>Mihindukulasooriya</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Naseem</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Kapanipathi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Rossiello</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Fokoue</surname>
          </string-name>
          ,
          <article-title>A two-stage approach towards generalization in knowledge base question answering</article-title>
          ,
          <source>in: Findings of the ACL: EMNLP</source>
          <year>2022</year>
          , ACL,
          <year>2022</year>
          , pp.
          <fpage>5571</fpage>
          -
          <lpage>5580</lpage>
          . doi:
          <volume>10</volume>
          .18653/ v1/
          <year>2022</year>
          .findings- emnlp.408.
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>D.</given-names>
            <surname>Banerjee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. A.</given-names>
            <surname>Nair</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Usbeck</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Biemann</surname>
          </string-name>
          , GETT-QA:
          <article-title>Graph embedding based T2T transformer for knowledge graph question answering</article-title>
          ,
          <source>in: The Semantic Web: 20th Int. Conf., ESWC</source>
          <year>2023</year>
          , Hersonissos, Crete, Grece, May 
          <fpage>28</fpage>
          -June 1,
          <year>2023</year>
          , Proceedings, Springer,
          <year>2023</year>
          , pp.
          <fpage>279</fpage>
          -
          <lpage>297</lpage>
          . doi:
          <volume>10</volume>
          .1007/978- 3-
          <fpage>031</fpage>
          - 33455- 9.
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>Y.-H.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E. J.-L.</given-names>
            <surname>Lu</surname>
          </string-name>
          , K.-H. Cheng,
          <article-title>Enhancing SPARQL query generation for question answering with a hybrid encoder-decoder and cross-attention model</article-title>
          ,
          <source>J. Web Semant</source>
          .
          <volume>87</volume>
          (
          <year>2025</year>
          )
          <article-title>100869</article-title>
          . doi:
          <volume>10</volume>
          .1016/j.websem.
          <year>2025</year>
          .
          <volume>100869</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>J.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Shin</surname>
          </string-name>
          ,
          <article-title>SPARKLE: Enhancing SPARQL generation with direct KG integration in decoding</article-title>
          ,
          <source>Expert Syst. Appl</source>
          .
          <volume>289</volume>
          (
          <year>2025</year>
          )
          <article-title>128263</article-title>
          . doi:
          <volume>10</volume>
          .1016/j.eswa.
          <year>2025</year>
          .
          <volume>128263</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <surname>T. B. Brown</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Mann</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          <string-name>
            <surname>Ryder</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Subbiah</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Kaplan</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Dhariwal</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Neelakantan</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Shyam</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          <string-name>
            <surname>Sastry</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Askell</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Agarwal</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Herbert-Voss</surname>
            , G. Krueger,
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Henighan</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Child</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Ramesh</surname>
            ,
            <given-names>D. M.</given-names>
          </string-name>
          <string-name>
            <surname>Ziegler</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Wu</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Winter</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Hesse</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Chen</surname>
            , E. Sigler,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Litwin</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Gray</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Chess</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Clark</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Berner</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>McCandlish</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Radford</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          <string-name>
            <surname>Sutskever</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Amodei</surname>
          </string-name>
          ,
          <article-title>Language models are few-shot learners</article-title>
          ,
          <source>in: Proc. 34th Int. Conf. Neural Information Processing Systems</source>
          , NIPS '20, Curran Associates Inc.,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>G.</given-names>
            <surname>Lima</surname>
          </string-name>
          ,
          <string-name>
            <surname>J. M. B. Rodrigues</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Machado</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          <string-name>
            <surname>Soares</surname>
            ,
            <given-names>S. R.</given-names>
          </string-name>
          <string-name>
            <surname>Fiorini</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Thiago</surname>
            ,
            <given-names>L. G.</given-names>
          </string-name>
          <string-name>
            <surname>Azevedo</surname>
          </string-name>
          , V. T. da
          <string-name>
            <surname>Silva</surname>
          </string-name>
          , R. Cerqueira,
          <article-title>KIF: A Wikidata-based framework for integrating heterogeneous knowledge sources</article-title>
          ,
          <year>2024</year>
          . arXiv:
          <volume>2403</volume>
          .
          <fpage>10304</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>S.</given-names>
            <surname>Kim</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Chen</surname>
          </string-name>
          , T. Cheng, A. Gindulyte,
          <string-name>
            <given-names>J.</given-names>
            <surname>He</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>He</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B. A.</given-names>
            <surname>Shoemaker</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. A.</given-names>
            <surname>Thiessen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Yu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Zaslavsky</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , E. E. Bolton,
          <article-title>PubChem 2023 update</article-title>
          ,
          <source>Nucleic Acids Res</source>
          .
          <volume>51</volume>
          (
          <year>2023</year>
          )
          <fpage>D1373</fpage>
          -
          <lpage>D1380</lpage>
          . doi:
          <volume>10</volume>
          .1093/nar/gkac956.
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>A.</given-names>
            <surname>Bordes</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Usunier</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Chopra</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Weston</surname>
          </string-name>
          ,
          <article-title>Large-scale simple question answering with memory networks</article-title>
          ,
          <year>2015</year>
          . arXiv:
          <fpage>1506</fpage>
          .
          <year>02075</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>D.</given-names>
            <surname>Diefenbach</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T. P.</given-names>
            <surname>Tanon</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K. D.</given-names>
            <surname>Singh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Maret</surname>
          </string-name>
          ,
          <article-title>Question answering benchmarks for Wikidata, in: Proc</article-title>
          .
          <article-title>ISWC 2017 Posters, Demonstrations, and Industry Tracks co-located with 16th International Semantic Web Conference (ISWC</article-title>
          <year>2017</year>
          ), Vienna, Austria,
          <source>October 23-25</source>
          ,
          <year>2017</year>
          , Springer,
          <year>2017</year>
          . URL: https://ceur-ws.
          <source>org/</source>
          Vol-1963/paper555.pdf.
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>M.</given-names>
            <surname>Azmy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Shi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Lin</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Ilyas</surname>
          </string-name>
          , Farewell Freebase:
          <article-title>Migrating the SimpleQuestions dataset to DBpedia</article-title>
          , in
          <source>: Proc. 27th Int. Conf. Computational Linguistics</source>
          ,
          <string-name>
            <surname>ACL</surname>
          </string-name>
          ,
          <year>2018</year>
          , pp.
          <fpage>2093</fpage>
          -
          <lpage>2103</lpage>
          . URL: https://aclanthology.org/C18-1178/.
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>H.</given-names>
            <surname>Touvron</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Lavril</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Izacard</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Martinet</surname>
          </string-name>
          , M.
          <article-title>-</article-title>
          <string-name>
            <surname>A. Lachaux</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Lacroix</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Rozière</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          <string-name>
            <surname>Goyal</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          <string-name>
            <surname>Hambro</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Azhar</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Rodriguez</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Joulin</surname>
          </string-name>
          , E. Grave, G. Lample,
          <source>LLaMA: Open and eficient foundation language models</source>
          ,
          <year>2023</year>
          . arXiv:
          <volume>2302</volume>
          .
          <fpage>13971</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <given-names>M.</given-names>
            <surname>Lewis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Goyal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Ghazvininejad</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Mohamed</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Levy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Stoyanov</surname>
          </string-name>
          , L. Zettlemoyer, BART:
          <article-title>Denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension</article-title>
          ,
          <source>in: Proc. 58th Annual Meeting of the ACL, ACL</source>
          ,
          <year>2020</year>
          , pp.
          <fpage>7871</fpage>
          -
          <lpage>7880</lpage>
          . doi:
          <volume>10</volume>
          . 18653/v1/
          <year>2020</year>
          .acl- main.703.
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [24]
          <string-name>
            <given-names>L.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Petroni</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Josifoski</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Riedel</surname>
          </string-name>
          , L. Zettlemoyer,
          <article-title>Scalable zero-shot entity linking with dense entity retrieval</article-title>
          ,
          <source>in: Proc. 2020 Conf. Empirical Methods in Natural Language Processing</source>
          , ACL,
          <year>2020</year>
          , pp.
          <fpage>6397</fpage>
          -
          <lpage>6407</lpage>
          . doi:
          <volume>10</volume>
          .18653/v1/
          <year>2020</year>
          .emnlp- main.519.
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          [25]
          <string-name>
            <given-names>C.</given-names>
            <surname>Rafel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Shazeer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Roberts</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Narang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Matena</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. J.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <article-title>Exploring the limits of transfer learning with a unified text-to-text transformer</article-title>
          ,
          <source>J. Mach. Learn. Res</source>
          .
          <volume>21</volume>
          (
          <year>2020</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          [26]
          <string-name>
            <given-names>R.</given-names>
            <surname>Omar</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Dhall</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Kalnis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Mansour</surname>
          </string-name>
          ,
          <article-title>A universal question-answering platform for knowledge graphs</article-title>
          ,
          <source>Proc. ACM Manag. Data</source>
          <volume>1</volume>
          (
          <year>2023</year>
          ). doi:
          <volume>10</volume>
          .1145/3588911.
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          [27]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Hua</surname>
          </string-name>
          , G. Qi,
          <article-title>Formal query building with query structure prediction for complex question answering over knowledge base</article-title>
          ,
          <source>in: Proc. Twenty-Ninth Int. Joint Conf. on Artificial Intelligence, IJCAI '20</source>
          ,
          <year>2021</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          [28]
          <string-name>
            <given-names>T.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Ma</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Zhuang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Gu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Su</surname>
          </string-name>
          , W. Chen,
          <article-title>Few-shot in-context learning on knowledge base question answering</article-title>
          ,
          <source>in: Proc. 61st Annual Meeting of the ACL (Volume</source>
          <volume>1</volume>
          :
          <string-name>
            <surname>Long</surname>
            <given-names>Papers)</given-names>
          </string-name>
          ,
          <source>ACL</source>
          ,
          <year>2023</year>
          , pp.
          <fpage>6966</fpage>
          -
          <lpage>6980</lpage>
          . doi:
          <volume>10</volume>
          .18653/v1/
          <year>2023</year>
          .acl- long.385.
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          [29] W3C RDF Working Group,
          <article-title>Resource Description Framework (RDF): Concepts and Abstract Syntax</article-title>
          ,
          <source>W3C Recommendation, W3C</source>
          ,
          <year>2014</year>
          . URL: https://www.w3.org/TR/rdf10-concepts/.
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          [30]
          <article-title>The UniProt Consortium</article-title>
          ,
          <source>UniProt: The Universal Protein Knowledgebase in 2025, Nucleic Acids Res</source>
          .
          <volume>53</volume>
          (
          <year>2024</year>
          )
          <fpage>D609</fpage>
          -
          <lpage>D617</lpage>
          . doi:
          <volume>10</volume>
          .1093/nar/gkae1010.
        </mixed-citation>
      </ref>
      <ref id="ref31">
        <mixed-citation>
          [31]
          <string-name>
            <given-names>M.</given-names>
            <surname>Petrochuk</surname>
          </string-name>
          , L. Zettlemoyer,
          <article-title>SimpleQuestions nearly solved: A new upperbound and baseline approach</article-title>
          ,
          <source>in: Proc. 2018 Conf. Empirical Methods in Natural Language Processing</source>
          , Brussels, Belgium,
          <source>October 31-November 4</source>
          ,
          <year>2018</year>
          , ACL,
          <year>2018</year>
          , pp.
          <fpage>554</fpage>
          -
          <lpage>558</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref32">
        <mixed-citation>
          [32]
          <string-name>
            <given-names>C.</given-names>
            <surname>Badenes-Olmedo</surname>
          </string-name>
          ,
          <string-name>
            <surname>O. Corcho,</surname>
          </string-name>
          <article-title>MuHeQA: Zero-shot question answering over multiple and heterogeneous knowledge bases</article-title>
          ,
          <source>Semant. Web</source>
          <volume>15</volume>
          (
          <year>2024</year>
          )
          <fpage>1547</fpage>
          -
          <lpage>1561</lpage>
          . doi:
          <volume>10</volume>
          .3233/SW- 233379.
        </mixed-citation>
      </ref>
      <ref id="ref33">
        <mixed-citation>
          [33]
          <string-name>
            <given-names>M.</given-names>
            <surname>Machado</surname>
          </string-name>
          ,
          <string-name>
            <surname>J. M. B. Rodrigues</surname>
            , G. Lima,
            <given-names>S. R.</given-names>
          </string-name>
          <string-name>
            <surname>Fiorini</surname>
          </string-name>
          , V. T. da Silva, LLM Store:
          <article-title>Leveraging large language models as sources of Wikidata-structured knowledge</article-title>
          ,
          <source>in: Joint Proc. 2nd Workshop</source>
          on
          <article-title>Knowledge Base Construction from Pre-Trained Language Models (KBC-LM 2024) and the 3rd Challenge on Language Models for Knowledge Base Construction (LM-KBC 2024) co-located with the 23nd International Semantic Web Conference (ISWC</article-title>
          <year>2024</year>
          ), Baltimore, USA, November
          <volume>12</volume>
          ,
          <year>2024</year>
          , CEUR-WS.org,
          <year>2023</year>
          . URL: https://ceur-ws.
          <source>org/</source>
          Vol-
          <volume>3853</volume>
          /paper6.pdf.
        </mixed-citation>
      </ref>
      <ref id="ref34">
        <mixed-citation>
          [34]
          <string-name>
            <given-names>N.</given-names>
            <surname>Reimers</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Gurevych</surname>
          </string-name>
          , Sentence-BERT:
          <article-title>Sentence embeddings using Siamese BERT-networks</article-title>
          ,
          <source>in: Proc. 2019 Conf. Empirical Methods in Natural Language Processing</source>
          , ACL,
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref35">
        <mixed-citation>
          [35]
          <string-name>
            <given-names>K.</given-names>
            <surname>Bollacker</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Evans</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Paritosh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Sturge</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Taylor</surname>
          </string-name>
          , Freebase:
          <article-title>A collaboratively created graph</article-title>
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>