<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <issn pub-type="ppub">1613-0073</issn>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>with Schema Filtering</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Makbule Gulcin Ozsoy</string-name>
          <email>makbule.ozsoy@neo4j.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="editor">
          <string-name>Text2Cypher, Large Language Model, Schema Filtering</string-name>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Neo4j</institution>
          ,
          <addr-line>London</addr-line>
          ,
          <country country="UK">UK</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Knowledge graphs represent complex data using nodes, relationships, and properties. Cypher, a powerful query language for graph databases, enables eficient modeling and querying. Recent advancements in large language models allow translation of natural language questions into Cypher queries-Text2Cypher. A common approach is incorporating database schema into prompts. However, complex schemas can introduce noise, increase hallucinations, and raise computational costs. Schema filtering addresses these challenges by including only relevant schema elements, improving query generation while reducing token costs. This work explores various schema filtering methods for Text2Cypher task and analyzes their impact on token length, performance, and cost. Results show that schema filtering efectively optimizes Text2Cypher, especially for smaller models. Consistent with prior research, we find that larger models benefit less from schema filtering due to their longer context capabilities. However, schema filtering remains valuable for both larger and smaller models in cost reduction.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Databases are an essential part of modern computer systems for storing and managing data. They are
typically accessed via query languages like SQL (for relational databases), SPARQL (for RDF graphs) or
Cypher (for graph databases) which allow users to store and query data for insight [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. Advancements
in LLMs have enabled the translation of natural language questions into database queries (Text2SQL,
Text2SPARQL, Text2Cypher), allowing non-expert users to query data models on their own terms.
      </p>
      <p>
        To help contextualize an LLM when generating database queries from natural language, a common
practice is to incorporate database schema information. Figure 1 shows an example schema where nodes
(e.g., Organization, Person) connect through relations (e.g., Has_CEO, Has_Investor) with their properties
(e.g., name, age). Schemas can be provided to LLMs via prompting, but complex schemas introduce
noise, increase hallucinations, and raise costs [
        <xref ref-type="bibr" rid="ref2 ref3">2, 3</xref>
        ]. Schema filtering addresses these challenges by
selecting only relevant elements, improving query generation while reducing token costs.
      </p>
      <p>In this paper, we apply five schema linking and filtering approaches that improve Text2Cypher: Two
static methods that extract the full database schema in diferent formats and three dynamic methods
that prune the schema based on the input question. We evaluate their impact on a Text2Cypher dataset,
analyzing token distribution, Cypher generation performance, and cost. Our main contributions are:
• We propose new schema filtering techniques. The two static methods use the full database schema
in diferent formats, while our three dynamic methods prune it based on the input question.
• We analyze their impact on Text2Cypher task, specifically on prompt token length distribution,
query generation performance, and computational cost.
• Our results show that schema filtering improves Text2Cypher eficiency. While larger models
benefit less due to their extended context windows, smaller models perform better with shorter
prompts. Nevertheless, schema filtering remains a cost-efective strategy for all models.</p>
      <p>The paper is structured as follows: Section 2 covers related work, and Section 3 details our
schemaifltering approaches for the Text2Cypher task. Section
4 presents our experiments and results, and
Section 5 concludes the paper.</p>
      <p>LLM-TEXT2KG 2025: 4th International Workshop on LLM-Integrated Knowledge Graph Generation from Text (Text2KG), June 1</p>
      <p>CEUR</p>
      <p>ceur-ws.org</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related Work</title>
      <sec id="sec-2-1">
        <title>2.1. Natural Language to Database Query Language</title>
        <p>
          Recent advances in large language models (LLMs) have significantly improved the ability to translate
natural language into database query languages. For instance, there has been extensive research on
the Text2SQL and Text2SPARQL tasks, which translates natural language queries to SQL or SPARQL,
respectively [
          <xref ref-type="bibr" rid="ref10 ref11 ref12 ref4 ref5 ref6 ref7 ref8 ref9">4, 5, 6, 7, 8, 9, 10, 11, 12</xref>
          ]. Until recently, the Text2Cypher task, which translates natural
language into Cypher; the query language used by Neo4j and other graph database systems; had received
less attention. However, with advancements in the integration of large language models (LLMs) and
knowledge graphs, text-to-graph query language (GQL) tasks, particularly Text2Cypher, have gained
increasing interest. Several datasets have been developed to support Text2Cypher research, including
Opitz and Hochgeschwender [
          <xref ref-type="bibr" rid="ref13">13</xref>
          ], S2CTrans [
          <xref ref-type="bibr" rid="ref14">14</xref>
          ], CySpider [
          <xref ref-type="bibr" rid="ref15">15</xref>
          ], Rel2Graph [
          <xref ref-type="bibr" rid="ref16">16</xref>
          ], SyntheT2C [
          <xref ref-type="bibr" rid="ref17">17</xref>
          ], and
Text2Cypher [
          <xref ref-type="bibr" rid="ref18">18</xref>
          ]. Additionally, studies have explored benchmarking and fine-tuning models for this
task, with contributions such as GPT4Graph [
          <xref ref-type="bibr" rid="ref19">19</xref>
          ], TopoChat [
          <xref ref-type="bibr" rid="ref20">20</xref>
          ], Baraki et al. [
          <xref ref-type="bibr" rid="ref21">21</xref>
          ], FCAV [22], Liang
et al. [23] and Text2Cypher [
          <xref ref-type="bibr" rid="ref18">18</xref>
          ]. In most cases, the baseline model is fine-tuned using prompts that
include natural language questions, database schema information, and ground-truth Cypher queries.
        </p>
      </sec>
      <sec id="sec-2-2">
        <title>2.2. Schema Filtering in Query Generation</title>
        <p>
          Schema information is essential for accurate query generation, ensuring correct linking of query terms
to database structures [
          <xref ref-type="bibr" rid="ref2 ref3">2, 3</xref>
          ]. This process, known as schema linking, plays a key role in Text2SQL and
Text2Cypher tasks by mapping query words to relevant database elements [
          <xref ref-type="bibr" rid="ref2">24, 2</xref>
          ]. While providing the
full schema in the prompt is possible, schema filtering is often preferred to reduce noise, computational
cost, and hallucinations [
          <xref ref-type="bibr" rid="ref2">2, 25</xref>
          ]. However, we must remain aware that excessive filtering can remove
essential components, harming accuracy [26]. Early Text2SQL schema filtering relied on heuristics like
string matching, as seen in IRNet [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ] and TypeSQL [27]. Later, learning-based methods such as Dong et
(a) Enhanced Schema
(b) Base Schema
al. [28], Bogin et al. [29], and RAT-SQL [30] were proposed. Recent approaches utilize LLMs through
prompting, fine-tuning, or agent-based techniques, such as DIN-SQL [ 31], RESDSQL [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ], CHESS [32],
E-SQL [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ], ExSL [33]and KaSLA [34]. While schema filtering is common, studies suggest it is less
necessary for LLMs with long context windows but remains valuable for smaller models [
          <xref ref-type="bibr" rid="ref2 ref3">26, 2, 3</xref>
          ]. The
trade-of is, however, that larger context sizes increase latency and computational cost for complex
databases, making filtering highly beneficial [
          <xref ref-type="bibr" rid="ref3">3</xref>
          ].
        </p>
        <p>Research on schema filtering for Text2Cypher or other graph query languages (Text2GQL) is presently
limited compared to Text2SQL. Liang et al. [23] explored aligning LLMs for a Text2GQL task in Chinese,
using a schema filtering module that executes: (i) extraction of the database schema as a dictionary,
(ii) extraction of the named entities from the query, and (iii) mapping these entities to the schema
dictionary. For queries requiring multiple nodes and relations, they used A* algorithm [35] to find the
shortest path. NAT-NL2GQL [36] includes a module for preprocessing inputs and executing schema
extraction, following a similar approach to Liang et al. [23]. Additionally, they use an LLM for filtering
multiple matched schema items before proceeding with the Text2GQL task. In this work, we examine
the impact of schema filtering on the Text2Cypher task, focusing on both performance and cost.</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Schema Filtering for Text2Cypher</title>
      <p>
        We now present schema filtering for Text2Cypher using a template from [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ] (Table 1), focusing on the
schema field with two static and three dynamic formats.
      </p>
      <sec id="sec-3-1">
        <title>3.1. Static Schemas</title>
        <p>Cypher is the query language for Neo4j, a graph database. Neo4j ofers various tools for retrieving
database schema information, based on the database structure rather than the input query. While this
allows eficient caching, it leads to longer schema representations, increasing token length and context
requirements for LLMs. We utilized two static schema formats provided by Neo4j frameworks:
• Enhanced Schema: This is one of the default schema types provided by Neo4j. It provides an
enhanced view of the database schema, including list of nodes, relationships and their properties.
It additionally provides example values for the fields. For instance, if the property is the ’name’
of the ’Actor’ node, examples might include: [’Tom Hanks’, ’Julia Roberts’, ...]. An example
enhanced schema is presented in Figure 2a.</p>
        <p>(a) Pruned By Exact-Match Schema
(b) NER Masked &amp; Pruned by Exact-Match Schema
(c) Pruned by Similarity Schema
• Base Schema: This is another default schema types provided by Neo4j. It provides similar
information as the Enhanced Schema, except it does not include examples of properties, and the
formatting is diferent. An example for this schema format is presented in Figure 2b.</p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Dynamically Pruned Schemas</title>
        <p>We implement three dynamic schema filtering approaches, which prune the baseline schemas based on
the input natural language question.</p>
        <p>• Pruned By Exact-Match: This approach compares node labels, relationship types, and properties
to words in the input question. Similar to Liang et al. [23] and NAT-NL2GQL [36], if an exact
case-insensitive match is found, the corresponding schema elements are retained; otherwise,
they are removed. Our method also considers properties as well as labels, and we retain multiple
matching elements (e.g., synonyms) to prevent excessive pruning. See Figure 3a for an example.
• NER Masked &amp; Pruned By Exact-Match: This approach replaces named entities with their
entity types before applying exact-match filtering. NER-masking prevents irrelevant matches. For
example, in the query ”List the articles that mention the organization ’Acme Energy’,” it avoids
incorrect matches, such as retaining properties of a node labeled ’Energy,’ which is unrelated.</p>
        <p>See Figure 3b for an example.
• Pruned by Similarity: This approach extends exact-match pruning by incorporating
similaritybased filtering. Instead of requiring an exact match, it computes similarity scores between query
(a) Token Distribution on Training Set
(b) Token Distribution on Test Set
terms and schema elements, retaining only those above a predefined threshold. While various
similarity measures could be used, we rely on embedding-based similarity. An example of this
schema filtering approach is shown in Figure 3c.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Experimental Setup and Results</title>
      <sec id="sec-4-1">
        <title>4.1. Experimental Setup and Evaluation Metrics</title>
        <p>
          We conducted experiments using a publicly available Text2Cypher dataset [
          <xref ref-type="bibr" rid="ref18">18</xref>
          ], focusing on a subset with
accessible databases for query execution, resulting in 22,093 training and 2,471 test samples. Schema
ifltering was assessed using the ’unsloth/Meta-Llama-3.1-8B-Instruct-bnb-4bit’, ’
unsloth/Qwen2.5-7BInstruct-bnb-4bit’ and ’GoogleAIStudio/Gemini-1.5-Flash’ models, referred as Llama-3.1-8B,
Qwen2.57B and Gemini-1.5-Flash, respectively, in the remainder of the paper. For Cypher generation, after
utilizing the LLMs, an additional post-processing step is executed to remove unwanted text, such as
’cypher:’ sufix. Furthermore, the spaCy framework is used for named entity extraction and similarity
computations. To compute evaluation metrics, we used the Hugging Face Evaluate library [37]. We
employed two evaluation procedures: (i) Translation-based (Lexical) evaluation: Compares generated
Cypher queries with reference queries based on text content. We used Google-BLEU score while
presenting the results. (ii) Execution-based evaluation: Executes both generated and reference queries
on target databases and compares their outputs (sorted lexicographically) using the same metrics as the
translation-based evaluation. We used ExactMatch score while presenting the results.
        </p>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. Evaluation Results</title>
        <p>We evaluate the proposed schema formats based on (i) token distribution and cost, and (ii) performance.</p>
        <sec id="sec-4-2-1">
          <title>4.2.1. Impact on Token Distribution &amp; Cost</title>
          <p>Schema format impacts both prompt length and token count. For example, with the Llama-3.1-8B
tokenizer, the base prompt is about 150 tokens, but adding schema information increases it to over 2,700
tokens. Figure 4 shows token distributions for training and test sets. Table 2 provides additional token
details for the test set. Results show that the Enhanced Schema leads to the longest prompts, while
switching to the Base Schema reduces the P95 token length by one-third. Exact-match pruning (with or
without NER masking) further reduces the P95 token length to 1/6th of the original. Similarity-based
pruning increases schema length but reduces the P95 token length to about 1/4th of the original.</p>
          <p>Reducing the token count reduces costs, whether for LLM vendor payments or infrastructure expenses
for self-hosted models (e.g., storage and GPU access). In a scenario with 20,000 instances, where input
token length aligns with the median in Table 2, we compare costs across models (Table 3). In the table,
we assume output lengths remain constant and only input tokens contribute to the cost. The results
show that cost scales linearly with token usage, but factors like output token count, caching, and batch
processing can afect this. Shorter prompts lead to significant cost reductions.</p>
          <p>While dynamic pruning reduces token length and costs, it may introduce computational overhead as
a side-efect. Unlike Enhanced or Base Schema (which are cached), dynamic pruning is performed for
each query, which might increase latency. However, we observe this overhead is minimal, especially
for methods like ‘Pruned by Exact-Match,’ which uses regular expression matching.</p>
        </sec>
        <sec id="sec-4-2-2">
          <title>4.2.2. Impact on Performance</title>
          <p>We evaluate the impact of proposed schema formats on Text2Cypher performance using the
Llama-3.18B model. Figure 5 presents the results, showing that longer prompts lead to lower performance. The
highest accuracy is achieved with the ‘Pruned by Exact-Match Schema.’ NER masking and
SimilarityBased Matching did not improve performance but may be beneficial for other datasets.</p>
          <p>
            We further compared the performance of diferent LLMs on a selected subset of schema formats. In
addition to Llama-3.1-8B, we evaluated Qwen2.5-7B and Gemini-1.5-Flash. While Llama-3.1-8B and
Qwen2.5-7B are similar in size, they difer in multiple ways, such as tokenization strategies.
Gemini-1.5Flash, in contrast, has a larger model size and a significantly longer context window. For comparison,
we used three schema formats—Enhanced, Base, and Pruned by Exact-Match—with decreasing token
lengths. Figure 6 presents the results, highlighting key trends: (i) In terms of lexical (translation-based)
comparison, performance of Llama-3.1-8B and Qwen2.5-7B models are improved as prompt length
decreased. However, Gemini-1.5-Flash had the opposite trend, performing better with longer prompts.
The drop in Gemini-1.5-Flash for shorter prompts was minor, remaining below 5%. (ii) In terms of
execution-based evaluation, Llama-3.1-8B model showed improved performance with shorter prompts,
while Qwen2.5-7B and Gemini-1.5-Flash experienced slight declines, both around 2%. These findings
align with observation made by previous research [
            <xref ref-type="bibr" rid="ref2 ref3">2, 3</xref>
            ]: The impact of schema length varies across
models, with Gemini-1.5-Flash potentially benefiting from longer context while the other smaller models
perform better with shorter inputs.
          </p>
          <p>(a) Schema formats</p>
          <p>(b) Translation-based - Google-Bleu score (c) Execution-based - Exact-match score</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Conclusion</title>
      <p>
        We presented schema filtering for Text2Cypher and analyzed its efects on token length, performance,
and cost. We found that reducing schema size improved performance for most models, and reduced
cost for all of those we tested. Comparison of various models revealed that smaller models performed
better with shorter prompts, while larger models benefited from longer contexts. Dynamically pruning
schemas reduced both token counts and cost, introducing slightly more latency but remained the most
eficient overall. This work has two main limitations. First, experiments used a subset of a public dataset
[
        <xref ref-type="bibr" rid="ref18">18</xref>
        ], selecting instances with accessible databases for schema extraction. However, these demo-oriented
databases often have simpler schemas than real-world ones. The longest schema had around 2700
tokens, whereas real databases likely have more complex structures, making schema filtering more
critical. Second, our filtering methods are heuristic-based. More advanced techniques, like those in
Text2SQL (see Section 2), may yield better results and require further exploration. In the future, we
will explore adaptive schema selection based on model characteristics, as well as the impact of schema
ifltering on the fine-tuning process and its efects on fine-tuned models.
      </p>
    </sec>
    <sec id="sec-6">
      <title>Declaration on Generative AI</title>
      <p>During the preparation of this work, the author(s) used Chat-GPT in order to: ’Improve writing style’
and ’Paraphrase and reword’. After using these tool(s)/service(s), the author(s) reviewed and edited the
content as needed and take(s) full responsibility for the publication’s content.
//www.diva-portal.org/smash/get/diva2:1881385/FULLTEXT01.pdf.
[22] Y. Liu, X. Wang, J. Ge, H. Wang, D. Xu, Y. Jia, Text to graph query using filter condition attributes,</p>
      <p>Proceedings of the VLDB Endowment. ISSN 2150 (2024) 8097.
[23] Y. Liang, K. Tan, T. Xie, W. Tao, S. Wang, Y. Lan, W. Qian, Aligning large language models to
a domain-specific graph database for nl2gql, in: Proceedings of the 33rd ACM International
Conference on Information and Knowledge Management, 2024, pp. 1367–1377.
[24] W. Lei, W. Wang, Z. Ma, T. Gan, W. Lu, M.-Y. Kan, T.-S. Chua, Re-examining the role of schema
linking in text-to-sql, in: Proceedings of the 2020 Conference on Empirical Methods in Natural
Language Processing (EMNLP), 2020, pp. 6943–6954.
[25] Z. Cao, Y. Zheng, Z. Fan, X. Zhang, W. Chen, X. Bai, Rsl-sql: Robust schema linking in text-to-sql
generation, arXiv preprint arXiv:2411.00073 (2024).
[26] K. Maamari, F. Abubaker, D. Jaroslawicz, A. Mhedhbi, The death of schema linking? text-to-sql in
the age of well-reasoned language models, arXiv preprint arXiv:2408.07702 (2024).
[27] T. Yu, Z. Li, Z. Zhang, R. Zhang, D. Radev, Typesql: Knowledge-based type-aware neural text-to-sql
generation, arXiv preprint arXiv:1804.09769 (2018).
[28] Z. Dong, S. Sun, H. Liu, J.-G. Lou, D. Zhang, Data-anonymous encoding for text-to-sql generation,
in: Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing
and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP),
2019, pp. 5405–5414.
[29] B. Bogin, M. Gardner, J. Berant, Global reasoning over database structures for text-to-sql parsing,
arXiv preprint arXiv:1908.11214 (2019).
[30] B. Wang, R. Shin, X. Liu, O. Polozov, M. Richardson, Rat-sql: Relation-aware schema encoding
and linking for text-to-sql parsers, arXiv preprint arXiv:1911.04942 (2019).
[31] M. Pourreza, D. Rafiei, Din-sql: Decomposed in-context learning of text-to-sql with self-correction,</p>
      <p>Advances in Neural Information Processing Systems 36 (2023) 36339–36348.
[32] S. Talaei, M. Pourreza, Y.-C. Chang, A. Mirhoseini, A. Saberi, Chess: Contextual harnessing for
eficient sql synthesis, arXiv preprint arXiv:2405.16755 (2024).
[33] M. Glass, M. Eyceoz, D. Subramanian, G. Rossiello, L. Vu, A. Gliozzo, Extractive schema linking
for text-to-sql, arXiv preprint arXiv:2501.17174 (2025).
[34] Z. Yuan, H. Chen, Z. Hong, Q. Zhang, F. Huang, X. Huang, Knapsack optimization-based schema
linking for llm-based text-to-sql generation, arXiv preprint arXiv:2502.12911 (2025).
[35] P. E. Hart, N. J. Nilsson, B. Raphael, A formal basis for the heuristic determination of minimum
cost paths, IEEE transactions on Systems Science and Cybernetics 4 (1968) 100–107.
[36] Y. Liang, T. Xie, G. Peng, Z. Huang, Y. Lan, W. Qian, Nat-nl2gql: A novel multi-agent framework
for translating natural language to graph query language, arXiv preprint arXiv:2412.10434 (2024).
[37] HuggingFace, Huggingface evaluate, 2024. https://huggingface.co/evaluate-metric.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>A.</given-names>
            <surname>Hogan</surname>
          </string-name>
          , E. Blomqvist,
          <string-name>
            <given-names>M.</given-names>
            <surname>Cochez</surname>
          </string-name>
          , C. d'Amato,
          <string-name>
            <given-names>G. D.</given-names>
            <surname>Melo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Gutierrez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Kirrane</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. E. L.</given-names>
            <surname>Gayo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Navigli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Neumaier</surname>
          </string-name>
          , et al.,
          <string-name>
            <surname>Knowledge</surname>
            <given-names>graphs</given-names>
          </string-name>
          ,
          <source>ACM Computing Surveys (Csur) 54</source>
          (
          <year>2021</year>
          )
          <fpage>1</fpage>
          -
          <lpage>37</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>H. A.</given-names>
            <surname>Caferoğlu</surname>
          </string-name>
          , Ö. Ulusoy, E-sql:
          <article-title>Direct schema linking via question enrichment in text-to-sql</article-title>
          ,
          <source>arXiv preprint arXiv:2409.16751</source>
          (
          <year>2024</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Chung</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G. T.</given-names>
            <surname>Kakkar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Gan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Milne</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Ozcan</surname>
          </string-name>
          ,
          <article-title>Is long context all you need? leveraging llm's extended context for nl2sql</article-title>
          ,
          <source>arXiv preprint arXiv:2501.12372</source>
          (
          <year>2025</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>T.</given-names>
            <surname>Yu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Yasunaga</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Ma</surname>
          </string-name>
          ,
          <string-name>
            <given-names>I.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Yao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Roman</surname>
          </string-name>
          , et al.,
          <article-title>Spider: A large-scale human-labeled dataset for complex and cross-domain semantic parsing and text-to-sql task</article-title>
          , arXiv preprint arXiv:
          <year>1809</year>
          .
          <volume>08887</volume>
          (
          <year>2018</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>J.</given-names>
            <surname>Guo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Zhan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Gao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Xiao</surname>
          </string-name>
          , J.-G. Lou, T. Liu,
          <string-name>
            <given-names>D.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <article-title>Towards complex text-to-sql in cross-domain database with intermediate representation</article-title>
          , arXiv preprint arXiv:
          <year>1905</year>
          .
          <volume>08205</volume>
          (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>H.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Chen</surname>
          </string-name>
          , Resdsql:
          <article-title>Decoupling schema linking and skeleton parsing for text-to-sql</article-title>
          ,
          <source>in: Proceedings of the AAAI Conference on Artificial Intelligence</source>
          , volume
          <volume>37</volume>
          ,
          <year>2023</year>
          , pp.
          <fpage>13067</fpage>
          -
          <lpage>13075</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Fan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>He</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Ren</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Huang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Jing</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X. S.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <article-title>Metasql: A generate-then-rank framework for natural language to sql translation</article-title>
          ,
          <source>arXiv preprint arXiv:2402.17144</source>
          (
          <year>2024</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>K.</given-names>
            <surname>Shen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Kejriwal</surname>
          </string-name>
          , Select-sql:
          <article-title>Self-correcting ensemble chain-of-thought for text-to-sql</article-title>
          ,
          <source>arXiv preprint arXiv:2409.10007</source>
          (
          <year>2024</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>J.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>I.</given-names>
            <surname>Baek</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Kim</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <article-title>Safe-sql: Self-augmented in-context learning with fine-grained example selection for text-to-sql</article-title>
          ,
          <source>arXiv preprint arXiv:2502.11438</source>
          (
          <year>2025</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>F.</given-names>
            <surname>Brei</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Frey</surname>
          </string-name>
          , L.-P. Meyer,
          <article-title>Leveraging small language models for text2sparql tasks to improve the resilience of ai assistance</article-title>
          ,
          <source>arXiv preprint arXiv:2405.17076</source>
          (
          <year>2024</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>L.-P.</given-names>
            <surname>Meyer</surname>
          </string-name>
          , J.
          <string-name>
            <surname>Frey</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Brei</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          <string-name>
            <surname>Arndt</surname>
          </string-name>
          ,
          <article-title>Assessing sparql capabilities of large language models</article-title>
          ,
          <source>arXiv preprint arXiv:2409.05925</source>
          (
          <year>2024</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>V.</given-names>
            <surname>Emonet</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Bolleman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Duvaud</surname>
          </string-name>
          ,
          <string-name>
            <surname>T. M. de Farias</surname>
            ,
            <given-names>A. C.</given-names>
          </string-name>
          <string-name>
            <surname>Sima</surname>
          </string-name>
          ,
          <article-title>Llm-based sparql query generation from natural language over federated knowledge graphs</article-title>
          ,
          <source>arXiv preprint arXiv:2410.06062</source>
          (
          <year>2024</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>D.</given-names>
            <surname>Opitz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Hochgeschwender</surname>
          </string-name>
          ,
          <article-title>From zero to hero: generating training data for question-to-cypher models</article-title>
          ,
          <source>in: Proceedings of the 1st International Workshop on Natural Language-based Software Engineering</source>
          ,
          <year>2022</year>
          , pp.
          <fpage>17</fpage>
          -
          <lpage>20</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Zhao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Ge</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Shen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Hu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <article-title>S2ctrans: Building a bridge from sparql to cypher</article-title>
          ,
          <source>in: International Conference on Database and Expert Systems Applications</source>
          , Springer,
          <year>2023</year>
          , pp.
          <fpage>424</fpage>
          -
          <lpage>430</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Zhao</surname>
          </string-name>
          , W. Liu,
          <string-name>
            <given-names>T.</given-names>
            <surname>French</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Stewart</surname>
          </string-name>
          ,
          <article-title>Cyspider: A neural semantic parsing corpus with baseline models for property graphs</article-title>
          ,
          <source>in: Australasian Joint Conference on Artificial Intelligence</source>
          , Springer,
          <year>2023</year>
          , pp.
          <fpage>120</fpage>
          -
          <lpage>132</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Zhao</surname>
          </string-name>
          , W. Liu,
          <string-name>
            <given-names>T.</given-names>
            <surname>French</surname>
          </string-name>
          ,
          <string-name>
            <surname>M.</surname>
          </string-name>
          <article-title>Stewart, Rel2graph: Automated mapping from relational databases to a unified property knowledge graph</article-title>
          ,
          <source>arXiv preprint arXiv:2310.01080</source>
          (
          <year>2023</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Zhong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Zhong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Sun</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Jin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Qin</surname>
          </string-name>
          ,
          <string-name>
            <surname>X. Zhang,</surname>
          </string-name>
          <article-title>Synthet2c: Generating synthetic data for ifne-tuning large language models on the text2cypher task</article-title>
          ,
          <source>arXiv preprint arXiv:2406.10710</source>
          (
          <year>2024</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>M. G.</given-names>
            <surname>Ozsoy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Messallem</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Besga</surname>
          </string-name>
          , G. Minneci,
          <article-title>Text2cypher: Bridging natural language and graph databases</article-title>
          ,
          <source>in: Proceedings of the Workshop on Generative AI and Knowledge Graphs (GenAIK)</source>
          ,
          <year>2025</year>
          , pp.
          <fpage>100</fpage>
          -
          <lpage>108</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>J.</given-names>
            <surname>Guo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Du</surname>
          </string-name>
          , H. Liu,
          <string-name>
            <given-names>M.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>He</surname>
          </string-name>
          , S. Han,
          <article-title>Gpt4graph: Can large language models understand graph structured data? an empirical evaluation and benchmarking</article-title>
          ,
          <source>arXiv preprint arXiv:2305.15066</source>
          (
          <year>2023</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>H.</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Jin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Zhu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Weng</surname>
          </string-name>
          , Topochat:
          <article-title>Enhancing topological materials retrieval with large language model and multi-source knowledge</article-title>
          ,
          <source>arXiv preprint arXiv:2409.13732</source>
          (
          <year>2024</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>W. W.</given-names>
            <surname>Baraki</surname>
          </string-name>
          ,
          <article-title>Leveraging large language models for accurate Cypher query generation: Natural language query to Cypher statements, Master degree project</article-title>
          , University of Skövde,
          <year>2024</year>
          . https:
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>