<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>LLMASP: A Framework for Mitigating Hallucinations and Supporting Symbolic Reasoning in Large Language Models</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Lorenzo Grillo</string-name>
          <email>loregrillo28@gmail.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>University of Calabria</institution>
          ,
          <addr-line>Rende (CS)</addr-line>
          <country country="IT">Italy</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2025</year>
      </pub-date>
      <fpage>9</fpage>
      <lpage>13</lpage>
      <abstract>
        <p>LLMASP is a framework that integrates Large Language Models (LLMs) with Answer Set Programming (ASP) to combine the strengths of natural language understanding and formal logical reasoning. It enables the extraction of structured, domain-specific facts from natural language input using YAML-based specifications and constrained output formats defined by GBNF grammars. These facts are then fed into an ASP solver to perform precise reasoning, and the results are translated back into natural language by the LLM to improve interpretability. Currently, the proposal evaluates the performance and accuracy of diferent GBNF grammars by analyzing the quality of the ASP facts generated from a user input. The evaluation is done on a set of ASP problems and shows that JSON-formatted outputs achieve higher extraction accuracy, except for longer inference times, whereas CSV outputs ofer faster performance with a slight reduction in precision.</p>
      </abstract>
      <kwd-group>
        <kwd>Natural Language processing</kwd>
        <kwd>Answer Set Programming</kwd>
        <kwd>Large Language Models</kwd>
        <kwd>Mitigating hallucinations and</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction and problem description</title>
      <p>
        Large Language Models (LLMs) and Answer Set Programming (ASP) represent two distinct yet
complementary paradigms in Artificial Intelligence. LLMs, such as GPT [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], PaLM [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ], and LLaMa [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ],
have transformed natural language processing (NLP) by achieving unprecedented levels of fluency and
understanding in textual data. Their rise was largely stimulated by the introduction of the Transformer
architecture, as described in the seminal paper “Attention is All You Need” by Vaswani et al. (2017)
[
        <xref ref-type="bibr" rid="ref4">4</xref>
        ], which introduced the self-attention mechanism as a powerful alternative to recurrent structures
in sequence modeling. Since then, attention mechanisms have evolved significantly. Variants such
as Active-Dormant Attention [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ], Flash Attention [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ], and ALiBi (Attention with Linear Biases) [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]
have been introduced to enhance eficiency, scalability, and context preservation. In addition to
architectural innovations, reasoning capabilities of LLMs have been improved through techniques such
as Chain-of-Thought (CoT) prompting [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ], which enables models to break down complex reasoning
tasks into intermediate steps, significantly enhancing performance on tasks requiring multi-step logic.
Moreover, LLMs can be guided to produce structured outputs [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] by adhering to formal grammars
[
        <xref ref-type="bibr" rid="ref10">10</xref>
        ], such as GBNF (GGML Backus-Naur Form) specifications [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ], allowing for constrained generation
in applications requiring syntax compliance. In contrast, Answer Set Programming (ASP) provides
a declarative paradigm for knowledge representation and reasoning. ASP enables systems to model
complex domains using logical rules, infer consequences, and solve combinatorial problems with high
eficiency [
        <xref ref-type="bibr" rid="ref12 ref13">12, 13</xref>
        ]. It has demonstrated success in applications such as planning [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ], diagnosis [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ],
configuration [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ] and decision-making. This proposal introduces a comprehensive framework that
combines Large Language Models (LLMs) and Answer Set Programming (ASP), leveraging the strengths
of both paradigms to address their respective limitations [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ]. It presents a method for encoding
domain-specific knowledge into input prompts using a YAML-based format, allowing LLMs to generate
structured relational facts that are then processed by the ASP solver for reasoning. The outcomes
inferred by ASP are subsequently translated back into natural language by the LLM, enhancing the user
      </p>
      <p>CEUR
Workshop</p>
      <p>ISSN1613-0073
experience and improving the interpretability of the results [18], Figure 1.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Background and overview of the existing literature</title>
      <p>Large Language Models (LLMs) are sophisticated artificial intelligence systems based on the Transformer
architecture. They are trained on extensive collections of textual data to learn complex patterns and
structures of human language. Given an input prompt, optionally accompanied by a history consisting
of the list of previous messages, they return a completion, that is, a textual response that probabilistically
continues or answers the input.</p>
      <p>Moreover, LLM frameworks such as Ollama ofer the possibility of constraining the generation process
to strictly adhere to a formal grammar, specifically the Grammar Backus–Naur Form (GBNF), which can
be automatically generated from a JSON structure defined using JSON Schema. By patching the Ollama
framework, it is also possible to send user-defined grammars via the API, allowing for greater control
over the structure of the response. This is crucial for optimizing token usage, for example by specifying
output formats such as CSV. Since the GBNF grammar imposes strict constraints, it is assumed that the
generated output always conforms to the expected structure.</p>
      <p>Let’s consider the following example: the task is to extract the person and the action they are performing
from the user prompt. A GBNF grammar for this task is defined as follows:
root ::= who action
who ::= "Alice" | "Marco" | "Luca"
action ::= "eats pasta" | "plays computer games" | "sleeps"</p>
      <sec id="sec-2-1">
        <title>The GBNF file is passed to the patched Ollama API in the following way:</title>
        <p>gbnf_grammar = open("grammar.gbnf")
chat.completion(messages=["I've seen Alice eating pasta"], grammar=gbnf_grammar)</p>
      </sec>
      <sec id="sec-2-2">
        <title>The output of that query will be:</title>
        <p>Alice eats pasta.
which follows the format previously defined by the grammar.</p>
        <p>Answer Set Programming (ASP) is a declarative, rule-based language designed for knowledge
representation and reasoning, whose correctness is based on formal semantics. An ASP program consists of a
collection of rules that derive new facts from a given input. It can be read by humans and the logic can
be validated by domain experts. The solutions generated by an ASP program are formally referred to as
answer sets, that is, collections of inferred facts that satisfy all the rules and constraints of the program.
Here’s an example of an ASP program:
a :- not b. % means a is true if we cannot prove b.
b :- not a. % means b is true if we cannot prove a.</p>
        <p>c :- a. % means c is true if a is true.</p>
      </sec>
      <sec id="sec-2-3">
        <title>This program has exactly two answer sets:</title>
        <p>{, }
and {}.</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Goal of the research</title>
      <p>The goal of this research is to enable LLMs to perform structured reasoning by combining their strong
pattern-matching abilities with the logical reasoning capabilities of Answer Set Programming (ASP).
This approach allows for solving complex problems defined in ASP while also reducing hallucinations
in LLM outputs [19, 20]. Since the resolution is ofloaded to ASP, the problem doesn’t need to be part of
the training set. Users can input plain natural language, which is then translated into ASP for reasoning
and resolution. To achieve this, the research will focus on the following sub-goals:
• Conversion back to natural language: Methods to translate the ASP result back to natural
language by using a diferent behavior YAML. The translation makes sure that all the ASP facts
are correctly included inside the produced natural language output (Hallucination mitigation).
• Automatic generation of the ASP code: Taking a natural language problem description,
generate the required ASP logic to solve the problem.
• Loop LLM-ASP Feedback: A closed loop interaction where LLM outputs are iteratively refined
through ASP reasoning, enabling more accurate and controlled information extraction.
• Neural Network Layer: Reduces latency by eliminating repeated external calls and would allow
direct control over the internal mechanisms of the Transformer, such as dynamic adjustment of
token weights.</p>
    </sec>
    <sec id="sec-4">
      <title>4. Current status of the research</title>
      <p>(a) LLM result without LLMASP
(b) The ASP solution of the 4-Queens problem</p>
      <p>The current state of the research involves the definition of a structured format to orchestrate the
interaction between the LLM and the ASP solver. User input is treated as a textual database. Leveraging
GBNF grammar and a predefined protocol, the LLM is able to query and extract relevant atoms, if present
in the input, and convert them into ASP facts. These atoms are specified within a YAML configuration
ifle, which also defines the model’s behavior, including the System Prompt and concise descriptions of
each atom to be extracted. The extracted facts are then passed to the ASP solver, which computes and
returns the solution to the defined problem.</p>
      <p>To illustrate the proposed approach, consider the following example. The task is to solve the well-known
N-Queens problem, described as follows:
”The N-Queens puzzle consists in placing n queens on an n×n chessboard such that no two queens
threaten each other. Given an integer n, return one solution to the N-Queens puzzle.”
The LLM used in this example is LLAMA 3.1 7B. When provided with the problem description and the
following prompt:</p>
      <p>Return all the solutions for n=4.
the LLM produces the output shown in Figure 2a, which is incorrect because the queens threaten each
other diagonally.</p>
      <p>Now, consider the same LLM combined with the LLMASP framework. LLMASP requires an Application
YAML file containing the problem statement and the definition of the facts to be extracted 3a, as well as
a behavior file specifying how the LLM should extract the data and which intermediate format should
be used 3b. The intermediate format can be either CSV or JSON, and the format specification is used to
control the output structure of the LLM. When CSV is selected, the LLM produces a simple, tabular
output such as:
which contains only the essential components required to construct an ASP fact, separated by tab
separator. When JSON is selected, the LLM follows the rules defined by the GBNF grammar automatically
generated by Ollama. The output is structured as:</p>
      <p>{ "list_fact1": [{"head" : "body"}, ] }</p>
      <sec id="sec-4-1">
        <title>Using the same prompt, the LLM extracts the following fact:</title>
        <p>which is subsequently passed to the ASP solver, returning:</p>
        <p>queen(1, 2). queen(2, 4). queen(3, 1). queen(4, 3).</p>
        <p>This solution is correct, as the queens do not threaten each other, as shown in figure 2b.</p>
        <sec id="sec-4-1-1">
          <title>Problem</title>
        </sec>
        <sec id="sec-4-1-2">
          <title>JSON (PD) JSON CSV CSV (PD)</title>
          <p>ACC
F1 ACC
F1 ACC
F1 ACC
F1</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Preliminary results accomplished</title>
      <p>In order to verify the correct extraction of the information encoded in ASP, several tests have been
conducted to determine the most suitable setup for this component. The tests were conducted on the
following hardware configuration: an RTX 3070 GPU, an Intel i7-11700KF CPU, and 32GB of DDR4
RAM. They focused on evaluating the precision of ASP fact extraction from user input. It also measured
performance and examined which additional information, in this case the problem description, could
improve the accuracy of the LLM’s extraction.</p>
      <p>The execution time of each method (JSON and CSV) is evaluated by repeating the tests 10 times to
verify consistency. The language model produced identical outputs across all runs, as it was configured
with a temperature value of 0 and a top_p parameter set to 0.98.</p>
      <p>These parameters are commonly used to regulate the stochasticity of the model’s responses. Specifically,
the temperature parameter influences the degree of creativity in the output: higher values lead to
more diverse and imaginative responses, while lower values favor determinism and repeatability[21].
Similarly, the top_p parameter limits the sampling space to the top p most probable tokens, with lower
values introducing greater variability and higher values promoting more stable and predictable results.
The precision of the extracted facts was evaluated using standard classification metrics, including true
positives, true negatives, false positives, and false negatives, from which the F1-Score was derived. In
the initial test configuration, the problem description was included in the system prompt to provide
additional context to the language model.</p>
      <p>The results shown in the table 1, indicate that the JSON format with the problem description inside the
system prompt (PD) provided an higher level of precision compared to the CSV (PD) format also with
the same setup. However, this improvement in accuracy was accompanied by an increase in execution
time. Specifically, the JSON GBNF grammar required an average of 665 seconds (approximately 11
minutes) to complete the tests, whereas the CSV GBNF grammar approach completed in an average of
540 seconds (approximately 9 minutes). In the second test, the problem description was omitted from
the system prompt. The JSON format maintained approximately the same level of precision. In some
cases, performance slightly decreased, while in others it improved. Similarly, the CSV format exhibited
a decline in performance for certain problems but demonstrated improved precision in others.
As can be observed in table 2, the JSON-based method produces approximately 100,000 more tokens
than the CSV-based method. This significant diference reflects the greater verbosity and structural
complexity typically associated with the JSON format. While the increased token count may contribute
to slightly higher precision by providing the language model with more explicit and structured
information, it also leads to longer processing times and greater computational cost.</p>
      <p>On the other hand, the CSV format, being more concise and lightweight, results in faster response times
and reduced token usage. However, this compactness may occasionally limit the model’s ability to fully
interpret the intended meaning, potentially leading to a loss in extraction precision.</p>
    </sec>
    <sec id="sec-6">
      <title>6. Open issues and expected achievements</title>
      <p>The proposed framework still has several unresolved issues. Notably, the need to invoke the LLM once
per atom significantly slows down the extraction process compared to a single-pass LLM inference.
Additionally, the LLM may occasionally enter an infinite generation loop, repeatedly producing the
same token. This behavior has been partially mitigated by using the streaming mode of the Ollama API
to detect token repetition: when a loop is identified, the chat.completion function is forcibly stopped,
and the system returns the partial output extracted up to that point. When the information extracted is
not suficient to solve the problem, an Active-Prompting system can be employed to request additional
input in order to satisfy the requirements.</p>
      <p>The expected achievements include reaching logic problem solving performance comparable to larger
models, showing that complex reasoning can be done eficiently also with the smaller ones. Additionally,
the framework can be designed to run on low powered devices, enabling embedded hardware to handle
complex tasks without relying on powerful hardware.</p>
    </sec>
    <sec id="sec-7">
      <title>Declaration on Generative AI</title>
      <p>During the preparation of this work, the author used ChatGPT, Grammarly in order to: Grammar and
spelling check, Paraphrase and reword. After using this tool/service, the author reviewed and edited
the content as needed and take full responsibility for the publication’s content.
[18] M. Alviano, L. Grillo, F. Lo Scudo, L. A. R. Reiners, Integrating answer set programming and
large language models for enhanced structured representation of complex knowledge in natural
language, in: IJCAI 2025, Montreal, Canada, August 16-22, 2025, https://tinyurl.com/ijcai25-llmasp,
2025.
[19] Y. Bang, Z. Ji, A. Schelten, A. Hartshorn, T. Fowler, C. Zhang, N. Cancedda, P. Fung, Hallulens:</p>
      <p>Llm hallucination benchmark, 2025. URL: https://arxiv.org/abs/2504.17550. arXiv:2504.17550.
[20] G. Perković, A. Drobnjak, I. Botički, Hallucinations in llms: Understanding and addressing
challenges, in: 2024 47th MIPRO ICT and Electronics Convention (MIPRO), 2024, pp. 2084–2088.
doi:10.1109/MIPRO60963.2024.10569238.
[21] L. Li, L. Sleem, N. Gentile, G. Nichil, R. State, Exploring the impact of temperature on large language
models:hot or cold?, 2025. URL: https://arxiv.org/abs/2506.07295. arXiv:2506.07295.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <surname>T. B. Brown</surname>
          </string-name>
          , et al.,
          <article-title>Language models are few-shot learners</article-title>
          , CoRR abs/
          <year>2005</year>
          .14165 (
          <year>2020</year>
          ). URL: https://arxiv.org/abs/
          <year>2005</year>
          .14165. arXiv:
          <year>2005</year>
          .14165.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>A.</given-names>
            <surname>Chowdhery</surname>
          </string-name>
          , et al.,
          <article-title>Palm: Scaling language modeling with pathways</article-title>
          ,
          <source>J. Mach. Learn. Res</source>
          .
          <volume>24</volume>
          (
          <year>2023</year>
          )
          <volume>240</volume>
          :
          <fpage>1</fpage>
          -
          <lpage>240</lpage>
          :
          <fpage>113</fpage>
          . URL: http://jmlr.org/papers/v24/
          <fpage>22</fpage>
          -
          <lpage>1144</lpage>
          .html.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>H.</given-names>
            <surname>Touvron</surname>
          </string-name>
          , et al.,
          <article-title>Llama: Open and eficient foundation language models</article-title>
          ,
          <source>CoRR abs/2302</source>
          .13971 (
          <year>2023</year>
          ). doi:
          <volume>10</volume>
          .48550/ARXIV.2302.13971. arXiv:
          <volume>2302</volume>
          .
          <fpage>13971</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>A.</given-names>
            <surname>Vaswani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Shazeer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Parmar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Uszkoreit</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Jones</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. N.</given-names>
            <surname>Gomez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Kaiser</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Polosukhin</surname>
          </string-name>
          ,
          <article-title>Attention is all you need</article-title>
          , https://arxiv.org/abs/1706.03762,
          <year>2023</year>
          . arXiv:
          <volume>1706</volume>
          .
          <fpage>03762</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>T.</given-names>
            <surname>Guo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Pai</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Bai</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Jiao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. I.</given-names>
            <surname>Jordan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Mei</surname>
          </string-name>
          ,
          <article-title>Active-dormant attention heads: Mechanistically demystifying extreme-token phenomena in llms</article-title>
          , https://arxiv.org/pdf/2410.13835,
          <year>2024</year>
          . arXiv:
          <volume>2410</volume>
          .
          <fpage>13835</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>T.</given-names>
            <surname>Dao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. Y.</given-names>
            <surname>Fu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Ermon</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Rudra</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Ré</surname>
          </string-name>
          ,
          <article-title>Flashattention: Fast and memory-eficient exact attention with io-awareness</article-title>
          , https://arxiv.org/pdf/2205.14135,
          <year>2022</year>
          . arXiv:
          <volume>2205</volume>
          .
          <fpage>14135</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            O. Press,
            <given-names>N. A.</given-names>
            <surname>Smith</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Lewis</surname>
          </string-name>
          ,
          <article-title>Train short, test long: Attention with linear biases enables input length extrapolation</article-title>
          , https://arxiv.org/abs/2108.12409,
          <year>2022</year>
          . arXiv:
          <volume>2108</volume>
          .
          <fpage>12409</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>J.</given-names>
            <surname>Wei</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Schuurmans</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Bosma</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Ichter</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Xia</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Chi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Le</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <article-title>Chain-ofthought prompting elicits reasoning in large language models</article-title>
          , https://arxiv.org/pdf/2201.11903,
          <year>2023</year>
          . arXiv:
          <volume>2201</volume>
          .
          <fpage>11903</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>M. X.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. J.</given-names>
            <surname>Fiannaca</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Koo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Dixon</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Terry</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C. J.</given-names>
            <surname>Cai</surname>
          </string-name>
          , ”
          <article-title>we need structured output”: Towards user-centered constraints on large language model output</article-title>
          ,
          <source>in: Extended Abstracts of the CHI Conference on Human Factors in Computing Systems, CHI EA '24</source>
          ,
          <string-name>
            <surname>Association</surname>
          </string-name>
          for Computing Machinery, New York, NY, USA,
          <year>2024</year>
          . URL: https://doi.org/10.1145/3613905.3650756. doi:
          <volume>10</volume>
          .1145/3613905.3650756.
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>D.</given-names>
            <surname>Deutsch</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Upadhyay</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Roth</surname>
          </string-name>
          , A
          <article-title>general-purpose algorithm for constrained sequential inference</article-title>
          , in: M.
          <string-name>
            <surname>Bansal</surname>
            ,
            <given-names>A</given-names>
          </string-name>
          . Villavicencio (Eds.),
          <source>Proceedings of the 23rd Conference on Computational Natural Language Learning</source>
          ,
          <source>CoNLL</source>
          <year>2019</year>
          ,
          <string-name>
            <given-names>Hong</given-names>
            <surname>Kong</surname>
          </string-name>
          , China, November 3-
          <issue>4</issue>
          ,
          <year>2019</year>
          , Association for Computational Linguistics,
          <year>2019</year>
          , pp.
          <fpage>482</fpage>
          -
          <lpage>492</lpage>
          . URL: https://doi.org/10.18653/v1/
          <fpage>K19</fpage>
          -1045. doi:
          <volume>10</volume>
          .18653/V1/K19-1045.
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <surname>D. D. McCracken</surname>
            ,
            <given-names>E. D.</given-names>
          </string-name>
          <string-name>
            <surname>Reilly</surname>
          </string-name>
          ,
          <article-title>Backus-Naur form (BNF)</article-title>
          , John Wiley and Sons Ltd.,
          <string-name>
            <surname>GBR</surname>
          </string-name>
          ,
          <year>2003</year>
          , p.
          <fpage>129</fpage>
          -
          <lpage>131</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>V.</given-names>
            <surname>Marek</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Truszczyński</surname>
          </string-name>
          ,
          <article-title>Stable models and an alternative logic programming paradigm</article-title>
          ,
          <source>in: The Logic Programming Paradigm: a 25-year Perspective</source>
          ,
          <year>1999</year>
          , pp.
          <fpage>375</fpage>
          -
          <lpage>398</lpage>
          . doi:
          <volume>10</volume>
          .1007/ 978-3-
          <fpage>642</fpage>
          -60085-2_
          <fpage>17</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <surname>I. Niemelä</surname>
          </string-name>
          ,
          <article-title>Logic programming with stable model semantics as a constraint programming paradigm</article-title>
          ,
          <source>Annals of Mathematics and Artificial Intelligence</source>
          <volume>25</volume>
          (
          <year>1999</year>
          )
          <fpage>241</fpage>
          -
          <lpage>273</lpage>
          . doi:
          <volume>10</volume>
          .1023/A:
          <fpage>1018930122475</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>P.</given-names>
            <surname>Cappanera</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Gavanelli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Nonato</surname>
          </string-name>
          , M. Roma,
          <article-title>Logic-based benders decomposition in answer set programming for chronic outpatients scheduling</article-title>
          ,
          <source>TPLP</source>
          <volume>23</volume>
          (
          <year>2023</year>
          )
          <fpage>848</fpage>
          -
          <lpage>864</lpage>
          . doi:
          <volume>10</volume>
          .1017/ S147106842300025X.
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>F.</given-names>
            <surname>Wotawa</surname>
          </string-name>
          ,
          <article-title>On the use of answer set programming for model-based diagnosis</article-title>
          , in: H.
          <string-name>
            <surname>Fujita</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Fournier-Viger</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Ali</surname>
          </string-name>
          , J. Sasaki (Eds.), IEA/AIE 2020, Kitakyushu, Japan,
          <source>September 22-25</source>
          ,
          <year>2020</year>
          , Proceedings, volume
          <volume>12144</volume>
          <source>of LNCS</source>
          , Springer,
          <year>2020</year>
          , pp.
          <fpage>518</fpage>
          -
          <lpage>529</lpage>
          . doi:
          <volume>10</volume>
          .1007/ 978-3-
          <fpage>030</fpage>
          -55789-8_
          <fpage>45</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>R.</given-names>
            <surname>Taupe</surname>
          </string-name>
          , G. Friedrich,
          <string-name>
            <given-names>K.</given-names>
            <surname>Schekotihin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Weinzierl</surname>
          </string-name>
          ,
          <article-title>Solving configuration problems with ASP and declarative domain specific heuristics</article-title>
          , in: M.
          <string-name>
            <surname>Aldanondo</surname>
            ,
            <given-names>A. A.</given-names>
          </string-name>
          <string-name>
            <surname>Falkner</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Felfernig</surname>
          </string-name>
          , M. Stettinger (Eds.),
          <source>Proceedings of the 23rd International Configuration Workshop</source>
          (CWS/ConfWS
          <year>2021</year>
          ), Vienna, Austria,
          <fpage>16</fpage>
          -
          <lpage>17</lpage>
          September,
          <year>2021</year>
          , volume
          <volume>2945</volume>
          <source>of CEUR Workshop Proceedings, CEUR-WS.org</source>
          ,
          <year>2021</year>
          , pp.
          <fpage>13</fpage>
          -
          <lpage>20</lpage>
          . URL: https://ceur-ws.
          <source>org/</source>
          Vol-
          <volume>2945</volume>
          /
          <fpage>21</fpage>
          -RT-ConfWS21_
          <article-title>paper_4</article-title>
          .pdf.
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>S.</given-names>
            <surname>Williams</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Huckle</surname>
          </string-name>
          ,
          <article-title>Easy problems that llms get wrong</article-title>
          ,
          <year>2024</year>
          . URL: https://arxiv.org/abs/2405. 19616. arXiv:
          <volume>2405</volume>
          .
          <fpage>19616</fpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>