<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>July</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Improving the Accuracy of Black-Box Language Models with Ontologies: A Preliminary Roadmap</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Marco Monti</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Oliver Kutz</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Guendalina Righetti</string-name>
          <xref ref-type="aff" rid="aff3">3</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Nicolas Troquard</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Free University of Bozen-Bolzano (UNIBZ)</institution>
          ,
          <addr-line>Piazza Università, 1-39100, Bozen-Bolzano (BZ)</addr-line>
          ,
          <country country="IT">Italy</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Gran Sasso Science Institute (GSSI)</institution>
          ,
          <addr-line>Viale Francesco Crispi, 7-67100, LAquila (AQ)</addr-line>
          ,
          <country country="IT">Italy</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>IBM</institution>
          ,
          <addr-line>Circonvallazione Idroscalo, 20090, Segrate (MI)</addr-line>
          ,
          <country country="IT">Italy</country>
        </aff>
        <aff id="aff3">
          <label>3</label>
          <institution>University of Oslo, Department of Philosophy</institution>
          ,
          <addr-line>Oslo</addr-line>
          ,
          <country country="NO">Norway</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2024</year>
      </pub-date>
      <volume>1</volume>
      <fpage>5</fpage>
      <lpage>19</lpage>
      <abstract>
        <p>Large Language Models (LLMs) have revolutionised natural language generation. But their statistical and auto-regressive nature makes them unreliable. It has become clear to the research community that in order to produce reliably correct answers, LLMs need to be enriched in some way with 'world models' reflecting the semantics of the domains being queried. We here propose a simple workflow to address this problem through a neuro-symbolic interaction protocol with the LLM treated as a blackbox. Answers given by an LLM are checked against accepted knowledge provided by a domain ontology. The approach aims to combine conflict detection with explanation extraction and formal repairs presented to the LLM in the form of specific artificial speech acts. The goal is to build constraining, incremental prompts that improve repeatability and veracity in the LLM's output.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;LLM</kwd>
        <kwd>ontologies</kwd>
        <kwd>neuro-symbolic reasoning</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        LLMs ofer serious potential in knowledge discovery and information retrieval given the training
on vast corpora of text. Examples include simple lookups of basic facts, summarisation in the
style of Wikipedia abstracts, producing reformulations of dificult-to-understand documents
such as those found in medical diagnosis, etc. [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. It is therefore no surprise that some of the
apparent skills of LLMs are also being explored for complementing or assisting ontological
reasoning tasks, namely in particular learning new subsumptions and building concept taxonomies,
or populating existing ontologies with entities [
        <xref ref-type="bibr" rid="ref2 ref3">2, 3</xref>
        ].
      </p>
      <p>
        Limitations. Despite their remarkable achievements, LLMs also exhibit a number of
significant drawbacks. Like other models of artificial neural networks, LLMs are susceptible to
ontology

answers
biases and have limited contextual understanding [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. But perhaps the most critical limitation
concerns their lack of accuracy, manifested by so-called hallucinations [
        <xref ref-type="bibr" rid="ref5 ref6">5, 6</xref>
        ].
      </p>
      <p>
        Yann LeCun’s “unpopular opinion” in a recent series of talks has been that “auto-regressive
LLMs are doomed” [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]. An argument for this claim goes roughly as follows. Assume a language
model  , and suppose that  is the probability that any token produced by  takes us outside
of the set of correct answers. The probability that an answer of length  provided by the
language model  is correct is then (1 − ), which converges to zero with the length of the
answer. Of course, with a suficiently small value of  (close to 0) and a ‘not too long’ answer,
the performance of such a model  can still be very high. Figure 1 illustrates this general
situation. Our core concern thus is to study how to use ontologies and formal reasoning to steer
the LLM to remain in, or at least close to, the space of ‘correct’ answers. We first provide some
context on existing mitigation methods for hallucinations which are knowledge-based.
Some existing hallucination mitigation methods. Yin et al. [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] propose a generative
question answering system. To provide correct answers, the system is connected to a triple
store of true facts, from which it retrieves a set of candidate facts and generates an answer
to the question. Further methods exploring how to enhance machine reading comprehension
systems by incorporating external knowledge sources are presented by Bi et al. [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]. Li et al.
[
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] also address the issue of semantic drift in generative question answering by incorporating
external knowledge. Martino et al. [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] use knowledge injection to counter hallucinations in
large language models. Retrieval-Augmented Generation (RAG) [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ] uses expert knowledge and
domain-specific related documents to augment answers to queries, which are then processed
together by the LLM to better contextualize them.
      </p>
      <p>
        Ji et al. [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ] propose a method for mitigating LLM hallucinations via self-reflection. This
approach involves three self-reflective loops: factual knowledge acquisition, knowledge-consistent
answering, and question-entailment answering. Galitsky [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ] presents a fact-checking system
that exploits web mining to find correct information to suggest to the LLM. The system
capitalises on argumentation analysis and defeasible logic programming to handle inconsistent
sources. For a more complete survey of existing hallucination mitigation methods, refer to [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ].
prompt 
 =
makeprompt()
Using formal ontologies as world models. It has become clear that LLMs need world
models to generate factual responses. Ontologies are natural candidates for being the providers
of these world models.
      </p>
      <p>
        Ontologies are formal descriptions of the entities present within a domain of interest. They
describe the concepts, the individuals that populate them, and the relationships that hold
between them. Ontologies may contain implicit knowledge (a set of general axioms) and
explicit knowledge (a set of factual statements). The latter can be represented as rows in a
relational database, as a knowledge graph, or a Description Logic (DL) ABox. The former
must be represented as a set of logical formulas, e.g., as a DL TBox. Description Logics are
natural candidates because their reasoning problems (consistency checking, entailment, instance
checking, etc.) are usually decidable and eficient algorithms and implementations exist. Yet,
they maintain a reasonable expressivity. DLs also form the theoretical underpinning of the W3C
Web Ontology Language (OWL). See [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ] for an introduction to Description Logics. Moreover,
some of the key technical elements that are required to interact and converse with an LLM are
more readily available in the DL context, as described further below. This includes non-classical
reasoning approaches for conflict detection, knowledge debugging, formal argumentation, or
knowledge weakening [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ].
      </p>
      <p>We thus suggest to use logical reasoning with background ontological domain knowledge to
detect inconsistent answers (Figure 1), and iteratively nudge the LLMs’ answers back on a path
of correct answers.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Steering LMs towards accuracy: a generic workflow</title>
      <p>A circular workflow is depicted in Figure 2, illustrating the possible enhancement of a Language
Model’s capabilities through interfacing with ontological reasoning. This workflow iteratively
enriches the LLM’s inference abilities by fostering a symbiotic relationship between linguistic
proficiency and ontological reasoning.</p>
      <sec id="sec-2-1">
        <title>We briefly discuss the key elements of our workflow:</title>
        <p>Prompt : The interaction begins with the user providing input in the form of text. This input
can in general range from simple queries to complex prompts, questions, or commands.
In our scenario, the prompt may be designed to:
1. ask for a succinct answer, so as to limit the issue with auto-regressive LLMs,
2. use only simple concepts and relations that have a direct counterpart in the given
ontology,
3. target certain central concepts in the specific domain of knowledge according to
subject-matter-experts.</p>
        <p>LLM: The LLM is treated as a blackbox. The workflow does not interfere with learning and is
not intended for fine-tuning purposes.</p>
        <p>Answer :  is a textual response to the prompt  generated by the LLM. This response 
typically appears to be a coherent and relevant piece of text that addresses the user’s
prompt. Ideally, this answer is short, as requested by , to limit the issue with
autoregressive LLMs.</p>
        <p>
          Formulizer formul: is a computational module designed to convert English responses  into
formal expressions formul() represented (largely) within the signature and logical
language used in the . This problem has been addressed by the DL research community
[
          <xref ref-type="bibr" rid="ref18 ref19">18, 19</xref>
          ]. Transforming a textual response into a given formal language can of course be
also achieved through appropriately training a network [
          <xref ref-type="bibr" rid="ref20">20</xref>
          ].
        </p>
        <p>Coherence validation module: This module is aimed at evaluating aspects of consistency
and coherence of the answer  to prompt . Besides the answer , it also takes as input a
domain ontology  related to the topic of the prompt, here expected to be written in
DL and using certain known concepts, roles, and individuals.</p>
        <p>
          We assume here that the outcome  is in a form that makes it amenable to semantic
analysis in order to generate a formal response, as follows:
1. LLM outcome: The generated text by the LLM, , follows a structure and
vocabulary that allows one to extract a formalised version formul() written in the same
language as .
2. Semantic analysis: ∪{formul()} is evaluated for semantic defects which include
inconsistency but also weaker notions such as ‘of topic’ or ‘incoherent’.
3. Coherence evaluation: The module may provide a coherence score (or other
quantitative evaluation metrics) indicating the degree of alignment or agreement between
the LLM outcome, domain ontology, and other constraints. Such metrics should
helps assess the quality and reliability of the generated text and help steer the
feedback. Scores for coherence were for instance proposed in [
          <xref ref-type="bibr" rid="ref21">21</xref>
          ].
4. Feedback  about incoherence: If incoherences are detected, the module may
provide feedback  highlighting the areas where the LLM outcome diverges from
the ontology or violates logical rules. This feedback can be used to refine the
generated text or improve the LLM’s understanding of the domain.
        </p>
        <p>
          Verbalizer verbal: The use of verbalisation techniques for translating symbolic facts,
ontological rules and logic entailments into natural language is a core aspect of the workflow.
Verbalisations are readily available for the DL framework; in the simplest form,
Manchester Syntax can be almost directly translated to regular English [
          <xref ref-type="bibr" rid="ref22">22</xref>
          ].
        </p>
        <p>
          Prompt adaptation  = makeprompt(): We create a speech act generated from the
semantic analysis. This can be a verbalisation of an explanation of a proof, an announcement
that certain facts need to be accepted, or that other facts need to be rejected.
Extracting formal explanations is arguably a challenge of its own [
          <xref ref-type="bibr" rid="ref23 ref24">23, 24</xref>
          ]. Fortunately,
we are not interested in a post hoc explanation of the response of the LLM, but in
logical derivations of the found inconsistency. Some simple forms of ‘explanation’ can be
considered, like the extraction of a minimal inconsistent set, for which there are eficient
methods (e.g., [
          <xref ref-type="bibr" rid="ref25">25</xref>
          ]). Thus, in the workflow,  could be for example a minimal
inconsistent set of  ∪ {formul()}, and the new prompt  = makeprompt() could be
“But verbal(¬)!". To further help the LLM, we might want to suggest a repair of the
inconsistency, and perhaps some weakened assumptions of the claims that the LLM had
made, using, e.g., the repair and weakening techniques of [
          <xref ref-type="bibr" rid="ref17">17</xref>
          ].
        </p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Possible limitations</title>
      <p>
        Our approach itself certainly has some limitations. Some of those possible limitations concern
assumptions we are making about the future (capabilities) of LLMs themselves.
• Our workflow involves some sort of (automated) “arguing” with the LLM. We started
this note by reporting on the lack of accuracy of LLMs. And yet, we must rely on some
accuracy, or at least logical consistency. Indeed, for our workflow to work as expected,
the LLM would need to have a basic ‘understanding’ of logic. (E.g., the updated prompt
attempting to point out a contradiction to the LLM: “But, verbal(¬)!”) Unfortunately,
current LLMs are deficient in this regard [
        <xref ref-type="bibr" rid="ref26 ref27">26, 27</xref>
        ] and perform especially poorly in the
presence of negations [
        <xref ref-type="bibr" rid="ref28">28</xref>
        ]. The logical understanding they have is limited to statistical
patterns in language rather than true logical comprehension. However, improving exactly
this skill is a core research problem in the field [
        <xref ref-type="bibr" rid="ref29">29</xref>
        ].
• Abstraction (and, thus, logic) is still very dificult to handle by LLMs, as is clear also
from studying their mathematics capabilities [
        <xref ref-type="bibr" rid="ref27">27</xref>
        ] but also when describing verbally the
description of an object using variables. To illustrate this, Fig 3a shows the result of a
text-to-image generation where the text prompt is a verbalisation of a formalisation of
the concept ‘fishvehicle’ as created by a symbolic blending algorithm and using phrases
such as ‘the object  should be such that [. . . ]’. In contrast, Fig 3b shows the direct
text-to-image production using ‘a fish that is also a vehicle’, relying directly on the bias
of the model what a ‘fish’ respectively a ‘vehicle’ look like. Both artefacts were produced
with SDXL-Lightning1.
1See https://huggingface.co/spaces/ByteDance/SDXL-Lightning.
      </p>
      <p>(a) A symbolic representation of a ‘fishvehicle’
produced by the blending algorithm of [33],
verbalised and fed back into a text-to-image
generation algorithm.</p>
      <p>(b) A textual representation of a ‘fishvehicle’,
namely described as ‘a fish that is also a
vehicle’, fed directly into a text-to-image
generation algorithm.</p>
      <p>
        • LeCun [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] recommends to abandon auto-regressive LLMs for text generation altogether.
      </p>
      <p>Our approach does not act upon this recommendation. Instead it tries (maybe naively) to
put back text generation on the path of truth.</p>
      <p>
        To summarise, to be put in practice, one needs the following elements:
• A domain ontology in a signature  for every domain  that is addressed in prompt .
• A reasoner for the specific Description Logic (or corresponding OWL profile) in which
the domain ontology is written. Fortunately, consistency checking is a standard reasoning
task, and eficient reasoners exist for DLs (e.g., Hermit, Fact++, Pellet), but also to some
extent for First Order Logic (e.g., Vampire, Z3).
• A verbalizer verbal, to transform a set of logical formulas in Description Logic over
the signature  into natural language. Some readily available technologies have been
proposed for controlled natural languages. Examples are the OWL-Verbalizer [
        <xref ref-type="bibr" rid="ref30">30</xref>
        ] for
Attempto Controlled English (ACE),2 or the mapping proposed by Cregan et al. [
        <xref ref-type="bibr" rid="ref31">31</xref>
        ]
between the Sydney OWL Syntax and OWL 1.1 functional syntax. The NaturalOWL
System [
        <xref ref-type="bibr" rid="ref32">32</xref>
        ] is specifically aimed for generating coherent multi-sentence translations of
OWL axioms.
• A formulizer formul, to transform English text of the domain  in a logical representation,
in Description Logic, over the signature . If we can assume that the answer provided
by the LLM is in controlled English, then the verbalizer, like Kaljurand’s OWL-Verbalizer
is reversible, meaning that it can convert ACE English back into (ACE) OWL.
      </p>
    </sec>
    <sec id="sec-4">
      <title>4. Outlook</title>
      <p>We are interested in addressing the challenge of making LLMs more reliable. This paper lays
the groundwork for future research by proposing a preliminary roadmap. To this end, we have
proposed a high-level architecture for interacting with an LLM through a conversational pipeline</p>
      <sec id="sec-4-1">
        <title>2See also https://www.w3.org/2001/sw/wiki/ACE.</title>
        <p>that incorporates artificial speech acts, including feedback from symbolic components. To fully
assess the potential of this architecture, concrete examples and instantiations are needed.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>Acknowledgments</title>
      <p>The authors thank Ruslan Idelfonso Magaa Vsevolodovna for valuable feedback.
[33] G. Righetti, D. Porello, N. Troquard, O. Kutz, M. Hedblom, P. Galliani, Asymmetric Hybrids:
Dialogues for Computational Concept Combination, in: B. Brodaric, F. Neuhaus (Eds.),
12th International Conference on Formal Ontology in Information Systems - FOIS 2021,
Frontiers in Artificial Intelligence and Applications, IOS Press, 2021.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>S.</given-names>
            <surname>Minaee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Mikolov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Nikzad</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Chenaghlu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Socher</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Amatriain</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Gao</surname>
          </string-name>
          ,
          <article-title>Large language models: A survey</article-title>
          ,
          <source>arXiv preprint arXiv:2402.06196</source>
          (
          <year>2024</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>M.</given-names>
            <surname>Funk</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Hosemann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. C.</given-names>
            <surname>Jung</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Lutz</surname>
          </string-name>
          ,
          <article-title>Towards ontology construction with language models</article-title>
          ,
          <source>arXiv preprint arXiv:2309.09898</source>
          (
          <year>2023</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>H.</given-names>
            <surname>Babaei Giglou</surname>
          </string-name>
          ,
          <string-name>
            <surname>J. D'Souza</surname>
            ,
            <given-names>S. Auer,</given-names>
          </string-name>
          <article-title>LLMs4OL: Large language models for ontology learning</article-title>
          , in: International Semantic Web Conference, Springer,
          <year>2023</year>
          , pp.
          <fpage>408</fpage>
          -
          <lpage>427</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>I. O.</given-names>
            <surname>Gallegos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. A.</given-names>
            <surname>Rossi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Barrow</surname>
          </string-name>
          ,
          <string-name>
            <surname>M. M. Tanjim</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Kim</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Dernoncourt</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Yu</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Zhang</surname>
            ,
            <given-names>N. K.</given-names>
          </string-name>
          <string-name>
            <surname>Ahmed</surname>
          </string-name>
          ,
          <article-title>Bias and fairness in large language models: A survey</article-title>
          ,
          <source>CoRR abs/2309</source>
          .00770 (
          <year>2023</year>
          ). URL: https://doi.org/10.48550/arXiv.2309.00770. doi:
          <volume>10</volume>
          .48550/ ARXIV.2309.00770. arXiv:
          <volume>2309</volume>
          .
          <fpage>00770</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>S.</given-names>
            <surname>Lin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Hilton</surname>
          </string-name>
          ,
          <string-name>
            <surname>O. Evans,</surname>
          </string-name>
          <article-title>TruthfulQA: Measuring how models mimic human falsehoods</article-title>
          , in: S. Muresan,
          <string-name>
            <given-names>P.</given-names>
            <surname>Nakov</surname>
          </string-name>
          ,
          <string-name>
            <surname>A</surname>
          </string-name>
          . Villavicencio (Eds.),
          <source>Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume</source>
          <volume>1</volume>
          :
          <string-name>
            <surname>Long</surname>
            <given-names>Papers)</given-names>
          </string-name>
          ,
          <source>Association for Computational Linguistics</source>
          , Dublin, Ireland,
          <year>2022</year>
          , pp.
          <fpage>3214</fpage>
          -
          <lpage>3252</lpage>
          . URL: https://aclanthology. org/
          <year>2022</year>
          .
          <article-title>acl-long</article-title>
          .
          <volume>229</volume>
          . doi:
          <volume>10</volume>
          .18653/v1/
          <year>2022</year>
          .
          <article-title>acl-long</article-title>
          .
          <volume>229</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>H.</given-names>
            <surname>Alkaissi</surname>
          </string-name>
          ,
          <string-name>
            <surname>S. McFarlane</surname>
          </string-name>
          , Artificial Hallucinations in ChatGPT: Implications in Scientific Writing,
          <source>Cureus</source>
          <volume>15</volume>
          (
          <year>2023</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>Y. LeCun</given-names>
            ,
            <surname>Objective-Driven</surname>
          </string-name>
          <string-name>
            <surname>AI</surname>
          </string-name>
          :
          <article-title>Towards Machines that Can Learn, Reason, and</article-title>
          <string-name>
            <surname>Plan</surname>
          </string-name>
          ,
          <year>2024</year>
          . AAAI/IAAI Invited Talk.
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>J.</given-names>
            <surname>Yin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Jiang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Lu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Shang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <article-title>Neural generative question answering</article-title>
          , in: M.
          <string-name>
            <surname>Iyyer</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          <string-name>
            <surname>He</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Boyd-Graber</surname>
          </string-name>
          , H. Daumé III (Eds.),
          <source>Proceedings of the Workshop on Human-Computer Question Answering, Association for Computational Linguistics</source>
          , San Diego, California,
          <year>2016</year>
          , pp.
          <fpage>36</fpage>
          -
          <lpage>42</lpage>
          . URL: https://aclanthology.org/W16-0106. doi:
          <volume>10</volume>
          . 18653/v1/
          <fpage>W16</fpage>
          -0106.
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>B.</given-names>
            <surname>Bi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Yan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Xia</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <article-title>Incorporating external knowledge into machine reading for generative question answering</article-title>
          , in: K. Inui,
          <string-name>
            <given-names>J.</given-names>
            <surname>Jiang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Ng</surname>
          </string-name>
          ,
          <string-name>
            <surname>X.</surname>
          </string-name>
          Wan (Eds.),
          <source>Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP)</source>
          ,
          <article-title>Association for Computational Linguistics</article-title>
          , Hong Kong, China,
          <year>2019</year>
          , pp.
          <fpage>2521</fpage>
          -
          <lpage>2530</lpage>
          . URL: https://aclanthology.org/D19-1255. doi:
          <volume>10</volume>
          .18653/v1/
          <fpage>D19</fpage>
          -1255.
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>C.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Bi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Yan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Huang</surname>
          </string-name>
          ,
          <article-title>Addressing semantic drift in generative question answering with auxiliary extraction</article-title>
          , in: C.
          <string-name>
            <surname>Zong</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Xia</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          <string-name>
            <surname>Li</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Navigli</surname>
          </string-name>
          (Eds.),
          <source>Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing</source>
          (Volume
          <volume>2</volume>
          :
          <string-name>
            <surname>Short</surname>
            <given-names>Papers)</given-names>
          </string-name>
          ,
          <source>Association for Computational Linguistics</source>
          , Online,
          <year>2021</year>
          , pp.
          <fpage>942</fpage>
          -
          <lpage>947</lpage>
          . URL: https: //aclanthology.org/
          <year>2021</year>
          .acl-short.
          <volume>118</volume>
          . doi:
          <volume>10</volume>
          .18653/v1/
          <year>2021</year>
          .acl-short.
          <volume>118</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>A.</given-names>
            <surname>Martino</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Iannelli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Truong</surname>
          </string-name>
          ,
          <article-title>Knowledge injection to counter large language model (LLM) hallucination</article-title>
          , in: C. Pesquita,
          <string-name>
            <given-names>H.</given-names>
            <surname>Skaf-Molli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Efthymiou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Kirrane</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Ngonga</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Collarana</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Cerqueira</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Alam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Trojahn</surname>
          </string-name>
          , S. Hertling (Eds.),
          <source>The Semantic Web: ESWC 2023 Satellite Events - Hersonissos</source>
          , Crete, Greece, May 28 - June 1,
          <year>2023</year>
          , Proceedings, volume
          <volume>13998</volume>
          of Lecture Notes in Computer Science, Springer,
          <year>2023</year>
          , pp.
          <fpage>182</fpage>
          -
          <lpage>185</lpage>
          . URL: https: //doi.org/10.1007/978-3-
          <fpage>031</fpage>
          -43458-7_
          <fpage>34</fpage>
          . doi:
          <volume>10</volume>
          .1007/978-3-
          <fpage>031</fpage>
          -43458-7\_
          <fpage>34</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>P. S. H.</given-names>
            <surname>Lewis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Perez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Piktus</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Petroni</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Karpukhin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Goyal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Küttler</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Lewis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Yih</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Rocktäschel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Riedel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Kiela</surname>
          </string-name>
          ,
          <article-title>Retrieval-augmented generation for knowledge-intensive NLP tasks</article-title>
          , in: H.
          <string-name>
            <surname>Larochelle</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Ranzato</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Hadsell</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Balcan</surname>
          </string-name>
          , H. Lin (Eds.),
          <source>Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems</source>
          <year>2020</year>
          ,
          <article-title>NeurIPS 2020</article-title>
          , December 6-
          <issue>12</issue>
          ,
          <year>2020</year>
          , virtual,
          <year>2020</year>
          . URL: https://proceedings.neurips.cc/paper/2020/hash/ 6b493230205f780e1bc26945df7481e5-Abstract.html.
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Ji</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Yu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Ishii</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Fung</surname>
          </string-name>
          ,
          <article-title>Towards mitigating LLM hallucination via self reflection</article-title>
          , in: H.
          <string-name>
            <surname>Bouamor</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Pino</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          Bali (Eds.),
          <source>Findings of the Association for Computational Linguistics: EMNLP</source>
          <year>2023</year>
          ,
          <article-title>Association for Computational Linguistics</article-title>
          , Singapore,
          <year>2023</year>
          , pp.
          <fpage>1827</fpage>
          -
          <lpage>1843</lpage>
          . URL: https://aclanthology.org/
          <year>2023</year>
          .findings-emnlp.
          <volume>123</volume>
          . doi:
          <volume>10</volume>
          .18653/v1/
          <year>2023</year>
          .findings-emnlp.
          <volume>123</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>B. A.</given-names>
            <surname>Galitsky</surname>
          </string-name>
          ,
          <string-name>
            <surname>Truth-O-Meter</surname>
          </string-name>
          :
          <article-title>Collaborating with LLM in Fighting its Hallucinations</article-title>
          ,
          <source>Preprints</source>
          (
          <year>2023</year>
          ). URL: https://doi.org/10.20944/preprints202307.
          <fpage>1723</fpage>
          .v1. doi:
          <volume>10</volume>
          .20944/ preprints202307.
          <fpage>1723</fpage>
          .
          <year>v1</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Ji</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Frieske</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Yu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Su</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Ishii</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y. J.</given-names>
            <surname>Bang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Madotto</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Fung</surname>
          </string-name>
          ,
          <article-title>Survey of hallucination in natural language generation</article-title>
          ,
          <source>ACM Comput. Surv</source>
          .
          <volume>55</volume>
          (
          <year>2023</year>
          ). URL: https://doi.org/10.1145/3571730. doi:
          <volume>10</volume>
          .1145/3571730.
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>S.</given-names>
            <surname>Rudolph</surname>
          </string-name>
          , Foundations of Description Logics, Springer Berlin Heidelberg, Berlin, Heidelberg,
          <year>2011</year>
          , pp.
          <fpage>76</fpage>
          -
          <lpage>136</lpage>
          . URL: https://doi.org/10.1007/978-3-
          <fpage>642</fpage>
          -23032-
          <issue>5</issue>
          _2. doi:
          <volume>10</volume>
          .1007/ 978-3-
          <fpage>642</fpage>
          -23032-
          <issue>5</issue>
          _
          <fpage>2</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>N.</given-names>
            <surname>Troquard</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Confalonieri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Galliani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Penaloza</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Porello</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Kutz</surname>
          </string-name>
          , Repairing Ontologies via Axiom Weakening,
          <source>in: Proc. of the Thirty-Second AAAI Conference on Artificial Intelligence</source>
          , AAAI Press,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <surname>R. R. de Azevedo</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Freitas</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Rocha</surname>
            ,
            <given-names>J. A</given-names>
          </string-name>
          .
          <string-name>
            <surname>A. de Menezes</surname>
            ,
            <given-names>L. F. A.</given-names>
          </string-name>
          <string-name>
            <surname>Pereira</surname>
          </string-name>
          ,
          <article-title>Generating description logic ℒ from text in natural language</article-title>
          , in: T. Andreasen,
          <string-name>
            <given-names>H.</given-names>
            <surname>Christiansen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.-C.</given-names>
            <surname>Cubero</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z. W.</given-names>
            <surname>Raś</surname>
          </string-name>
          (Eds.),
          <source>Foundations of Intelligent Systems</source>
          , Springer International Publishing, Cham,
          <year>2014</year>
          , pp.
          <fpage>305</fpage>
          -
          <lpage>314</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>B.</given-names>
            <surname>Gyawali</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Shimorina</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Gardent</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Cruz-Lara</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Mahfoudh</surname>
          </string-name>
          ,
          <article-title>Mapping natural language to description logic</article-title>
          ,
          <source>in: The Semantic Web: 14th International Conference, ESWC</source>
          <year>2017</year>
          , Portorož, Slovenia, May 28-June 1,
          <year>2017</year>
          , Proceedings,
          <source>Part I 14</source>
          , Springer,
          <year>2017</year>
          , pp.
          <fpage>273</fpage>
          -
          <lpage>288</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>A.</given-names>
            <surname>Ranta</surname>
          </string-name>
          ,
          <article-title>Translating between language and logic: What is easy and what is dificult</article-title>
          , in: N.
          <string-name>
            <surname>Bjørner</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          <string-name>
            <surname>Sofronie-Stokkermans</surname>
          </string-name>
          (Eds.),
          <source>Automated Deduction - CADE-23</source>
          , Springer Berlin Heidelberg, Berlin, Heidelberg,
          <year>2011</year>
          , pp.
          <fpage>5</fpage>
          -
          <lpage>25</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>R.</given-names>
            <surname>Confalonieri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Kutz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Troquard</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Galliani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Porello</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Peñaloza</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Schorlemmer</surname>
          </string-name>
          , Coherence, Similarity, and Concept Generalisation, in: A.
          <string-name>
            <surname>Artale</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Glimm</surname>
          </string-name>
          , R. Kontchakov (Eds.),
          <source>Proc. of the 30th International Workshop on Description Logics (DL</source>
          <year>2017</year>
          ), volume
          <year>1879</year>
          ,
          <article-title>CEUR-WS, Montpellier</article-title>
          , France,
          <source>July 18-21</source>
          ,
          <year>2017</year>
          ,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>M.</given-names>
            <surname>Horridge</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Drummond</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Goodwin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Rector</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Stevens</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <source>The Manchester OWL Syntax, in: OWL: Experiences and Directions</source>
          ,
          <year>2006</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <given-names>W.</given-names>
            <surname>Saeed</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C. W.</given-names>
            <surname>Omlin</surname>
          </string-name>
          ,
          <string-name>
            <surname>Explainable</surname>
            <given-names>AI</given-names>
          </string-name>
          (
          <article-title>XAI): A systematic meta-survey of current challenges and future opportunities</article-title>
          ,
          <source>Knowl. Based Syst</source>
          .
          <volume>263</volume>
          (
          <year>2023</year>
          )
          <article-title>110273</article-title>
          . doi:
          <volume>10</volume>
          .1016/ J.KNOSYS.
          <year>2023</year>
          .
          <volume>110273</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [24]
          <string-name>
            <given-names>S.</given-names>
            <surname>Ali</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Abuhmed</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. H. A.</given-names>
            <surname>El-Sappagh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Muhammad</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. M.</given-names>
            <surname>Alonso-Moral</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Confalonieri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Guidotti</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. D.</given-names>
            <surname>Ser</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N. D.</given-names>
            <surname>Rodríguez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Herrera</surname>
          </string-name>
          ,
          <article-title>Explainable artificial intelligence (XAI): what we know and what is left to attain trustworthy artificial intelligence</article-title>
          ,
          <source>Inf. Fusion</source>
          <volume>99</volume>
          (
          <year>2023</year>
          )
          <article-title>101805</article-title>
          . URL: https://doi.org/10.1016/j.infus.
          <year>2023</year>
          .
          <volume>101805</volume>
          . doi:
          <volume>10</volume>
          .1016/J.INFFUS.
          <year>2023</year>
          .
          <volume>101805</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          [25]
          <string-name>
            <surname>K. M. Shchekotykhin</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Jannach</surname>
          </string-name>
          , T. Schmitz, Mergexplain:
          <article-title>Fast computation of multiple conflicts for diagnosis</article-title>
          , in: Q.
          <string-name>
            <surname>Yang</surname>
            ,
            <given-names>M. J.</given-names>
          </string-name>
          <string-name>
            <surname>Wooldridge</surname>
          </string-name>
          (Eds.),
          <source>Proceedings of the TwentyFourth International Joint Conference on Artificial Intelligence, IJCAI</source>
          <year>2015</year>
          ,
          <string-name>
            <given-names>Buenos</given-names>
            <surname>Aires</surname>
          </string-name>
          , Argentina,
          <source>July 25-31</source>
          ,
          <year>2015</year>
          , AAAI Press,
          <year>2015</year>
          , pp.
          <fpage>3221</fpage>
          -
          <lpage>3228</lpage>
          . URL: http://ijcai.org/Abstract/ 15/454.
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          [26]
          <string-name>
            <given-names>M. E.</given-names>
            <surname>Jang</surname>
          </string-name>
          , T. Lukasiewicz,
          <source>Consistency Analysis of ChatGPT</source>
          ,
          <year>2023</year>
          . arXiv:
          <volume>2303</volume>
          .
          <fpage>06273</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          [27]
          <string-name>
            <given-names>S.</given-names>
            <surname>Frieder</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Berner</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Petersen</surname>
          </string-name>
          , T. Lukasiewicz,
          <source>Large language models for mathematicians</source>
          ,
          <year>2024</year>
          . arXiv:
          <volume>2312</volume>
          .
          <fpage>04556</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          [28]
          <string-name>
            <given-names>T. H.</given-names>
            <surname>Truong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Baldwin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Verspoor</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Cohn</surname>
          </string-name>
          ,
          <article-title>Language models are not naysayers: an analysis of language models on negation benchmarks</article-title>
          , in: A.
          <string-name>
            <surname>Palmer</surname>
          </string-name>
          , J. Camachocollados (Eds.),
          <source>Proceedings of the 12th Joint Conference on Lexical and Computational Semantics (*SEM</source>
          <year>2023</year>
          ),
          <article-title>Association for Computational Linguistics</article-title>
          , Toronto, Canada,
          <year>2023</year>
          , pp.
          <fpage>101</fpage>
          -
          <lpage>114</lpage>
          . URL: https://aclanthology.org/
          <year>2023</year>
          .starsem-
          <volume>1</volume>
          .10. doi:
          <volume>10</volume>
          .18653/v1/
          <year>2023</year>
          . starsem-
          <volume>1</volume>
          .
          <fpage>10</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          [29]
          <string-name>
            <given-names>L.</given-names>
            <surname>Pan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Albalak</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Wang</surname>
          </string-name>
          , Logic-LM:
          <article-title>Empowering large language models with symbolic solvers for faithful logical reasoning</article-title>
          , in: H.
          <string-name>
            <surname>Bouamor</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Pino</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          Bali (Eds.),
          <source>Findings of the Association for Computational Linguistics: EMNLP</source>
          <year>2023</year>
          ,
          <article-title>Association for Computational Linguistics</article-title>
          , Singapore,
          <year>2023</year>
          , pp.
          <fpage>3806</fpage>
          -
          <lpage>3824</lpage>
          . URL: https://aclanthology.org/
          <year>2023</year>
          .findings-emnlp.
          <volume>248</volume>
          . doi:
          <volume>10</volume>
          .18653/v1/
          <year>2023</year>
          .findings-emnlp.
          <volume>248</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          [30]
          <string-name>
            <given-names>K.</given-names>
            <surname>Kaljurand</surname>
          </string-name>
          ,
          <article-title>Attempto Controlled English as a Semantic Web Language</article-title>
          ,
          <source>Ph.D. thesis, Faculty of Mathematics and Computer Science</source>
          , University of Tartu,
          <year>2007</year>
          . URL: http: //attempto.ifi.uzh.ch/site/pubs/papers/phd_kaljurand.pdf.
        </mixed-citation>
      </ref>
      <ref id="ref31">
        <mixed-citation>
          [31]
          <string-name>
            <given-names>A.</given-names>
            <surname>Cregan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Schwitter</surname>
          </string-name>
          , T. Meyer, Sydney OWL syntax
          <article-title>- towards a controlled natural language syntax for OWL 1.1</article-title>
          , in: C.
          <string-name>
            <surname>Golbreich</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Kalyanpur</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          Parsia (Eds.),
          <source>Proceedings of the OWLED 2007 Workshop on OWL: Experiences and Directions</source>
          , Innsbruck, Austria, June 6-7,
          <year>2007</year>
          , volume
          <volume>258</volume>
          <source>of CEUR Workshop Proceedings, CEUR-WS.org</source>
          ,
          <year>2007</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref32">
        <mixed-citation>
          [32]
          <string-name>
            <given-names>I.</given-names>
            <surname>Androutsopoulos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Lampouras</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Galanis</surname>
          </string-name>
          ,
          <article-title>Generating natural language descriptions from OWL ontologies: the naturalowl system</article-title>
          ,
          <source>J. Artif. Intell. Res</source>
          .
          <volume>48</volume>
          (
          <year>2013</year>
          )
          <fpage>671</fpage>
          -
          <lpage>715</lpage>
          . URL: https://doi.org/10.1613/jair.4017. doi:
          <volume>10</volume>
          .1613/JAIR.4017.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>