<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Towards Explainable and Ontologically Grounded Language Models⋆</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Walid S. Saba</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Institute for Experiential AI, Northeastern University</institution>
          ,
          <addr-line>100 Fore St, Portland, ME 04101</addr-line>
          <country country="US">USA</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>We argue that the relative success of large language models (LLMs) is not a reflection on the symbolic vs. subsymbolic debate but a reflection on employing an appropriate bottom-up strategy of a reverse engineering of language at scale. However, and due to their subsymbolic nature whatever knowledge these systems acquire about language will always be buried in millions of weights none of which is meaningful on its own, rendering such systems utterly unexplainable. Furthermore, and due to their stochastic nature, LLMs will often fail in making the correct inferences in various linguistic contexts that require reasoning in intensional, temporal, or modal contexts. To remedy these shortcomings we suggest employing the successful bottomup strategy employed in LLMs but in a symbolic setting, resulting in explainable, languageagnostic, and ontologically grounded language models.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Large language models</kwd>
        <kwd>ontology</kwd>
        <kwd>bottom-up reverse engineering 1</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        To arrive at a scientific explanation there are
generally two approaches we can adopt, a top-down
approach or a bottom-up approach
        <xref ref-type="bibr" rid="ref20">(Salmon, 1989)</xref>
        .
However, for a top-down approach to work, there
must be a set of established general principles that
one can start with, which is clearly not the case when
it comes to language and how our minds externalize
our thoughts in language. In retrospect, therefore, it is
not surprising that decades of top-down work in
natural language processing (NLP) failed to produce
satisfactory results since most of this work was
inspired by theories that made questionable
assumptions where, for example, an innate universal
grammar was assumed
        <xref ref-type="bibr" rid="ref4">(Chomsky, 1957)</xref>
        , or that we
metaphorically build our linguistic competence based
on a set of idealized cognitive models
        <xref ref-type="bibr" rid="ref11">(Lakoff, 1987)</xref>
        ,
or that natural language could be formally described
using the tools of formal logic
        <xref ref-type="bibr" rid="ref17">(Montague, 1973)</xref>
        . In a
similar vein, it is perhaps for the same reason that
decades of top-down work in ontology and knowledge
representation
        <xref ref-type="bibr" rid="ref12 ref23">(Lenat and Guha, 1990 and Sowa,
1995)</xref>
        also faltered since most of this work amounted
to pushing, in a top-down manner, metaphysical
theories of how the world is supposedly structured
and represented in our minds, and again without any
agreed upon general principles to start with. On the
other hand, unprecedented progress has been made
in only a few years of NLP work that employed a
datadriven bottom-up strategy, as exemplified by recent
advances in large language models (LLMs) that are
essentially a massive experiment of a bottom-up
reverse engineering of language at scale (e.g., ChatGPT
and GPT-4)2.
      </p>
      <sec id="sec-1-1">
        <title>1.1. Issues with LLMs</title>
        <p>
          Despite their relative success, LLMs do not tell us
anything about how language works since these
models are not really models of language but are
statistical models of regularities found in language3. In
2 GPT stands for ‘Generative Pre-trained Transformer’, an
architecture that OpenAI built on top of the transformer
architecture
          <xref ref-type="bibr" rid="ref25">(Vaswani, A. et. al., 2017)</xref>
          .
3 In looking inside the neural network (NN) of an LLM one does
not find concepts, meanings, linguistic structures, etc. but
weights associated with neural connections, which is exactly
what one will find in an object recognition or any other NN.
fact, and due to their subsymbolic nature, whatever
‘knowledge’ these models acquire about language will
always be buried in millions of weights
(microfeatures) none of which is meaningful on its
own, rendering these models utterly unexplainable
          <xref ref-type="bibr" rid="ref13 ref8">(Guizzardia and Guarino, 2024)</xref>
          . Besides
unexplainability, LLMs are also oblivious to truth
          <xref ref-type="bibr" rid="ref3">(Borji, 2023)</xref>
          , since for LLMs all text (factual or
nonfactual), is treated equally. Finally, and while LLMs
have been shown to do poorly in a number of tasks
that require high-level reasoning such as planning
          <xref ref-type="bibr" rid="ref24">(Valmeekam et. al., 2023)</xref>
          , analogies
          <xref ref-type="bibr" rid="ref13 ref8">(Lewis and
Mitchell, 2024)</xref>
          and formal reasoning (Arkoudas,
2023) what concerns here is the failure of LLMs in
making the right inferences in various linguistic
contexts. As an illustration of the kinds of failures in
deep language understanding we consider here three
linguistic contexts involving copredication, intension
and prepositional attitudes.
        </p>
        <p>Example 1. Show the entities and the relations that are
implicit in the following text: “I threw away the
newspaper I was reading because they fired my
favorite columnist”.</p>
        <p>
          Example 2. Since Madrid is the capital of Spain, can I
replace one for the other in the following: “Maria thinks
Madrid was not always the capital of Spain”?
Example 3. Suppose Devon knows that if someone is a
client, then s/he is a student, and suppose that Olga is a
client. Then what does Devon know?
The first example involves a phenomenon called
copredication
          <xref ref-type="bibr" rid="ref1">(see Asher and Pustejovsky, 2005)</xref>
          which occurs when the same entity is used in the same
context to refer to more than one semantic
(ontological) type. All LLMs tested4 failed in
recognizing that ‘newspaper’ in the text is used to
simultaneously refer to three entities: (i) the physical
object I threw away; (ii) the content of the newspaper
I was reading; and (iii) the ‘editorial board’ of the
newspaper that did the firing of the columnist. Note
that the failure of the LLMs was more acute when the
LLMs were asked to draw a graph showing all entities
and relations implied by the text since to show all the
relations in the text all the different types of entities
must be extracted. Here all LLMs tested showed the
same newspaper (physical) object doing the firing of
the columnist.
        </p>
        <p>
          In example 2 all LLMs we tested approved replacing
‘the capital of Spain’ by ‘Madrid’ resulting in ‘Maria
thinks that Madrid was not always Madrid’. It is worth
noting that the LLMs tested were consistently oblivion
to intension. For example, in ‘Perhaps Socrates was not
the tutor of Alexander the Great’, ‘Socrates’ and ‘the
tutor of Alexander the Great’ were also deemed
replaceable (since they are extensionally equal) resulting
in ‘Perhaps Socrates was not Socrates’. These results
were expected since neural networks (deep or
otherwise), that are the computing architecture behind
all LLMs, are purely extensional models and are based
on the ‘empiricist theory of abstraction’ where their
similarity semantics has no notion of ‘object identity’
          <xref ref-type="bibr" rid="ref14">(Lopes, 2023)</xref>
          .
        </p>
        <p>Finally, example 3 illustrates failures of LLMs in
making the correct inferences in modal (belief) contexts:
the response of the LLMs tested was that ‘Devon knows
that Olga is a student’ which is clearly the wrong
inference since inferring K(Devon, student(Olga)) from
K(Devon, client(Olga)student(Olga)) requires K(Devon,
client(Olga)), i.e., it requires Devon knowing that Olga is
a client. We have collected many other tests that we
make available elsewhere for the sake of saving space.5</p>
      </sec>
      <sec id="sec-1-2">
        <title>1.2. LLMs: A Glass Half Empty, Half Full</title>
        <p>So where do we stand now? On one hand, LLMs have
clearly proven that one can get a handle on syntax and
quite a bit of semantics in a bottom-up reverse
engineering of language at scale; yet on the other hand
what we have are unexplainable models that do not
shed any light on how language actually works.
Moreover, it would seem that due to their purely
extensional and statistical nature, LLMs will always
fail in making the correct inferences in many linguistic
contexts. Since we believe the relative success of LLMs
is not a reflection on the symbolic vs. subsymbolic
debate but is a reflection on a successful bottom-up
reverse engineering strategy, we think that combining
the advantages of symbolic and ontologically
grounded representations with a bottom-up reverse
engineering strategy is a worthwhile effort. In fact, the
idea that word meaning can be extracted from how
words are actually used in language is not exclusive to
linguistic work in the empirical tradition, but in fact it
can be traced back to Frege.</p>
        <p>In the rest of the paper we will (i) first argue that
current word embeddings that are the genesis of
modern-day large language models can be
constructed in a symbolic setting instead of being the
result of statistical cooccurrences; (ii) we will show
that symbolic vectors perform better than current
embeddings on a well-known word similarity
benchmark; (iii) we will discuss how our symbolic
4 Our experiments were conducted on GPT-4o (chat.openai.com).
5 https://shorturl.at/ejmH8
vectors can be used to discover the ontological
structure that is implicit in our ordinary language.</p>
      </sec>
    </sec>
    <sec id="sec-2">
      <title>2. Concerning 'the Company a Word</title>
    </sec>
    <sec id="sec-3">
      <title>Keeps'</title>
      <p>
        The genesis of modern LLMs is the distributional
semantics hypothesis which states that the more
semantically similar words are, the more they tend to
occur in similar contexts – or, similarity in meaning is
similarity in linguistic distribution
        <xref ref-type="bibr" rid="ref9">(Harris, 1954)</xref>
        .
This is usually summarized by a saying that is
attributed to the British linguist John R. Firth that “you
shall know a word by the company it keeps”. When
processing a large corpus, this idea can be used by
analyzing co-occurrences and contexts of use to
approximate word meanings by word embeddings
(vectors or tensors), that are essentially points in
multidimensional space. Thus, at the root of LLMs is a
bottom-up reverse engineering of language strategy
where, unlike top-down approaches, “reverse
engineers the process and induces semantic
representations from contexts of use”
        <xref ref-type="bibr" rid="ref2">(Boleda, 2020)</xref>
        .
But nothing precludes this idea from being carried out
in a symbolic setting. In other words, the ‘company a
word keeps’ can be measured in several ways, other
than the correlational and statistical measures that
underlie modern word embeddings.
      </p>
      <sec id="sec-3-1">
        <title>2.1. Symbolic Dimensions of Meaning</title>
        <p>
          In discussing possible models of the world that can be
employed in computational linguistics Hobbs (1985)
once suggested that there are two alternatives: (i) on
one extreme we could attempt building a “correct”
theory that would entail a full description of the
world, something that would involve quantum
physics and all the sciences; (ii) on the other hand, we
could have a promiscuous model of the world that is
isomorphic to the way we talk it about in natural
language (emphasis is ours). Since the first option is a
project that is most likely impossible to complete,
what Hobbs is clearly suggesting here is a reverse
engineering of language to discover how we actually
use language to talk about the world we live in. This is
also not much different from Frege’s Context Principal
that suggests “never ask for the meaning of words in
isolation”
          <xref ref-type="bibr" rid="ref5">(Dummett, 1981)</xref>
          but that a word gets its
meanings from analyzing all the contexts in which the
word can appear
          <xref ref-type="bibr" rid="ref16">(Milne, 1986)</xref>
          . Again, what this
suggests is that the meaning of words is embedded (to
use a modern terminology) in all the ways we use
these words in how we talk about the world. While
Hobbs’ and Frege’s observations might be a bit vague,
the proposal put forth by Fred Sommers (1963) was
very specific. Again, Sommers suggests that “to know
the meaning of a word is to know how to formulate
some sentences containing the word” and this would
lead, like in Frege’s case, to the conclusion that a
complete knowledge of some word w would be all the
ways w can be used. For Sommers, the process of
understanding the meaning of some word w starts by
analyzing all the properties P that can sensibly be
said of w. Thus, for example, [delicious Thursday] is not
sensible while [delicious apple] is, regardless of the
truth or falsity of the predication. Moreover, and since
[delicious cake] is also sensible, then there must be a
common type (perhaps food?) that subsumes both
apple and cake. This idea is similar to the idea of type
checking in strongly typed polymorphic programming
languages. For example, the types in an expression
such as ‘x + 3’ will only unify (or the expression will
only ‘make sense’) if/when x is an object of type
number (as opposed to a tuple, for example). As it was
suggested in
          <xref ref-type="bibr" rid="ref19">(Saba, 2007)</xref>
          , this type of analysis can
thus be used to ‘discover’ the ontology that seems to
be implicit in the language, as will be discussed below.
First, however, we describe how a bottom-up reverse
engineering of language can be done in a symbolic
setting.
        </p>
      </sec>
      <sec id="sec-3-2">
        <title>2.2. Symbolic Reverse Engineering of</title>
      </sec>
      <sec id="sec-3-3">
        <title>Language</title>
        <p>
          The procedure we have in mind assumes a
Platonic universe where all concepts, physical or
abstract, including states, activities, properties
(tropes)
          <xref ref-type="bibr" rid="ref15">(Moltmann, 2013)</xref>
          , processes, events, etc. are
considered entities that can be defined by a number of
language-agnostic primitives
          <xref ref-type="bibr" rid="ref21">(Smith, 2005)</xref>
          that we
call the ‘dimensions of meaning’. We consider here the
following dimensions: AGENTOF, OBJECTOF, HASPROP,
INSTATE, PARTOF, INSTATE, INPROCESS, and OFTYPE. For
every word w in the language, and for every
dimension D, a reverse-engineering process is
conducted to compute a set wD = {(x, t) | D(w, x)}
where t is a weight in [0,1]. Here are example sets
computed for ‘book’ along four dimensions of
meaning along with the masking prompt that queries
what an LLM has ‘learned’ about how we talk about
books:
book . HASPROP
Everyone likes to read a [MASK] book.
=&gt; {(popular, 0.9), (educational, 0.8), (famous, 0.8), ... }
book . OBJECTOF
Everyone I know enjoyed [MASK] ‘The Prince’.
=&gt; {(reading, 0.9), (writing, 0.8), (editing, 0.8), ... }
book . AGENTOF
Das Kapital has [MASK] many people over the years.
=&gt; {(influenced, 0.9), (inspired, 0.8), (changed, 0.8), ... }
book . PARTOF
Hamlet should be part of every [MASK].
=&gt; {(collection, 0.9), (archive, 0.8), (library, 0.8), ... }
book . INSTATE
I was told that my book is now in [MASK].
=&gt; {(print, 0.9), (circulation, 0.8), (review, 0.8), ... }
What the above says is the following (i) in ordinary
spoken language we speak of a ‘book’ that is popular,
educational, famous, etc.; (ii) we speak of reading,
writing, editing, etc. a ‘book’; (iii) we speak of ‘book’
that may change, influence, inspire, etc.; and (iv) we
speak of a b ‘book’ that is part of a collection, an
archive, or a library; and (v) a book can be in review, in
print, in circulation, etc. The nominalization process
can be conducted using the copular ‘is’ as shown in
table 1. For example, ‘John is famous’ can be restated
as ‘John has the property of fame’; ‘Jim is sad’ as ‘Jim is
in a state of sadness’; etc.
          <xref ref-type="bibr" rid="ref1 ref15 ref21">(see [Smith, 2005] for more
on the relationship between the copular and abstract
entities and [Moltmann, 2013] for more on abstract
objects.)</xref>
          What should be noted here is that even with
the simple conceptual structure discovered thus far
one can generate plausible text, such as the following:
(1) enjoyed the interesting reading of the new book
(2) completed a boring reading of a controversial book
The sensible (and meaningful) fragment in (1) can be
generated because a book can be ‘read’ and described
by ‘new’, and readings can be ‘interesting’ and the
object of enjoyment; and similarly for (2) where a
reading of a controversial book can be boring and the
object of a completion, etc. Note, however, that text
generation in this case is not a function of ‘predicting’
the most likely continuation, but a function of
plausible filling in of subjects, objects, agents,
descriptions, etc. to any propositional structure.
        </p>
      </sec>
      <sec id="sec-3-4">
        <title>2.3. Symbolic Embeddings</title>
        <p>
          The process we described thus far results in symbolic
word embeddings as the one shown in figure 2 below.
In figure 2(a) we show the symbolic embedding for
6 https://kaggle.com/datasets/julianschelb/wordsim353-crowd
‘boy’ and ‘lad’ along the HASPROP dimension. Thus, in
ordinary spoken language it is sensible to speak of a
‘handsome boy’ and a ‘funny boy’ as well as a ‘clever
lad’ and a ‘talented lad’. We note here that in this
process generic descriptions are removed using a
function that computes the information content of
some adjectives, where the information content of an
adjective adj is inversely proportional to the set of
types of adj can sensibly be applied to. For example,
‘beautiful’ will have a low information content score
since ‘beautiful’ can sensibly be said of many concepts,
both physical and abstract (e.g., car, movie, poem,
night, girl, …) while ‘tasty’ can sensibly be said of ‘food’
and just a few others. The symbolic embeddings in
figure 2(b) are those of ‘automobile’ and ‘car’ along
the OBJECTOF dimension. Note now that word
similarity along these symbolic dimensions can be
computed using cosine similarity as well as weighted
Jaccard similarity where max and min can be used in
fuzzy union and fuzzy intersection. We are currently
experimenting with the optimal number of
dimensions using a number of word similarity
benchmarks, including the WordSim353 dataset
          <xref ref-type="bibr" rid="ref6">(Finkelstein, Lev. et al., 2001)</xref>
          6.
        </p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>3. The Ontology of the Language of</title>
    </sec>
    <sec id="sec-5">
      <title>Thought?</title>
      <p>The reverse engineering process we have described
above would result in symbolic embeddings along
various dimensions, as the ones shown in figure 2. As
a result of this, however, we could then analyze the
subset relations between these embeddings to
discover the ontological structure that seems to be
implicit in our ordinary language. To illustrate,
consider the following:
(3) car . objectOf</p>
      <p>= {(driving, 0.9), (repairing, 0.8), (buying, 0.8), ... }
(4) book . objectOf</p>
      <p>= {(reading, 0.9), (writing, 0.8), (buying, 0.8), ... }
(5) person . AGENTOF</p>
      <p>= {(reading, 0.9),(writing, 0.8), (driving, 0.8), ... }
(6) person. HASPROP</p>
      <p>= {(popularity, 0.9), (fame, 0.8), (beautiful, 0.8), .. }
(7) car. hasProp</p>
      <p>= {(popularity, 0.9), (fame, 0.8), (beautiful, 0.8), .. }
(8) book . HASPROP</p>
      <p>= {(popularity, 0.9), (fame, 0.8), (beautiful, 0.8), .. }
Note that car can be the object of ‘buying’ and so can
be a book and this means that car and book must, at
some level of abstraction, share the same parent
(perhaps ‘artifact’?) Note also that a car as well as a
book and a person can be popular. An analysis along
these lines would result in the following:
(9) read(person, book)
(10) write(person, book)
(11) buy(person, T1 = car  book … )
(12) drive(person, car)
(13) beautiful(T2 = person  car  book … )
What the above says is the following: in ordinary
spoken language we speak of people reading and
writing books (9 and 10); we speak of people buying
cars and books, and thus of buying objects that are of
some type that subsumes both cars and books (11);
we speak of people driving cars (12); and we speak of
beautiful people, cars, and books (and thus beautiful
seems to be a property that can sensibly be said of
concepts that are at very high level of generality). As
suggested by Sommers (1963) this type of analysis
that can be fully automated with the help of LLMs can
help us discover what he called ‘the Tree of Language’
– which is essentially the ontology that seems to be
underneath our ordinary language. This might also be
what Hobbs (1985) was seeking when he suggested
building a model of the world that isomorphic to the
we talk about it in natural language.</p>
    </sec>
    <sec id="sec-6">
      <title>4. Concluding Remarks</title>
      <p>Large language models (LLMs) have shown
impressive capabilities that pioneers in artificial
intelligence and natural language processing would
marvel at.
(a)
(b)
However, we believe that LLMs are not the answer to
the language understanding problem nor to reasoning
in general and in particular commonsense reasoning.
Due to their paradigmatic unexplainability LLMs will
also not shed any light on how language works and
how we externalize our thoughts in language. Since, in
our opinion, the relative success of LLMs is not due to
their subsymbolic nature but due to applying a
successful bottom-up reverse engineering strategy,
we suggested here applying the same strategy but in a
symbolic setting, something that has been argued for
by logicians dating back to Frege. By combining the
successful bottom-up strategy and symbolic and
ontological methods we arrive at explainable and
ontologically grounded language models that can be
used in problems requiring commonsense reasoning.</p>
      <p>We are still in the early stage of this work, but we
currently have the tools to realize the dream of Frege
and Sommers and perhaps shed some light on the
‘language of thought’ Fodor (1998) – the internal
language that we use to construct and process our
thoughts.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>Nicholas</given-names>
            <surname>Asher</surname>
          </string-name>
          and
          <string-name>
            <given-names>James</given-names>
            <surname>Pustejovsky</surname>
          </string-name>
          .
          <year>2005</year>
          .
          <article-title>Word Meaning and Commonsense Metaphysics</article-title>
          . In:
          <article-title>Course Materials for Type Selection and the Semantics of Local Context</article-title>
          ,
          <string-name>
            <surname>ESSLLI</surname>
          </string-name>
          <year>2005</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>Gemma</given-names>
            <surname>Boleda</surname>
          </string-name>
          .
          <year>2020</year>
          .
          <article-title>Distributional Semantics and Linguistic Theory, Annual Review of Linguistics, 6</article-title>
          , pp.
          <fpage>213</fpage>
          -
          <lpage>34</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <surname>Borji</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <year>2023</year>
          .
          <article-title>A Categorical Archive of ChatGPT Failures</article-title>
          , Available online at https://arxiv.org/abs/2302.03494
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>Noam</given-names>
            <surname>Chomsky</surname>
          </string-name>
          .
          <year>1957</year>
          . Syntactic Structures, Mouton de Gruyter, NY.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>Michael</given-names>
            <surname>Dummett</surname>
          </string-name>
          .
          <year>1981</year>
          .
          <article-title>Frege: Philosophy of Language</article-title>
          . Harvard University Press.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <surname>Finkelstein</surname>
          </string-name>
          ,
          <string-name>
            <surname>Lev</surname>
          </string-name>
          , et al.
          <article-title>"Placing search in context: The concept revisited</article-title>
          .
          <source>" Proceedings of the 10th international conference on the World Wide Web. ACM</source>
          ,
          <year>2001</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>Jerry</given-names>
            <surname>Fodor</surname>
          </string-name>
          ,
          <year>1988</year>
          . Concepts:
          <source>Where Cognitive Science Went Wrong</source>
          , Oxford University Press.
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>Giancarlo</given-names>
            <surname>Guizzardia</surname>
          </string-name>
          and
          <string-name>
            <given-names>Nicola</given-names>
            <surname>Guarino</surname>
          </string-name>
          ,
          <year>2024</year>
          . Semantics, Ontology, and
          <string-name>
            <surname>Explanation</surname>
          </string-name>
          , Data &amp; Knowledge
          <string-name>
            <surname>Engineering</surname>
          </string-name>
          (to appear).
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <surname>Zellig</surname>
            <given-names>S. Harris. 1954. Distributional</given-names>
          </string-name>
          <string-name>
            <surname>Structure</surname>
          </string-name>
          .
          <source>Word 10</source>
          , pp.
          <fpage>146</fpage>
          -
          <lpage>62</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>Jerry</given-names>
            <surname>Hobbs</surname>
          </string-name>
          .
          <year>1985</year>
          .
          <article-title>Ontological promiscuity</article-title>
          .
          <source>In Proceedings. of the 23rd Annual Meeting of the Association for Computational Linguistics</source>
          , Chicago, Illinois,
          <year>1985</year>
          , pp.
          <fpage>61</fpage>
          -
          <lpage>69</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>George</given-names>
            <surname>Lakoff</surname>
          </string-name>
          .
          <year>1987</year>
          . Women, Fire, and Dangerous Things:
          <article-title>What Categories Reveal About the Mind</article-title>
          , University of Chicago Press.
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>Doug</given-names>
            <surname>Lenat</surname>
          </string-name>
          and Guha,
          <string-name>
            <surname>R. V.</surname>
          </string-name>
          <year>1990</year>
          .
          <article-title>Building Large Knowledge-Based Systems: Representation and Inference in the CYC Project</article-title>
          .
          <article-title>Addison-Wesley.</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>Martha</given-names>
            <surname>Lewis</surname>
          </string-name>
          and Melanie Mitchell.
          <year>2024</year>
          .
          <article-title>Using Counterfactual Tasks to Evaluate the Generality of Analogical Reasoning in LLMs</article-title>
          , https://arxiv.org/abs/2402.08955
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>Jesse</given-names>
            <surname>Lopes</surname>
          </string-name>
          .
          <year>2023</year>
          .
          <article-title>Can Deep CNNs Avoid Infinite Regress/Circularity in Content Constitution? Minds and Machines</article-title>
          , vol.
          <volume>33</volume>
          , pp.
          <fpage>507</fpage>
          -
          <lpage>524</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>Friederike</given-names>
            <surname>Moltmann</surname>
          </string-name>
          ,
          <year>2013</year>
          .
          <article-title>Abstract Objects and the Semantics of Natural Language</article-title>
          , Oxford University Press.
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>Peter</given-names>
            <surname>Milne</surname>
          </string-name>
          .
          <year>1986</year>
          .
          <article-title>Frege's Context Principle</article-title>
          , Mind, Vol.
          <volume>95</volume>
          , No.
          <volume>380</volume>
          , pp.
          <fpage>491</fpage>
          -
          <lpage>495</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>Richard</given-names>
            <surname>Montague</surname>
          </string-name>
          .
          <year>1973</year>
          .
          <article-title>The Proper Treatment of Quantification in Ordinary English</article-title>
          . In: Kulas,
          <string-name>
            <given-names>J.</given-names>
            ,
            <surname>Fetzer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.H.</given-names>
            ,
            <surname>Rankin</surname>
          </string-name>
          ,
          <string-name>
            <surname>T.L</surname>
          </string-name>
          . (eds) Philosophy, Language, and Artificial Intelligence.
          <source>Studies in Cognitive Systems</source>
          , vol
          <volume>2</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>Walid</given-names>
            <surname>Saba</surname>
          </string-name>
          .
          <year>2020</year>
          .
          <article-title>Language, Knowledge, and Ontology: Where Formal Semantics Went Wrong, and How to Go Forward, Again</article-title>
          ,
          <source>Journal of Knowledge Structures and Systems</source>
          ,
          <volume>1</volume>
          (
          <issue>1</issue>
          ):
          <fpage>40</fpage>
          -
          <lpage>62</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>Walid</given-names>
            <surname>Saba</surname>
          </string-name>
          .
          <year>2007</year>
          .
          <article-title>Language, logic and ontology: Uncovering the structure of commonsense knowledge</article-title>
          ,
          <source>International Journal of Human Computer Studies</source>
          ,
          <volume>65</volume>
          (
          <issue>7</issue>
          ):
          <fpage>610</fpage>
          -
          <lpage>623</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>Wesley</given-names>
            <surname>Salmon</surname>
          </string-name>
          .
          <year>1989</year>
          .
          <article-title>Four decades of scientific explanation</article-title>
          , in: P. KITCHER &amp; W. SALMON (Eds)
          <article-title>Minnesota Studies in the Philosophy of Science</article-title>
          , Vol. XIII (Minnesota, University of Minnesota Press), pp.
          <fpage>3</fpage>
          -
          <lpage>21</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>Barry</given-names>
            <surname>Smith</surname>
          </string-name>
          ,
          <year>2005</year>
          . Against Fantology, In J. Marek and
          <string-name>
            <surname>E. M.</surname>
          </string-name>
          Reicher (eds.),
          <source>Experience and Analysis</source>
          , pp.
          <fpage>153</fpage>
          -
          <lpage>170</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>Fred</given-names>
            <surname>Sommers</surname>
          </string-name>
          .
          <year>1963</year>
          .
          <article-title>Types and ontology</article-title>
          .
          <source>Philosophical Review</source>
          ,
          <volume>72</volume>
          (
          <issue>3</issue>
          ), pp.
          <fpage>327</fpage>
          -
          <lpage>363</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <given-names>John</given-names>
            <surname>Sowa</surname>
          </string-name>
          .
          <year>1995</year>
          .
          <article-title>Knowledge Representation: Logical, Philosophical and Computational Foundations</article-title>
          , PWS Publishing Company, Boston.
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [24]
          <string-name>
            <surname>Karthik</surname>
            <given-names>Valmeekam</given-names>
          </string-name>
          , Matthew Marquez, Sarath Sreedharan,
          <string-name>
            <given-names>Subbarao</given-names>
            <surname>Kambhampati</surname>
          </string-name>
          .
          <year>2023</year>
          .
          <article-title>On the Planning Abilities of Large Language Models - A Critical Investigation</article-title>
          ,
          <source>In Advances in Neural Information Processing Systems</source>
          <volume>36</volume>
          (NeurIPS
          <year>2023</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          [25]
          <string-name>
            <surname>Vaswani</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Shazeer</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          , et. al.
          <year>2017</year>
          .
          <article-title>Attention is All You Need</article-title>
          , Available online at https://arxiv.org/abs/1706.03762.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>