<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Towards Explainable and Ontologically Grounded Language Models ⋆</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author role="corresp">
							<persName><forename type="first">Walid</forename><forename type="middle">S</forename><surname>Saba</surname></persName>
							<email>saba@northeastern.edu</email>
							<affiliation key="aff0">
								<orgName type="department">Institute for Experiential AI</orgName>
								<orgName type="institution">Northeastern University</orgName>
								<address>
									<addrLine>100 Fore St</addrLine>
									<postCode>04101</postCode>
									<settlement>Portland</settlement>
									<region>ME</region>
									<country key="US">USA</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Towards Explainable and Ontologically Grounded Language Models ⋆</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">F9486B1ABE702AE4B0675EC4C6AFAE44</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T18:57+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Large language models</term>
					<term>ontology</term>
					<term>bottom-up reverse engineering 1</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>We argue that the relative success of large language models (LLMs) is not a reflection on the symbolic vs. subsymbolic debate but a reflection on employing an appropriate bottom-up strategy of a reverse engineering of language at scale. However, and due to their subsymbolic nature whatever knowledge these systems acquire about language will always be buried in millions of weights none of which is meaningful on its own, rendering such systems utterly unexplainable. Furthermore, and due to their stochastic nature, LLMs will often fail in making the correct inferences in various linguistic contexts that require reasoning in intensional, temporal, or modal contexts. To remedy these shortcomings we suggest employing the successful bottomup strategy employed in LLMs but in a symbolic setting, resulting in explainable, languageagnostic, and ontologically grounded language models.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>To arrive at a scientific explanation there are generally two approaches we can adopt, a top-down approach or a bottom-up approach <ref type="bibr" target="#b19">(Salmon, 1989)</ref>. However, for a top-down approach to work, there must be a set of established general principles that one can start with, which is clearly not the case when it comes to language and how our minds externalize our thoughts in language. In retrospect, therefore, it is not surprising that decades of top-down work in natural language processing (NLP) failed to produce satisfactory results since most of this work was inspired by theories that made questionable assumptions where, for example, an innate universal grammar was assumed <ref type="bibr" target="#b3">(Chomsky, 1957)</ref>, or that we metaphorically build our linguistic competence based on a set of idealized cognitive models <ref type="bibr" target="#b10">(Lakoff, 1987)</ref>, or that natural language could be formally described using the tools of formal logic <ref type="bibr" target="#b16">(Montague, 1973)</ref>. In a similar vein, it is perhaps for the same reason that decades of top-down work in ontology and knowledge representation <ref type="bibr" target="#b11">(Lenat and</ref><ref type="bibr" target="#b11">Guha, 1990 and</ref><ref type="bibr" target="#b22">Sowa, 1995)</ref> also faltered since most of this work amounted to pushing, in a top-down manner, metaphysical theories of how the world is supposedly structured and represented in our minds, and again without any agreed upon general principles to start with. On the other hand, unprecedented progress has been made in only a few years of NLP work that employed a datadriven bottom-up strategy, as exemplified by recent advances in large language models (LLMs) that are essentially a massive experiment of a bottom-up reverse engineering of language at scale (e.g., ChatGPT and GPT-4)<ref type="foot" target="#foot_0">2</ref> .</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.1.">Issues with LLMs</head><p>Despite their relative success, LLMs do not tell us anything about how language works since these models are not really models of language but are statistical models of regularities found in language <ref type="foot" target="#foot_1">3</ref> . In fact, and due to their subsymbolic nature, whatever 'knowledge' these models acquire about language will always be buried in millions of weights (microfeatures) none of which is meaningful on its own, rendering these models utterly unexplainable <ref type="bibr" target="#b7">(Guizzardia and Guarino, 2024)</ref>. Besides unexplainability, LLMs are also oblivious to truth <ref type="bibr" target="#b2">(Borji, 2023)</ref>, since for LLMs all text (factual or nonfactual), is treated equally. Finally, and while LLMs have been shown to do poorly in a number of tasks that require high-level reasoning such as planning <ref type="bibr" target="#b23">(Valmeekam et. al., 2023)</ref>, analogies <ref type="bibr" target="#b12">(Lewis and Mitchell, 2024)</ref> and formal reasoning (Arkoudas, 2023) what concerns here is the failure of LLMs in making the right inferences in various linguistic contexts. As an illustration of the kinds of failures in deep language understanding we consider here three linguistic contexts involving copredication, intension and prepositional attitudes. The first example involves a phenomenon called copredication (see <ref type="bibr" target="#b0">Asher and Pustejovsky, 2005)</ref> which occurs when the same entity is used in the same context to refer to more than one semantic (ontological) type. All LLMs tested <ref type="foot" target="#foot_2">4</ref> failed in recognizing that 'newspaper' in the text is used to simultaneously refer to three entities: (i) the physical object I threw away; (ii) the content of the newspaper I was reading; and (iii) the 'editorial board' of the newspaper that did the firing of the columnist. Note that the failure of the LLMs was more acute when the LLMs were asked to draw a graph showing all entities and relations implied by the text since to show all the relations in the text all the different types of entities must be extracted. Here all LLMs tested showed the same newspaper (physical) object doing the firing of the columnist.</p><p>In example 2 all LLMs we tested approved replacing 'the capital of Spain' by 'Madrid' resulting in 'Maria thinks that Madrid was not always Madrid'. It is worth noting that the LLMs tested were consistently oblivion to intension. For example, in 'Perhaps Socrates was not the tutor of Alexander the Great', 'Socrates' and 'the tutor of Alexander the Great' were also deemed replaceable (since they are extensionally equal) resulting in 'Perhaps Socrates was not Socrates'. These results were expected since neural networks (deep or otherwise), that are the computing architecture behind all LLMs, are purely extensional models and are based on the 'empiricist theory of abstraction' where their similarity semantics has no notion of 'object identity' <ref type="bibr" target="#b13">(Lopes, 2023)</ref>.</p><p>Finally, example 3 illustrates failures of LLMs in making the correct inferences in modal (belief) contexts: the response of the LLMs tested was that 'Devon knows that Olga is a student' which is clearly the wrong inference since inferring K(Devon, student(Olga)) from K(Devon, client(Olga)student(Olga)) requires K(Devon, client(Olga)), i.e., it requires Devon knowing that Olga is a client. We have collected many other tests that we make available elsewhere for the sake of saving space.<ref type="foot" target="#foot_3">5</ref> </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.2.">LLMs: A Glass Half Empty, Half Full</head><p>So where do we stand now? On one hand, LLMs have clearly proven that one can get a handle on syntax and quite a bit of semantics in a bottom-up reverse engineering of language at scale; yet on the other hand what we have are unexplainable models that do not shed any light on how language actually works. Moreover, it would seem that due to their purely extensional and statistical nature, LLMs will always fail in making the correct inferences in many linguistic contexts. Since we believe the relative success of LLMs is not a reflection on the symbolic vs. subsymbolic debate but is a reflection on a successful bottom-up reverse engineering strategy, we think that combining the advantages of symbolic and ontologically grounded representations with a bottom-up reverse engineering strategy is a worthwhile effort. In fact, the idea that word meaning can be extracted from how words are actually used in language is not exclusive to linguistic work in the empirical tradition, but in fact it can be traced back to Frege.</p><p>In the rest of the paper we will (i) first argue that current word embeddings that are the genesis of modern-day large language models can be constructed in a symbolic setting instead of being the result of statistical cooccurrences; (ii) we will show that symbolic vectors perform better than current embeddings on a well-known word similarity benchmark; (iii) we will discuss how our symbolic vectors can be used to discover the ontological structure that is implicit in our ordinary language.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Concerning 'the Company a Word Keeps'</head><p>The genesis of modern LLMs is the distributional semantics hypothesis which states that the more semantically similar words are, the more they tend to occur in similar contexts -or, similarity in meaning is similarity in linguistic distribution <ref type="bibr" target="#b8">(Harris, 1954)</ref>. This is usually summarized by a saying that is attributed to the British linguist John R. Firth that "you shall know a word by the company it keeps". When processing a large corpus, this idea can be used by analyzing co-occurrences and contexts of use to approximate word meanings by word embeddings (vectors or tensors), that are essentially points in multidimensional space. Thus, at the root of LLMs is a bottom-up reverse engineering of language strategy where, unlike top-down approaches, "reverse engineers the process and induces semantic representations from contexts of use" <ref type="bibr" target="#b1">(Boleda, 2020)</ref>. But nothing precludes this idea from being carried out in a symbolic setting. In other words, the 'company a word keeps' can be measured in several ways, other than the correlational and statistical measures that underlie modern word embeddings.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1.">Symbolic Dimensions of Meaning</head><p>In discussing possible models of the world that can be employed in computational linguistics <ref type="bibr" target="#b9">Hobbs (1985)</ref> once suggested that there are two alternatives: (i) on one extreme we could attempt building a "correct" theory that would entail a full description of the world, something that would involve quantum physics and all the sciences; (ii) on the other hand, we could have a promiscuous model of the world that is isomorphic to the way we talk it about in natural language (emphasis is ours). Since the first option is a project that is most likely impossible to complete, what Hobbs is clearly suggesting here is a reverse engineering of language to discover how we actually use language to talk about the world we live in. This is also not much different from Frege's Context Principal that suggests "never ask for the meaning of words in isolation" <ref type="bibr" target="#b4">(Dummett, 1981)</ref> but that a word gets its meanings from analyzing all the contexts in which the word can appear <ref type="bibr" target="#b15">(Milne, 1986)</ref>. Again, what this suggests is that the meaning of words is embedded (to use a modern terminology) in all the ways we use these words in how we talk about the world. While Hobbs' and Frege's observations might be a bit vague, the proposal put forth by Fred Sommers (1963) was very specific. Again, Sommers suggests that "to know the meaning of a word is to know how to formulate some sentences containing the word" and this would lead, like in Frege's case, to the conclusion that a complete knowledge of some word w would be all the ways w can be used. For Sommers, the process of understanding the meaning of some word w starts by analyzing all the properties P that can sensibly be said of w. Thus, for example, [delicious Thursday] is not sensible while [delicious apple] is, regardless of the truth or falsity of the predication. Moreover, and since [delicious cake] is also sensible, then there must be a common type (perhaps food?) that subsumes both apple and cake. This idea is similar to the idea of type checking in strongly typed polymorphic programming languages. For example, the types in an expression such as 'x + 3' will only unify (or the expression will only 'make sense') if/when x is an object of type number (as opposed to a tuple, for example). As it was suggested in <ref type="bibr" target="#b18">(Saba, 2007)</ref>, this type of analysis can thus be used to 'discover' the ontology that seems to be implicit in the language, as will be discussed below. First, however, we describe how a bottom-up reverse engineering of language can be done in a symbolic setting.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2.">Symbolic Reverse Engineering of Language</head><p>The procedure we have in mind assumes a Platonic universe where all concepts, physical or abstract, including states, activities, properties (tropes) <ref type="bibr" target="#b14">(Moltmann, 2013)</ref>, processes, events, etc. are considered entities that can be defined by a number of language-agnostic primitives <ref type="bibr" target="#b20">(Smith, 2005)</ref> that we call the 'dimensions of meaning'. We consider here the following dimensions: AGENTOF, OBJECTOF, HASPROP, INSTATE, PARTOF, INSTATE, INPROCESS, and OFTYPE. For every word w in the language, and for every dimension D, a reverse-engineering process is conducted to compute a set w D = {(x, t) | D(w, x)} where t is a weight in [0,1]. Here are example sets computed for 'book' along four dimensions of meaning along with the masking prompt that queries what an LLM has 'learned' about how we talk about books:</p><p>book . HASPROP Everyone likes to read a [MASK] book. =&gt; {(popular, 0.9), (educational, 0.8), (famous, 0.8), ... } book . OBJECTOF Everyone I know enjoyed [MASK] 'The Prince'. =&gt; {(reading, 0.9), (writing, 0.8), (editing, 0.8), ... } book . AGENTOF Das Kapital has [MASK] many people over the years. =&gt; {(influenced, 0.9), (inspired, 0.8), (changed, 0.8), ... } book . PARTOF Hamlet should be part of every <ref type="bibr">[MASK]</ref>. =&gt; {(collection, 0.9), (archive, 0.8), (library, 0.8), ... } <ref type="bibr">[MASK]</ref>. =&gt; {(print, 0.9), (circulation, 0.8), (review, 0.8), ... } What the above says is the following (i) in ordinary spoken language we speak of a 'book' that is popular, educational, famous, etc.; (ii) we speak of reading, writing, editing, etc. a 'book'; (iii) we speak of 'book' that may change, influence, inspire, etc.; and (iv) we speak of a b 'book' that is part of a collection, an archive, or a library; and (v) a book can be in review, in print, in circulation, etc. The nominalization process can be conducted using the copular 'is' as shown in table <ref type="table" target="#tab_0">1</ref>. For example, 'John is famous' can be restated as 'John has the property of fame'; 'Jim is sad' as 'Jim is in a state of sadness'; etc. (see <ref type="bibr" target="#b20">[Smith, 2005]</ref> for more on the relationship between the copular and abstract entities and <ref type="bibr" target="#b14">[Moltmann, 2013]</ref> for more on abstract objects.) What should be noted here is that even with the simple conceptual structure discovered thus far one can generate plausible text, such as the following:</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>book . INSTATE I was told that my book is now in</head><p>(1) enjoyed the interesting reading of the new book (2) completed a boring reading of a controversial book</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Table 1: From propositions to relations and entities</head><p>The sensible (and meaningful) fragment in (1) can be generated because a book can be 'read' and described by 'new', and readings can be 'interesting' and the object of enjoyment; and similarly for (2) where a reading of a controversial book can be boring and the object of a completion, etc. Note, however, that text generation in this case is not a function of 'predicting' the most likely continuation, but a function of plausible filling in of subjects, objects, agents, descriptions, etc. to any propositional structure.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.3.">Symbolic Embeddings</head><p>The process we described thus far results in symbolic word embeddings as the one shown in figure <ref type="figure" target="#fig_1">2</ref> below. In figure <ref type="figure" target="#fig_1">2</ref>(a) we show the symbolic embedding for 'boy' and 'lad' along the HASPROP dimension. Thus, in ordinary spoken language it is sensible to speak of a 'handsome boy' and a 'funny boy' as well as a 'clever lad' and a 'talented lad'. We note here that in this process generic descriptions are removed using a function that computes the information content of some adjectives, where the information content of an adjective adj is inversely proportional to the set of types of adj can sensibly be applied to. For example, 'beautiful' will have a low information content score since 'beautiful' can sensibly be said of many concepts, both physical and abstract (e.g., car, movie, poem, night, girl, …) while 'tasty' can sensibly be said of 'food' and just a few others. The symbolic embeddings in figure 2(b) are those of 'automobile' and 'car' along the OBJECTOF dimension. Note now that word similarity along these symbolic dimensions can be computed using cosine similarity as well as weighted Jaccard similarity where max and min can be used in fuzzy union and fuzzy intersection. We are currently experimenting with the optimal number of dimensions using a number of word similarity benchmarks, including the WordSim353 dataset <ref type="bibr" target="#b5">(Finkelstein, Lev. et al., 2001)</ref>  <ref type="foot" target="#foot_4">6</ref> .</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Figure 1:</head><p>We speak of a 'book' (i) that influence, change, convince; (ii) that is edited, read, written; (iii) that can be popular, controversial, famous; (iv) that is part of a library, an archive, etc.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">The Ontology of the Language of Thought?</head><p>The reverse engineering process we have described above would result in symbolic embeddings along various dimensions, as the ones shown in figure <ref type="figure" target="#fig_1">2</ref>. As a result of this, however, we could then analyze the subset relations between these embeddings to discover the ontological structure that seems to be implicit in our ordinary language. To illustrate, consider the following:</p><p>(3) car . objectOf = {(driving, 0.9), (repairing, 0.8), (buying, 0.8), ... } (4) book . objectOf = {(reading, 0.9), (writing, 0.8), (buying, 0.8), ... } (5) person . AGENTOF = {(reading, 0.9),(writing, 0.8), (driving, 0.8), ... } (6) person. HASPROP = {(popularity, 0.9), (fame, 0.8), (beautiful, 0.8), .. } (7) car. hasProp = {(popularity, 0.9), (fame, 0.8), (beautiful, 0.8), .. } (8) book . HASPROP = {(popularity, 0.9), (fame, 0.8), (beautiful, 0.8), .. } Note that car can be the object of 'buying' and so can be a book and this means that car and book must, at some level of abstraction, share the same parent (perhaps 'artifact'?) Note also that a car as well as a book and a person can be popular. An analysis along these lines would result in the following: What the above says is the following: in ordinary spoken language we speak of people reading and writing books (9 and 10); we speak of people buying cars and books, and thus of buying objects that are of some type that subsumes both cars and books (11); we speak of people driving cars (12); and we speak of beautiful people, cars, and books (and thus beautiful seems to be a property that can sensibly be said of concepts that are at very high level of generality). As suggested by <ref type="bibr" target="#b21">Sommers (1963)</ref> this type of analysis that can be fully automated with the help of LLMs can help us discover what he called 'the Tree of Language' -which is essentially the ontology that seems to be underneath our ordinary language. This might also be what <ref type="bibr" target="#b9">Hobbs (1985)</ref> was seeking when he suggested building a model of the world that isomorphic to the we talk about it in natural language.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Concluding Remarks</head><p>Large language models (LLMs) have shown impressive capabilities that pioneers in artificial intelligence and natural language processing would marvel at. However, we believe that LLMs are not the answer to the language understanding problem nor to reasoning in general and in particular commonsense reasoning. Due to their paradigmatic unexplainability LLMs will also not shed any light on how language works and how we externalize our thoughts in language. Since, in our opinion, the relative success of LLMs is not due to their subsymbolic nature but due to applying a successful bottom-up reverse engineering strategy, we suggested here applying the same strategy but in a symbolic setting, something that has been argued for by logicians dating back to Frege. By combining the successful bottom-up strategy and symbolic and ontological methods we arrive at explainable and ontologically grounded language models that can be used in problems requiring commonsense reasoning. We are still in the early stage of this work, but we currently have the tools to realize the dream of Frege and Sommers and perhaps shed some light on the 'language of thought' Fodor (1998) -the internal language that we use to construct and process our thoughts.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>( 9 )</head><label>9</label><figDesc>read(person, book) (10) write(person, book) (11) buy(person, T1 = car  book … ) (12) drive(person, car) (13) beautiful(T2 = person  car  book … )</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figure 2 :</head><label>2</label><figDesc>Figure 2: (a) the symbolic embeddings of 'boy' and 'lad' along the HASPROP dimension (with a weighted Jaccard similarity of 0.876) and (b) those of 'automobile' and 'car' along the OBJECTOF dimension (the weighted Jaccard similarity is 0.91)</figDesc><graphic coords="5,311.95,309.40,197.33,198.10" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Example 1 .</head><label>1</label><figDesc>Show the entities and the relations that are implicit in the following text: "I threw away the newspaper I was reading because they fired my favorite columnist". Since Madrid is the capital of Spain, can I replace one for the other in the following: "Maria thinks Madrid was not always the capital of Spain"?</figDesc><table><row><cell>Example 2. Example 3. Suppose Devon knows that if someone is a</cell></row><row><cell>client, then s/he is a student, and suppose that Olga is a</cell></row><row><cell>client. Then what does Devon know?</cell></row></table></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_0">GPT stands for 'Generative Pre-trained Transformer', an architecture that OpenAI built on top of the transformer architecture<ref type="bibr" target="#b24">(Vaswani, A. et. al., 2017)</ref>.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_1">In looking inside the neural network (NN) of an LLM one does not find concepts, meanings, linguistic structures, etc. but weights associated with neural connections, which is exactly what one will find in an object recognition or any other NN.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="4" xml:id="foot_2">Our experiments were conducted on GPT-4o (chat.openai.com).</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="5" xml:id="foot_3">https://shorturl.at/ejmH8</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="6" xml:id="foot_4">https://kaggle.com/datasets/julianschelb/wordsim353-crowd</note>
		</body>
		<back>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Word Meaning and Commonsense Metaphysics</title>
		<author>
			<persName><forename type="first">Nicholas</forename><surname>Asher</surname></persName>
		</author>
		<author>
			<persName><forename type="first">James</forename><surname>Pustejovsky</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Course Materials for Type Selection and the Semantics of Local Context</title>
				<meeting><address><addrLine>ESSLLI</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2005">2005. 2005</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Distributional Semantics and Linguistic Theory</title>
		<author>
			<persName><forename type="first">Gemma</forename><surname>Boleda</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Annual Review of Linguistics</title>
		<imprint>
			<biblScope unit="volume">6</biblScope>
			<biblScope unit="page" from="213" to="234" />
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<monogr>
		<title level="m" type="main">A Categorical Archive of ChatGPT Failures</title>
		<author>
			<persName><forename type="first">A</forename><surname>Borji</surname></persName>
		</author>
		<ptr target="https://arxiv.org/abs/2302.03494" />
		<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<monogr>
		<title level="m" type="main">Syntactic Structures</title>
		<author>
			<persName><forename type="first">Noam</forename><surname>Chomsky</surname></persName>
		</author>
		<imprint>
			<date type="published" when="1957">1957</date>
			<publisher>Mouton de Gruyter</publisher>
			<pubPlace>NY</pubPlace>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<monogr>
		<title level="m" type="main">Frege: Philosophy of Language</title>
		<author>
			<persName><forename type="first">Michael</forename><surname>Dummett</surname></persName>
		</author>
		<imprint>
			<date type="published" when="1981">1981</date>
			<publisher>Harvard University Press</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Placing search in context: The concept revisited</title>
		<author>
			<persName><forename type="first">Lev</forename><surname>Finkelstein</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 10th international conference on the World Wide Web</title>
				<meeting>the 10th international conference on the World Wide Web</meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2001">2001</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<monogr>
		<author>
			<persName><forename type="first">Jerry</forename><surname>Fodor</surname></persName>
		</author>
		<title level="m">Concepts: Where Cognitive Science Went Wrong</title>
				<imprint>
			<publisher>Oxford University Press</publisher>
			<date type="published" when="1988">1988</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<monogr>
		<author>
			<persName><forename type="first">Giancarlo</forename><surname>Guizzardia</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Nicola</forename><surname>Guarino</surname></persName>
		</author>
		<title level="m">Semantics, Ontology, and Explanation, Data &amp; Knowledge Engineering</title>
				<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
	<note>to appear</note>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Distributional Structure</title>
		<author>
			<persName><forename type="first">S</forename><surname>Zellig</surname></persName>
		</author>
		<author>
			<persName><surname>Harris</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Word</title>
		<imprint>
			<biblScope unit="volume">10</biblScope>
			<biblScope unit="page" from="146" to="162" />
			<date type="published" when="1954">1954</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Ontological promiscuity</title>
		<author>
			<persName><forename type="first">Jerry</forename><surname>Hobbs</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings. of the 23rd Annual Meeting of the Association for Computational Linguistics</title>
				<meeting>of the 23rd Annual Meeting of the Association for Computational Linguistics<address><addrLine>Chicago, Illinois</addrLine></address></meeting>
		<imprint>
			<date type="published" when="1985">1985. 1985</date>
			<biblScope unit="page" from="61" to="69" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<monogr>
		<author>
			<persName><forename type="first">George</forename><surname>Lakoff</surname></persName>
		</author>
		<title level="m">Women, Fire, and Dangerous Things: What Categories Reveal About the Mind</title>
				<imprint>
			<publisher>University of Chicago Press</publisher>
			<date type="published" when="1987">1987</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<monogr>
		<title level="m" type="main">Building Large Knowledge-Based Systems: Representation and Inference in the CYC Project</title>
		<author>
			<persName><forename type="first">Doug</forename><surname>Lenat</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">V</forename><surname>Guha</surname></persName>
		</author>
		<imprint>
			<date type="published" when="1990">1990</date>
			<publisher>Addison-Wesley</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<monogr>
		<title level="m" type="main">Using Counterfactual Tasks to Evaluate the Generality of Analogical Reasoning in LLMs</title>
		<author>
			<persName><forename type="first">Martha</forename><surname>Lewis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Melanie</forename><surname>Mitchell</surname></persName>
		</author>
		<ptr target="https://arxiv.org/abs/2402.08955" />
		<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">Can Deep CNNs Avoid Infinite Regress/Circularity in Content Constitution?</title>
		<author>
			<persName><forename type="first">Jesse</forename><surname>Lopes</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Minds and Machines</title>
		<imprint>
			<biblScope unit="volume">33</biblScope>
			<biblScope unit="page" from="507" to="524" />
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<monogr>
		<title level="m" type="main">Abstract Objects and the Semantics of Natural Language</title>
		<author>
			<persName><forename type="first">Friederike</forename><surname>Moltmann</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2013">2013</date>
			<publisher>Oxford University Press</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">Frege&apos;s Context Principle</title>
		<author>
			<persName><forename type="first">Peter</forename><surname>Milne</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Mind</title>
		<imprint>
			<biblScope unit="volume">95</biblScope>
			<biblScope unit="issue">380</biblScope>
			<biblScope unit="page" from="491" to="495" />
			<date type="published" when="1986">1986</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">The Proper Treatment of Quantification in Ordinary English</title>
		<author>
			<persName><forename type="first">Richard</forename><surname>Montague</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Philosophy, Language, and Artificial Intelligence. Studies in Cognitive Systems</title>
				<editor>
			<persName><forename type="first">J</forename><surname>Kulas</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">J</forename><forename type="middle">H</forename><surname>Fetzer</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">T</forename><forename type="middle">L</forename><surname>Rankin</surname></persName>
		</editor>
		<imprint>
			<date type="published" when="1973">1973</date>
			<biblScope unit="volume">2</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">Language, Knowledge, and Ontology: Where Formal Semantics Went Wrong, and How to Go Forward, Again</title>
		<author>
			<persName><forename type="first">Walid</forename><surname>Saba</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Knowledge Structures and Systems</title>
		<imprint>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page" from="40" to="62" />
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<analytic>
		<title level="a" type="main">Language, logic and ontology: Uncovering the structure of commonsense knowledge</title>
		<author>
			<persName><forename type="first">Walid</forename><surname>Saba</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">International Journal of Human Computer Studies</title>
		<imprint>
			<biblScope unit="volume">65</biblScope>
			<biblScope unit="issue">7</biblScope>
			<biblScope unit="page" from="610" to="623" />
			<date type="published" when="2007">2007</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<analytic>
		<title level="a" type="main">Four decades of scientific explanation</title>
		<author>
			<persName><forename type="first">Wesley</forename><surname>Salmon</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="s">Minnesota Studies in the Philosophy of Science</title>
		<editor>P. KITCHER &amp; W. SALMON</editor>
		<imprint>
			<biblScope unit="volume">XIII</biblScope>
			<biblScope unit="page" from="3" to="21" />
			<date type="published" when="1989">1989</date>
			<publisher>University of Minnesota Press</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<analytic>
		<title level="a" type="main">Against Fantology</title>
		<author>
			<persName><forename type="first">Barry</forename><surname>Smith</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Experience and Analysis</title>
				<editor>
			<persName><forename type="first">J</forename><surname>Marek</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">E</forename><forename type="middle">M</forename><surname>Reicher</surname></persName>
		</editor>
		<imprint>
			<date type="published" when="2005">2005</date>
			<biblScope unit="page" from="153" to="170" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b21">
	<analytic>
		<title level="a" type="main">Types and ontology</title>
		<author>
			<persName><forename type="first">Fred</forename><surname>Sommers</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Philosophical Review</title>
		<imprint>
			<biblScope unit="volume">72</biblScope>
			<biblScope unit="issue">3</biblScope>
			<biblScope unit="page" from="327" to="363" />
			<date type="published" when="1963">1963</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b22">
	<analytic>
		<title level="a" type="main">Knowledge Representation: Logical</title>
		<author>
			<persName><forename type="first">John</forename><surname>Sowa</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Philosophical and Computational Foundations</title>
				<meeting><address><addrLine>Boston</addrLine></address></meeting>
		<imprint>
			<publisher>PWS Publishing Company</publisher>
			<date type="published" when="1995">1995</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b23">
	<analytic>
		<title level="a" type="main">On the Planning Abilities of Large Language Models -A Critical Investigation</title>
		<author>
			<persName><forename type="first">Karthik</forename><surname>Valmeekam</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Matthew</forename><surname>Marquez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Sarath</forename><surname>Sreedharan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Subbarao</forename><surname>Kambhampati</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Advances in Neural Information Processing Systems</title>
		<imprint>
			<biblScope unit="volume">36</biblScope>
			<date type="published" when="2023">2023. 2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b24">
	<monogr>
		<title level="m" type="main">Attention is All You Need</title>
		<author>
			<persName><forename type="first">A</forename><surname>Vaswani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Shazeer</surname></persName>
		</author>
		<ptr target="https://arxiv.org/abs/1706.03762" />
		<imprint>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
