<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">SCIENCE IS EXPLORATION: Computational Frontiers for Conceptual Metaphor Theory</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Rebecca</forename><forename type="middle">M M</forename><surname>Hicke</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Department of Computer Science</orgName>
								<orgName type="institution">Cornell University</orgName>
								<address>
									<country key="US">USA</country>
								</address>
							</affiliation>
							<affiliation key="aff1">
								<orgName type="department">Center for Humanities Computing</orgName>
								<orgName type="institution">Aarhus University</orgName>
								<address>
									<country key="DK">Denmark</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Ross</forename><surname>Deans Kristensen-Mclachlan</surname></persName>
							<affiliation key="aff1">
								<orgName type="department">Center for Humanities Computing</orgName>
								<orgName type="institution">Aarhus University</orgName>
								<address>
									<country key="DK">Denmark</country>
								</address>
							</affiliation>
							<affiliation key="aff2">
								<orgName type="department">Department of Linguistics, Cognitive Science, and Semiotics</orgName>
								<orgName type="institution">Aarhus University</orgName>
								<address>
									<country key="DK">Denmark</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">R</forename><forename type="middle">D</forename><surname>Kristensen-Mclachlan</surname></persName>
						</author>
						<title level="a" type="main">SCIENCE IS EXPLORATION: Computational Frontiers for Conceptual Metaphor Theory</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">F8C3B8CA5902A00C002657C12CB45ADF</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T19:48+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>conceptual metaphor theory</term>
					<term>large language models</term>
					<term>pragglejaz</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Metaphors are everywhere. They appear extensively across all domains of natural language, from the most sophisticated poetry to seemingly dry academic prose. A significant body of research in the cognitive science of language argues for the existence of conceptual metaphors, the systematic structuring of one domain of experience in the language of another. Conceptual metaphors are not simply rhetorical flourishes but are crucial evidence of the role of analogical reasoning in human cognition. In this paper, we ask whether Large Language Models (LLMs) can accurately identify and explain the presence of such conceptual metaphors in natural language data. Using a novel prompting technique based on metaphor annotation guidelines, we demonstrate that LLMs are a promising tool for large-scale computational research on conceptual metaphors. Further, we show that LLMs are able to apply procedural guidelines designed for human annotators, displaying a surprising depth of linguistic knowledge.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>Metaphor is commonly understood as the description of one concept in the vocabulary of another, typically for poetic, rhetorical, or otherwise literary effect. However, some metaphors are so systematic and conventionalized that we barely recognize them as such. Consider, for example, the commonplace metaphor life is a journey. While this is certainly a metaphor in its own right, it can also be instantiated in ways which do not directly refer to either life or journeys: "We have come a long way"; "I'm going through a rough patch"; "She's at a crossroads. " A metaphor like life is a journey is more than simple figurative language. Instead, it allows us to make sense of the process of a normal human life through the rich language of journeys and voyages. These kinds of systematic mappings are known as conceptual metaphors <ref type="bibr" target="#b15">[16]</ref>.</p><p>Conceptual metaphors are so deeply ingrained in our everyday language that the most frequent usage of a word or phrase may be metaphorical. This can make it difÏcult to recognize the presence of metaphorical language, even for linguists and other specialists <ref type="bibr" target="#b13">[14]</ref>. Moreover, processing and understanding conceptual metaphors necessarily involves encyclopedic general knowledge of the world, as well as shared cultural norms and stereotypes between speaker/hearer or writer/reader. Identifying conceptual metaphors has therefore consistently proven to be a challenge for both natural language processing (NLP) and artificial intelligence (AI). Given that contemporary transformer-based models 'learn' by consuming huge amounts of natural language and build knowledge by looking at how words are commonly used, it seems that they should similarly struggle to recognize these metaphors.</p><p>Nevertheless, being able to consistently identify the presence of conceptual metaphors in discourse would be greatly beneficial for fields such as computational literary studies and cultural analytics more broadly. In this paper, we therefore set out to test empirically whether LLMs can be used productively in the context of conceptual metaphor theory (CMT). If the answer is no, it may be evidence that developing computational approaches to conceptual metaphor remains an intractable problem. If, on the other hand, LLMs can consistently identify the presence of conceptual metaphors, the question then arises of exactly how and where LLMs encode metaphoricity.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Related Work</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Cognitive Linguistic Foundations</head><p>The study of CMT was inaugurated in the 1980s in Lakoff and Johnson <ref type="bibr" target="#b15">[16]</ref>. Central to the argument in <ref type="bibr" target="#b15">[16]</ref> is the fundamental role that metaphor plays in structuring everyday discourse. Rather than simply being a literary or rhetorical flourish, Lakoff and Johnson argue that conceptual metaphors structure our common understanding of entities, processes, or entire semantic domains. CMT played a crucial role in the development of non-generativist cognitive linguistics <ref type="bibr" target="#b6">[7,</ref><ref type="bibr" target="#b16">17,</ref><ref type="bibr" target="#b17">18,</ref><ref type="bibr" target="#b26">27,</ref><ref type="bibr" target="#b27">28]</ref>. An important development came with the introduction of conceptual blending (otherwise known as conceptual integration), in which the kinds of analogical reasoning underlying conceptual metaphors are argued to be more fundamental to human language and cognition <ref type="bibr" target="#b10">[11,</ref><ref type="bibr" target="#b11">12,</ref><ref type="bibr" target="#b7">8]</ref>. While there are dissenting voices <ref type="bibr" target="#b12">[13]</ref>, recent decades have seen a growing body of empirical research which claims to provide experimental evidence in favour of CMT <ref type="bibr" target="#b3">[4,</ref><ref type="bibr" target="#b5">6,</ref><ref type="bibr" target="#b28">29]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Metaphor and Artificial Intelligence</head><p>The extensive role and function of metaphor in natural language has not escaped researchers in NLP. Indeed, a great deal of early work in AI concerned exactly how a computer could make sense of metaphorical language <ref type="bibr" target="#b4">[5,</ref><ref type="bibr" target="#b22">23,</ref><ref type="bibr" target="#b23">24]</ref> (see <ref type="bibr" target="#b1">[2]</ref> for a comprehensive overview). Much of this work relied on so-called symbolic AI, but struggled because GOFAI (good, old-fashioned AI) could not provide the analogous reasoning skills and extensive general knowledge required to process metaphorical language. In the context of conceptual blending specifically, there have been tentative attempts to model the underlying processes involved computationally <ref type="bibr" target="#b9">[10]</ref>.  However, the encyclopedic world knowledge required in CMT is challenging to encode since the domain is essentially unbounded. Moreover, a core issue at the heart of computational metaphor analysis has historically been the lack of a shared definition of what is being studied, along with the scarcity of robust data and evaluation strategies <ref type="bibr" target="#b25">[26]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Recent Developments</head><p>However, recent years have again seen a growing interest in quantitative and computational approaches to metaphor, primarily in more applied fields such as corpus linguistics and stylistics <ref type="bibr" target="#b8">[9,</ref><ref type="bibr" target="#b14">15,</ref><ref type="bibr" target="#b24">25]</ref>. With the advent of neural language models, research in NLP has increasingly attempted to discover if (and how) word embeddings capture aspects of metaphoricity <ref type="bibr" target="#b19">[20,</ref><ref type="bibr" target="#b29">30,</ref><ref type="bibr" target="#b20">21,</ref><ref type="bibr" target="#b21">22]</ref>. More recently, work has been done on explicitly probing pre-trained language models to inspect how metaphorical knowledge is encoded <ref type="bibr" target="#b0">[1]</ref> and creating shared task datasets <ref type="bibr" target="#b29">[30]</ref>. In the context of LLMs specifically, GPT-3 has been shown to perform reasonably well at predicting source domains for English metaphors, although it performs significantly worse for Spanish <ref type="bibr" target="#b30">[31]</ref>. These are promising developments and point to the increasing awareness of the practical and theoretical challenges underlying the nature of figurative language in NLP. To date, though, we are unaware of existing research which seeks to operationalize existing metaphor identification procedures for LLMs.</p><p>For each word in the following text, determine if it has a *more* basic contemporary meaning in other contexts than the one in the given context. For our purposes, basic meanings tend to be:</p><p>-more concrete; what they evoke is easier to imagine, see, here, feel, smell, and taste -related to bodily action -more precise (as opposed to vague) -historically older Basic meanings are not necessarily the most frequent meanings of the word. Remember to only report YES for words whose use in this context is not their most basic meaning.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Examples:</head><p>"I demolished his argument."</p><p>-I: NO (pronoun, basic meaning is the same) -demolished: YES (verb, more basic meaning refers to destroying a building) -his: NO (possessive pronoun, basic meaning is the same) -argument: NO (noun, basic meaning is the same)</p><p>"They write about their family."</p><p>-They: NO (pronoun, basic meaning is the same) -write: NO (verb, basic meaning is the same) -about: NO (preposition, basic meaning is the same) -their: NO (possessive pronoun, basic meaning is the same) -family: NO (noun, basic meaning is the same)</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Figure 2:</head><p>The modified prompt used in all experiments. All significant (non-formatting) changes from</p><p>Step 3b of the MIP procedure are highlighted in red. The sentence to be analyzed is appended in quotations to the bottom of the prompt when querying models.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Methods</head><p>We are interested in studying whether LLMs are capable of leveraging the contextual and cultural knowledge necessary to perform difÏcult metaphor identification tasks, such as recognizing conceptual metaphors. To do this, we operationalize the Metaphor Identification Procedure (MIP), a set of annotation guidelines introduced by the Pragglejaz Group in 2007 <ref type="bibr" target="#b13">[14]</ref>. The steps of MIP are detailed in Figure <ref type="figure" target="#fig_0">1</ref>. MIP and its variants are among the most commonly used procedures for metaphor annotation in work in corpus linguistics and stylistics; we therefore take it as a reasonable proxy for the metaphor identification process in humans.</p><p>Examining MIP, the first few steps have clear parallels in transformer-based LLMs.</p><p>Step 1 is performed by the attention mechanism. Creating the representation of each word, in traditional attention mechanisms, requires 'looking' at every other word in the text, thus allowing the model to "establish a general understanding of [its] meaning. " Step 2, determining the lexical units in a text, is directly paralleled by tokenization. Finally, Step 3a is achieved via contextual embeddings, whose explicit purpose is to represent the specific meaning of a word in context. However, Step 3b of MIP has no clear counterpart in transformer-based models. This is because the 'basic meaning' of a lexical unit, as defined by MIP, is not obviously stored or represented within these models.</p><p>We want to know if state-of-the-art LLMs are capable of replicating Step 3b of the MIP procedure despite these barriers, so we transform it into a prompt for instruction-tuned models (Figure <ref type="figure">2</ref>). In creating the prompt, we leave the text of Step 3b almost entirely intact. "Lexical units" are replaced by "words" in order to encourage consistent outputs. Emphasis is added in several locations to only annotating words which have a more basic meaning than their usage in the input text. Finally, two examples are appended to the prompt. These examples are meant to facilitate in-context learning and provide a structure for the models to mimic in their outputs. They instruct the model to provide a parenthetical expression next to each word consisting of the word's part of speech and its more basic meaning, if one exists. These expressions were introduced by the gpt-4o model itself during initial exploration and were then included in the prompt because they force the model to provide explanations for its annotations. This allows for further evaluation of the outputs and ultimately means they play a similar role to Chain of Thought prompts, which have been shown to improve model performance across a variety of tasks <ref type="bibr" target="#b31">[32]</ref>.</p><p>We use this prompt to query three models from OpenAI's GPT family, chosen because of their size, popularity, and high performance on a wide range of tasks. Specifically, we look at gpt-3.5-turbo, 1 gpt-4-turbo, 2 and gpt-4o. <ref type="foot" target="#foot_2">3</ref> We access each model using OpenAI's Chat Completion API. A system prompt ("You are a helpful assistant. You have extensive linguistic knowledge. ") is provided each time the model is queried. Then, the main prompt (Figure <ref type="figure">2</ref>) is passed as a user message, with the current text of interest appended in quotations at the end. All parameters except nuclear sampling (top_p) are left at their default values during prompting for all models. We set top_p to 0.1, which means that the models only select from the tokens which make up the top 10% of the probability mass when choosing each next token. We adjust the top_p value because we are not prompting for diverse or creative text, but instead for reliable and accurate outputs, which we hypothesize will be more likely produced from highprobability tokens. The cost of all experiments using the OpenAI API was ∼$100.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Data</head><p>We evaluate each model on two datasets. First, we wish to study the models' basic ability to use MIP to correctly identify metaphorical words. To this end, we use the Trope Finder (TroFi) dataset <ref type="bibr" target="#b2">[3]</ref>, which consists of sentences drawn from Wall Street Journal issues published between 1987 and 1989. Each sentence contains one of a list of fifty words whose usage is annotated as literal or non-literal. We only consider sentences from the dataset that have been labeled by humans and take their annotations as ground truth, excluding all sentences labeled only by the TroFi clustering system. This leaves us with 3,736 sentences. Each model is then prompted as described above for each sentence and evaluated only on the words annotated in TroFi. A response is marked as correct if the model says "YES" there is a more basic meaning for the word of interest and the label from TroFi is "literal" and vice versa. Occasionally, the models provide labels for a phrase in the input sentence instead of individual words; the label provided for the entire phrase is then applied to the word of interest if necessary. If a model</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Table 1</head><p>The precision and recall values for the identification of literal and metaphorical word usage in the TroFi dataset by all models.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Metaphorical Literal</head><p>Precision Recall Precision Recall does not provide a label for a word, we act as if the incorrect label was given. However, the TroFi dataset does not permit for a very fine-grained analysis of the models' capabilities. In order to facilitate such an analysis, we create a new dataset consisting of example sentences pulled from Lakoff and Johnson's original work on CMT <ref type="bibr" target="#b15">[16]</ref>. We collect all full sentence examples from the text that we believe can be understood as metaphorical without context, leaving us with 544 sentences, which we refer to as the MWLB dataset. The dataset is made publicly available, along with all model outputs and human annotations. <ref type="foot" target="#foot_3">4</ref>Each model is again prompted to provide word-level annotations for every sentence. Then, we perform detailed qualitative analysis on each model's output for a subset of 100 sentences. Specifically, we evaluate each output on five binary categories:</p><p>• L&amp;J Metaphor(s) -Identified: Whether the model has correctly identified all metaphors highlighted by Lakoff and Johnson. For metaphorical phrases, the model's response is considered correct if it has identified at at least one key word as having a more basic meaning. All annotations are performed by one of the authors, using MIP as a guideline. Confusing or difÏcult annotations are then discussed between both authors till a consensus is reached. These annotations are necessarily subjective, and we do not claim that they represent a ground truth. Instead, since these evaluations are informed by the same procedure used to create the outputs, we claim that they allow us to study the plausibility of the models' performance, if not strictly their accuracy. In future work, further evaluation of model outputs with multiple trained annotators may allow for a more concrete analysis of model performance.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Results and Discussion</head><p>On the TroFi dataset, all models demonstrate an ability to distinguish between literal and nonliteral word usage with generally better than chance performance (Table <ref type="table">1</ref>). Thus, they seem to be able to apply the MIP procedure for identifying metaphors and determine computationally whether a word has a more basic meaning. This is surprising since, as discussed above, basic word meanings are not necessarily the most frequent. Nothing about the language modeling approach to pre-training models guarantees that a model would be able to differentiate between the most frequent and most basic meaning of a word. We also find, however, that the models are prone to over-labeling words as metaphorical (Figure <ref type="figure" target="#fig_1">3</ref>). Whereas all three models achieve high recall for metaphor identification, they frequently label words used literally as having a more basic meaning, leading to low recall for the 'literal' category and low precision for the 'metaphorical' category. 3.5-turbo particularly struggles with accurately identifying literal word usage. It is worth noting that, although there are clear cases where the models have failed to properly apply MIP, there may also be a disconnect between annotations in the TroFi corpus and what is considered metaphorical by the MIP procedure, as MIP was not used for annotation in TroFi.</p><p>For the annotated subset of the MWLB dataset, all the models found all the Lakoff and Johnson metaphors in over 60% of sentences and further provided a correct basic meaning for the metaphorical words in over 50% of sentences (Figure <ref type="figure" target="#fig_2">4</ref>). 4o performed the best on this task, achieving a remarkable performance given the subtlety and complexity of many of the metaphors and the cultural knowledge required to identify both their metaphoricity and their basic meanings. 4o's high performance may be due in part to the fact that it was most likely to label words as metaphorical overall, having annotated additional words in the largest percent- age of sentences (Figure <ref type="figure" target="#fig_3">5</ref>). In only 48% of sentences did these additional annotations identify only plausibly metaphorical words. Nonetheless, the inclusion of correct basic meanings for such a large proportion of the Lakoff and Johnson metaphors suggests that MIP was overall being correctly applied and true 'knowledge' was being demonstrated, not just random chance.</p><p>3.5-turbo struggled the most at replicating the desired output structure, occasionally excluding words or truncating sentences, inconsistently formatting lists, and sometimes dropping the parenthetical explanations. This affected its accuracy, as the model sometimes failed to provide annotations for words central to the metaphors of interest. Additionally, 4-turbo  was least likely to annotate a word as having a more basic meaning (Figure <ref type="figure" target="#fig_3">5</ref>). This meant it identified fewer Lakoff and Johnson metaphors, but a greater proportion of the additional metaphors it annotated were plausible. The basic meanings provided by 4-turbo were usually of high quality, and tended to be more accurate and specific than those produced by 3.5-turbo. Despite the benefits, however, this model's 'caution' led it to perform worse than 4o overall.</p><p>Each model struggled to correctly identify when smaller function words were used metaphorically, particularly prepositions like 'in, ' . This made the models worse at identifying so-called container metaphors, such as life is a container and activities are containers. For example, in the sentence "That's in the center of my field of vision. ", labeled by Lakoff and Johnson as visual fields are containers, the word 'in' was overlooked as metaphorical by all three models.</p><p>The models were also challenged by metaphors in which an entity is being treated as a different kind of object, like instances of personification, place for institution, or producer for product metaphors. However, the GPT-4 models had a much greater ability to detect and provide accurate basic meanings for these. For example, in "Let's not let Thailand become another Vietnam. ", both 4-turbo and 4o correctly identified 'Vietnam' as a metaphor. 4o explained that the more basic meaning of the word "refers to the country, whereas in this context it refers to a situation similar to the Vietnam War" and 4-turbo provides a similar response. Likewise, in"I hate to read Heidegger. ", 4o recognized that 'Heidegger' is being used metaphorically and stated that the "more basic meaning refers to the person Martin Heidegger, a German philosopher, rather than his works. " Identifying and explaining both of these metaphors requires a nuanced understanding of both the semantics of the sentence and the cultural context surrounding them. The models cannot always perform this analysis (all three miss that 'the Alamo' is metaphorical in "Remember the Alamo!"), but it is impressive that they are ever able to do so. In addition, forcing the models to annotate word-by-word makes it challenging for them to identify metaphors comprising multi-word units. For example, it is difÏcult to determine which words should be marked as having more basic meanings in the sentence "Get the most out of life. " Analyzing texts word-by-word can also make identifying and providing the basic meanings of metaphorical compound words more complex. For example, both 'underage' and 'brainchild' cause problems for 3.5-turbo; it does not recognize that the under of underage is metaphorical and it says that the more basic meaning of brainchild is "a child conceived in the mind. " In contrast, 4-turbo says that the "more basic meanings of 'brain' and 'child' are more concrete and related to physical objects or beings" and notes that the "more basic meaning of 'under' is physically beneath something" for underage. 4o recognizes both words as metaphorical, but provides worse basic meanings.</p><p>While the models are clearly fallible, they nevertheless demonstrate an impressive ability to synthesize semantic, syntactic, and cultural information. They are frequently able to recognize when a word is being used metaphorically and often provide a correct basic meaning for the word. This ability sometimes holds even for nuanced and complex metaphors.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.">Conclusion</head><p>We find that large, generative LMs are capable of applying the classic metaphor annotation procedure, MIP. In doing so, they demonstrate an ability to discern the "basic meaning" of words and thus a depth of linguistic understanding that is not obviously gleaned from language modeling pre-training. Notably, they also demonstrate that LLMs are able to execute linguistic procedures designed for human annotators. This capacity means that generative LLMs are a promising tool for large-scale computational research on conceptual metaphors, which has previously been largely infeasible. In addition, it suggests that further research on how metaphoricity is learned by models may provide insight into their ability to acquire complex linguistic knowledge that often relies on people's embodied experiences. These findings also suggest several avenues for future research, including studies into where information about words' basic meanings or metaphoricity is stored by models, an exploration of open source models ability to annotate for conceptual metaphor, and research on operationalizing other linguistic procedures for execution by LLMs.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1:The metaphor identification procedure (MIP) introduced by the Pragglejaz Group in<ref type="bibr" target="#b13">[14]</ref>.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figure 3 :</head><label>3</label><figDesc>Figure 3: Confusion matrices for results on the TroFi dataset for each model. The rows represent the true values for each sample, and the columns represent the model labels.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>Figure 4 :</head><label>4</label><figDesc>Figure 4: The percentage of samples from Lakoff and Johnson where all metaphors and basic meanings are correctly identified.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_3"><head>Figure 5 :</head><label>5</label><figDesc>Figure 5: The percentage of samples from Lakoff and Johnson in which additional metaphors have been labeled. Smaller bars are included which represent the percentage of additional words for which the correct basic meaning is provided (upper bar) and which are plausible metaphors (lower bar).</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>• L&amp;J Metaphor(s) -Correct Basic Meanings:</head><label></label><figDesc>Whether the model has provided a correct, or plausible, more basic meaning for all metaphorical words identified in the category above. A label of 1 is applied only if all metaphors from Lakoff and Johnson have been correctly identified.• Additional Annotations: Whether the model has labeled any words not highlighted by Lakoff and Johnson as having a more basic meaning. •</figDesc><table /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_3"><head>Additional Annotations -Metaphorical:</head><label></label><figDesc>Whether all the additional words labeled by the model are plausibly metaphorical or may have a more basic meaning. •</figDesc><table /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_4"><head>Additional Annotations -Correct Basic Meanings: Whether</head><label></label><figDesc>the model has provided a correct, or plausible, basic meaning for all additional labeled words. It is possible for models to give the correct basic meaning for a word even if it is not plausibly metaphorical.</figDesc><table /></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_0">gpt-3.5-turbo-0125</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_1">gpt-4-turbo-2024-04-09</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_2">gpt-4o-2024-05-13</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="4" xml:id="foot_3">https://github.com/rmatouschekh/science-is-exploration</note>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgments</head><p>We would like to thank Lavinia Cerioni for her help in brainstorming this project and our colleagues at the Center for Humanities Computing for their support and advice. Additionally, part of the computation done for this project was performed on the UCloud interactive HPC system, which is managed by the eScience Center at the University of Southern Denmark.</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Metaphors in Pre-Trained Language Models: Probing and Generalization Across Datasets and Languages</title>
		<author>
			<persName><forename type="first">E</forename><surname>Aghazadeh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Fayyaz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Yaghoobzadeh</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/2022.acl-long.144</idno>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics</title>
				<meeting>the 60th Annual Meeting of the Association for Computational Linguistics<address><addrLine>Dublin, Ireland</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2022">2022</date>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="page" from="2037" to="2050" />
		</imprint>
	</monogr>
	<note>Association for Computational Linguistics</note>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Metaphor and Artificial Intelligence: Why They Matter to Each Other</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">A</forename><surname>Barnden</surname></persName>
		</author>
		<idno type="DOI">10.1017/cbo9780511816802.020</idno>
	</analytic>
	<monogr>
		<title level="m">The Cambridge Handbook of Metaphor and Thought</title>
				<imprint>
			<publisher>Cambridge University Press</publisher>
			<date type="published" when="2008">2008</date>
			<biblScope unit="page" from="311" to="338" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">A Clustering Approach for Nearly Unsupervised Recognition of Nonliteral Language</title>
		<author>
			<persName><forename type="first">J</forename><surname>Birke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Sarkar</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">11th Conference of the European Chapter of the Association for Computational Linguistics</title>
				<meeting><address><addrLine>Trento, Italy</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2006">2006</date>
			<biblScope unit="page" from="329" to="336" />
		</imprint>
	</monogr>
	<note>Association for Computational Linguistics</note>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Metaphoric Structuring: Understanding Time Through Spatial Metaphors</title>
		<author>
			<persName><forename type="first">L</forename><surname>Boroditsky</surname></persName>
		</author>
		<idno type="DOI">10.1016/s0010-0277(99)00073-6</idno>
	</analytic>
	<monogr>
		<title level="j">Cognition</title>
		<imprint>
			<biblScope unit="volume">75</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page" from="1" to="28" />
			<date type="published" when="2000">2000</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Metaphor: An Inescapable Phenomenon in Natural Language Comprehension</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">G</forename><surname>Carbonell</surname></persName>
		</author>
		<idno type="DOI">10.4324/9781315802671</idno>
	</analytic>
	<monogr>
		<title level="m">Strategies for Natural Language Processing</title>
				<meeting><address><addrLine>Pittsburgh, PA</addrLine></address></meeting>
		<imprint>
			<publisher>Psychology Press</publisher>
			<date type="published" when="1982">1982</date>
			<biblScope unit="page" from="415" to="434" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Time in the Mind: Using Space to Think About Time</title>
		<author>
			<persName><forename type="first">D</forename><surname>Casasanto</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Boroditsky</surname></persName>
		</author>
		<idno type="DOI">10.1016/j.cognition.2007.03.004</idno>
	</analytic>
	<monogr>
		<title level="j">Cognition</title>
		<imprint>
			<biblScope unit="volume">106</biblScope>
			<biblScope unit="issue">2</biblScope>
			<biblScope unit="page" from="579" to="593" />
			<date type="published" when="2008">2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<monogr>
		<title level="m" type="main">Cognitive Linguistics</title>
		<author>
			<persName><forename type="first">W</forename><surname>Croft</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">A</forename><surname>Cruse</surname></persName>
		</author>
		<idno type="DOI">10.1017/cbo9780511803864</idno>
		<imprint>
			<date type="published" when="2004">2004</date>
			<publisher>Cambridge University Press</publisher>
			<pubPlace>Cambridge, UK</pubPlace>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Mental Spaces in Grammar: Conditional Constructions</title>
		<author>
			<persName><forename type="first">B</forename><surname>Dancygier</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Sweetser</surname></persName>
		</author>
		<idno type="DOI">10.1017/cbo9780511486760</idno>
	</analytic>
	<monogr>
		<title level="s">Cambridge Studies in Linguistics</title>
		<imprint>
			<date type="published" when="2005">2005</date>
			<publisher>Cambridge University Press</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Corpus Techniques for Metaphor Analysis</title>
		<author>
			<persName><forename type="first">A</forename><surname>Deignan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Semino</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Metaphor Analysis</title>
				<meeting><address><addrLine>London, UK</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2010">2010</date>
			<biblScope unit="page" from="161" to="179" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">A Computational Framework for Conceptual Blending</title>
		<author>
			<persName><forename type="first">M</forename><surname>Eppe</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Maclean</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Confalonieri</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Kutz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Schorlemmer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Plaza</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K.-U</forename><surname>Kühnberger</surname></persName>
		</author>
		<idno type="DOI">10.1016/j.artint.2017.11.005</idno>
	</analytic>
	<monogr>
		<title level="j">Artificial Intelligence</title>
		<imprint>
			<biblScope unit="volume">256</biblScope>
			<biblScope unit="page" from="105" to="129" />
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<monogr>
		<author>
			<persName><forename type="first">G</forename><surname>Fauconnier</surname></persName>
		</author>
		<idno type="DOI">10.1017/cbo9780511624582</idno>
		<title level="m">Mental Spaces: Aspects of Meaning Construction in Natural Language</title>
				<meeting><address><addrLine>Cambridge, UK</addrLine></address></meeting>
		<imprint>
			<publisher>Cambridge University Press</publisher>
			<date type="published" when="1994">1994</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<monogr>
		<title level="m" type="main">The Way We Think: Conceptual Blending and the Mind&apos;s Hidden Complexities</title>
		<author>
			<persName><forename type="first">M</forename><surname>Fauconnier</surname></persName>
		</author>
		<author>
			<persName><surname>Turner</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2002">2002</date>
			<publisher>Basic Books</publisher>
			<pubPlace>New York City, NY</pubPlace>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Why Do Some People Dislike Conceptual Metaphor Theory?</title>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">W</forename><surname>Gibbs</surname></persName>
		</author>
		<idno type="DOI">10.1515/cogsem.2013.5.12.14</idno>
	</analytic>
	<monogr>
		<title level="j">Cognitive Semiotics</title>
		<imprint>
			<biblScope unit="volume">5</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page" from="14" to="36" />
			<date type="published" when="2009">2009</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">MIP: A Method for Identifying Metaphorically Used Words in Discourse</title>
		<author>
			<persName><forename type="first">P</forename><surname>Group</surname></persName>
		</author>
		<idno type="DOI">10.1207/s15327868ms2201\_1</idno>
	</analytic>
	<monogr>
		<title level="j">Metaphor and Symbol</title>
		<imprint>
			<biblScope unit="volume">22</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page" from="1" to="39" />
			<date type="published" when="2007">2007</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">What Corpus Linguistics Can Tell Us About Metaphor Use In Newspaper Texts</title>
		<author>
			<persName><forename type="first">T</forename><surname>Krennmayr</surname></persName>
		</author>
		<idno type="DOI">10.1080/1461670x.2014.937155</idno>
	</analytic>
	<monogr>
		<title level="j">Journalism Studies</title>
		<imprint>
			<biblScope unit="volume">16</biblScope>
			<biblScope unit="issue">4</biblScope>
			<biblScope unit="page" from="530" to="546" />
			<date type="published" when="2015">2015</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<monogr>
		<title level="m" type="main">Metaphors We Live By</title>
		<author>
			<persName><forename type="first">G</forename><surname>Lakoff</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Johnson</surname></persName>
		</author>
		<idno type="DOI">10.7208/chicago/9780226470993.001.0001</idno>
		<imprint>
			<date type="published" when="1980">1980</date>
			<publisher>The University of Chicago Press</publisher>
			<pubPlace>Chicago</pubPlace>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<monogr>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">W</forename><surname>Langacker</surname></persName>
		</author>
		<title level="m">Foundations of Cognitive Grammar: Volume I: Theoretical Prerequisites</title>
				<meeting><address><addrLine>Stanford, CA</addrLine></address></meeting>
		<imprint>
			<publisher>Stanford University Press</publisher>
			<date type="published" when="1987">1987</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<monogr>
		<title level="m" type="main">Foundations of Cognitive Grammar: Volume II: Descriptive Application</title>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">W</forename><surname>Langacker</surname></persName>
		</author>
		<idno type="DOI">10.1515/9780804764469</idno>
		<imprint>
			<date type="published" when="1991">1991</date>
			<publisher>Stanford University Press</publisher>
			<pubPlace>Stanford, CA</pubPlace>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<analytic>
		<title level="a" type="main">Semiology and Rhetoric</title>
		<author>
			<persName><forename type="first">P</forename><surname>De Man</surname></persName>
		</author>
		<idno type="DOI">10.2307/464524</idno>
	</analytic>
	<monogr>
		<title level="j">Diacritics</title>
		<imprint>
			<biblScope unit="volume">3</biblScope>
			<biblScope unit="issue">3</biblScope>
			<biblScope unit="page" from="27" to="33" />
			<date type="published" when="1973">1973</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<analytic>
		<title level="a" type="main">Word Embedding and WordNet Based Metaphor Identification and Interpretation</title>
		<author>
			<persName><forename type="first">R</forename><surname>Mao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Lin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Guerin</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/p18-1113</idno>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics</title>
		<title level="s">Long Papers</title>
		<meeting>the 56th Annual Meeting of the Association for Computational Linguistics<address><addrLine>Melbourne, Australia</addrLine></address></meeting>
		<imprint>
			<publisher>Association for Computational Linguistics</publisher>
			<date type="published" when="2018">2018</date>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="page" from="1222" to="1231" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<analytic>
		<title level="a" type="main">Towards Automatic Conceptual Metaphor Detection for Psychological Tasks</title>
		<author>
			<persName><forename type="first">P</forename><forename type="middle">V</forename><surname>Panicheva</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><forename type="middle">D</forename><surname>Mamaev</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">A</forename><surname>Litvinova</surname></persName>
		</author>
		<idno type="DOI">10.1016/j.ipm.2022.103191</idno>
	</analytic>
	<monogr>
		<title level="j">Information Processing &amp; Management</title>
		<imprint>
			<biblScope unit="volume">60</biblScope>
			<biblScope unit="issue">2</biblScope>
			<biblScope unit="page" from="1" to="17" />
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b21">
	<analytic>
		<title level="a" type="main">Methods of Annotating and Identifying Metaphors in the Field of Natural Language Processing</title>
		<author>
			<persName><forename type="first">M</forename><surname>Ptiček</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Dobša</surname></persName>
		</author>
		<idno type="DOI">10.3390/fi15060201</idno>
	</analytic>
	<monogr>
		<title level="j">Future Internet</title>
		<imprint>
			<biblScope unit="volume">15</biblScope>
			<biblScope unit="issue">6</biblScope>
			<biblScope unit="page">201</biblScope>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b22">
	<analytic>
		<title level="a" type="main">Computer Understanding of Metaphorically Used Verbs</title>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">W</forename><surname>Russell</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">American Journal of Computational Linguistics</title>
		<imprint>
			<biblScope unit="volume">3</biblScope>
			<biblScope unit="issue">3</biblScope>
			<biblScope unit="page" from="1" to="74" />
			<date type="published" when="1976">1976</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b23">
	<analytic>
		<title level="a" type="main">Information and Experience in Metaphor: A Perspective From Computer Analysis</title>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">W</forename><surname>Russell</surname></persName>
		</author>
		<idno type="DOI">10.1207/s15327868ms0104\_1</idno>
	</analytic>
	<monogr>
		<title level="j">Metaphor and Symbolic Activity</title>
		<imprint>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="issue">4</biblScope>
			<biblScope unit="page" from="227" to="270" />
			<date type="published" when="1986">1986</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b24">
	<analytic>
		<title level="a" type="main">Metaphor, Cancer and the End of Life: A Corpus-Based Study</title>
		<author>
			<persName><forename type="first">E</forename><surname>Semino</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Demjén</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Hardie</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Payne</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Rayson</surname></persName>
		</author>
		<idno type="DOI">10.4324/9781315629834</idno>
	</analytic>
	<monogr>
		<title level="m">Routledge Advances in Corpus Linguistics</title>
				<meeting><address><addrLine>Oxford, UK</addrLine></address></meeting>
		<imprint>
			<publisher>Routledge</publisher>
			<date type="published" when="2018">2018</date>
			<biblScope unit="volume">19</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b25">
	<analytic>
		<title level="a" type="main">Design and Evaluation of Metaphor Processing Systems</title>
		<author>
			<persName><forename type="first">E</forename><surname>Shutova</surname></persName>
		</author>
		<idno type="DOI">10.1162/coli\_a\_00233</idno>
	</analytic>
	<monogr>
		<title level="j">Computational Linguistics</title>
		<imprint>
			<biblScope unit="volume">41</biblScope>
			<biblScope unit="issue">4</biblScope>
			<biblScope unit="page" from="579" to="623" />
			<date type="published" when="2015">2015</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b26">
	<monogr>
		<title level="m" type="main">Concept Structuring Systems</title>
		<author>
			<persName><forename type="first">L</forename><surname>Talmy</surname></persName>
		</author>
		<idno type="DOI">10.7551/mitpress/6847.001.0001</idno>
		<imprint>
			<date type="published" when="2000">2000</date>
			<publisher>The MIT Press</publisher>
			<biblScope unit="volume">1</biblScope>
			<pubPlace>Cambridge, MA</pubPlace>
		</imprint>
	</monogr>
	<note>Toward a Cognitive Semantics</note>
</biblStruct>

<biblStruct xml:id="b27">
	<monogr>
		<title level="m" type="main">Typology and Process in Concept Structuring</title>
		<author>
			<persName><forename type="first">L</forename><surname>Talmy</surname></persName>
		</author>
		<idno type="DOI">10.7551/mitpress/6848.001.0001</idno>
		<imprint>
			<date type="published" when="2000">2000</date>
			<publisher>The MIT Press</publisher>
			<biblScope unit="volume">2</biblScope>
			<pubPlace>Cambridge, MA</pubPlace>
		</imprint>
	</monogr>
	<note>Toward a Cognitive Semantics</note>
</biblStruct>

<biblStruct xml:id="b28">
	<analytic>
		<title level="a" type="main">How Linguistic Metaphor Scaffolds Reasoning</title>
		<author>
			<persName><forename type="first">P</forename><forename type="middle">H</forename><surname>Thibodeau</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">K</forename><surname>Hendricks</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Boroditsky</surname></persName>
		</author>
		<idno type="DOI">10.1016/j.tics.2017.07.001</idno>
	</analytic>
	<monogr>
		<title level="j">Trends in Cognitive Sciences</title>
		<imprint>
			<biblScope unit="volume">21</biblScope>
			<biblScope unit="issue">11</biblScope>
			<biblScope unit="page" from="852" to="863" />
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b29">
	<analytic>
		<title level="a" type="main">Recent Advances in Neural Metaphor Processing: A Linguistic, Cognitive and Social Perspective</title>
		<author>
			<persName><forename type="first">X</forename><surname>Tong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Shutova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Lewis</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/2021.naacl-main.372</idno>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies. Online: Association for Computational Linguistics</title>
				<meeting>the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies. Online: Association for Computational Linguistics</meeting>
		<imprint>
			<date type="published" when="2021">2021</date>
			<biblScope unit="page" from="4673" to="4686" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b30">
	<analytic>
		<title level="a" type="main">Does GPT-3 Grasp Metaphors? Identifying Metaphor Mappings with Generative Language Models</title>
		<author>
			<persName><forename type="first">L</forename><surname>Wachowiak</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Gromann</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/2023.acl-long.58</idno>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics</title>
				<meeting>the 61st Annual Meeting of the Association for Computational Linguistics<address><addrLine>Toronto, Canada</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2023">2023</date>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="page" from="1018" to="1032" />
		</imprint>
	</monogr>
	<note>Association for Computational Linguistics</note>
</biblStruct>

<biblStruct xml:id="b31">
	<analytic>
		<title level="a" type="main">Chain-of-Thought Prompting Elicits Reasoning in Large Language Models</title>
		<author>
			<persName><forename type="first">J</forename><surname>Wei</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Schuurmans</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Bosma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Ichter</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Xia</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><forename type="middle">H</forename><surname>Chi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><forename type="middle">V</forename><surname>Le</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Zhou</surname></persName>
		</author>
		<idno type="DOI">10.48550/arXiv.2201.11903</idno>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 36th International Conference on Neural Information Processing Systems</title>
				<meeting>the 36th International Conference on Neural Information Processing Systems<address><addrLine>New Orleans, LA</addrLine></address></meeting>
		<imprint>
			<publisher>Curran Associates Inc</publisher>
			<date type="published" when="2024">2024</date>
			<biblScope unit="page" from="1" to="14" />
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
