<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>SCIENCE IS EXPLORATION: Computational Frontiers for Conceptual Metaphor Theory</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Rebecca M. M.Hicke</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Ross DeansKristensen-McLachl</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Center for Humanities Computing, Aarhus University</institution>
          ,
          <country country="DK">Denmark</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Department of Computer Science, Cornell University</institution>
          ,
          <country country="US">USA</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Department of Linguistics, Cognitive Science, and Semiotics, Aarhus University</institution>
          ,
          <country country="DK">Denmark</country>
        </aff>
      </contrib-group>
      <fpage>1105</fpage>
      <lpage>1116</lpage>
      <abstract>
        <p>Metaphors are everywhere. They appear extensively across all domains of natural language, from the most sophisticated poetry to seemingly dry academic prose. A significant body of research in the cognitive science of language argues for the existencocnecoepftual metaphors, the systematic structuring of one domain of experience in the language of another. Conceptual metaphors are not simply rhetorical lfourishes but are crucial evidence of the role of analogical reasoning in human cognition. In this paper, we ask whether Large Language Models (LLMs) can accurately identify and explain the presence of such conceptual metaphors in natural language data. Using a novel prompting technique based on metaphor annotation guidelines, we demonstrate that LLMs are a promising tool for large-scale computational research on conceptual metaphors. Further, we show that LLMs are able to apply procedural guidelines designed for human annotators, displaying a surprising depth of linguistic knowledge.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;conceptual metaphor theory</kwd>
        <kwd>large language models</kwd>
        <kwd>pragglejaz</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>Conceptual metaphors are so deeply ingrained in our everyday language that the most
frequent usage of a word or phrase may be metaphorical. This can make it difÏcult to recognize
the presence of metaphorical language, even for linguists and other sp1e4c]i.aMlisotrse[over,
processing and understanding conceptual metaphors necessarily involves encyclopedic general
knowledge of the world, as well as shared cultural norms and stereotypes between
speaker/hearer or writer/reader. Identifying conceptual metaphors has therefore consistently proven to
be a challenge for both natural language processing (NLP) and artificial intelligence (AI). Given
that contemporary transformer-based models ‘learn’ by consuming huge amounts of natural
language and build knowledge by looking at how words are commonly used, it seems that they
should similarly struggle to recognize these metaphors.</p>
      <p>Nevertheless, being able to consistently identify the presence of conceptual metaphors in
discourse would be greatly beneficial for fields such as computational literary studies and cultural
analytics more broadly. In this paper, we therefore set out to test empirically whether LLMs
can be used productively in the context of conceptual metaphor theory (CMT). If the answer
is no, it may be evidence that developing computational approaches to conceptual metaphor
remains an intractable problem. If, on the other handc,aLnLcMosnsistently identify the
presence of conceptual metaphors, the question then arises of exactly how and where LLMs encode
metaphoricity.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related Work</title>
      <sec id="sec-2-1">
        <title>Cognitive Linguistic Foundations</title>
        <p>
          The study of CMT was inaugurated in the 1980s in Lakof and Johns1o6n]. [Central to the
argument in 1[
          <xref ref-type="bibr" rid="ref6">6</xref>
          ] is the fundamental role that metaphor plays in structuring everyday discourse.
Rather than simply being a literary or rhetorical flourish, Lakof and Johnson argue that
conceptual metaphors structure our common understanding of entities, processes, or entire semantic
domains. CMT played a crucial role in the development of non-generativist cognitive
linguistics [
          <xref ref-type="bibr" rid="ref17 ref18 ref27 ref28 ref7">7, 17, 18, 27, 28</xref>
          ]. An important development came with the introductcioonnceoptfual
blending (otherwise known acsonceptual integration), in which the kinds of analogical
reasoning underlying conceptual metaphors are argued to be more fundamental to human language
and cognition1[
          <xref ref-type="bibr" rid="ref1 ref12 ref8">1, 12, 8</xref>
          ]. While there are dissenting voic1e3s],[recent decades have seen a
growing body of empirical research which claims to provide experimental evidence in favour
of CMT [
          <xref ref-type="bibr" rid="ref29 ref4 ref6">4, 6, 29</xref>
          ].
        </p>
      </sec>
      <sec id="sec-2-2">
        <title>Metaphor and Artificial Intelligence</title>
        <p>
          The extensive role and function of metaphor in natural language has not escaped researchers
in NLP. Indeed, a great deal of early work in AI concerned exactly how a computer could make
sense of metaphorical langua5g,e23[
          <xref ref-type="bibr" rid="ref24">, 24</xref>
          ] (see [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ] for a comprehensive overview). Much of this
work relied on so-called symbolic AI, but struggled because GOFAI (good, old-fashioned AI)
could not provide the analogous reasoning skills and extensive general knowledge required
to process metaphorical language. In the context of conceptual blending specifically, there
have been tentative attempts to model the underlying processes involved comput1a0]t.ionally [
1. Read the entire text-discourse to establish a general
        </p>
        <p>understanding of the meaning.
2. Determine the lexical units in the text-discourse.
3.</p>
        <p>a. For each lexical unit in the text, establish its
meaning in context, that is, how it applies to an
entity, relation, or attribute in the situation evoked
by the text (contextual meaning). Take into account
what comes before and after the lexical unit.
b. For each lexical unit, determine if it has a more
basic contemporary meaning in other contexts than the
one in the given context. For our purposes, basic
meanings tend to be
i. More concrete; what they evoke is easier to</p>
        <p>imagine, see, hear, feel, smell, and taste.
ii. Related to bodily action.
iii. More precise (as opposed to vague).
iv. Historically older.</p>
        <p>Basic meanings are not necessarily the most frequent
meanings of the lexical unit.
c. If the lexical unit has a more basic
currentcontemporary meaning in other contexts than the given
context, decide whether the contextual meaning
contrasts with the basic meaning but can be understood
in comparison with it.</p>
        <p>4. If yes, mark the lexical unit as metaphorical.
However, the encyclopedic world knowledge required in CMT is challenging to encode since
the domain is essentially unbounded. Moreover, a core issue at the heart of computational
metaphor analysis has historically been the lack of a shared definition of what is being studied,
along with the scarcity of robust data and evaluation st2r6a]t.egies [</p>
      </sec>
      <sec id="sec-2-3">
        <title>Recent Developments</title>
        <p>
          However, recent years have again seen a growing interest in quantitative and computational
approaches to metaphor, primarily in more applied fields such as corpus linguistics and
stylistics [
          <xref ref-type="bibr" rid="ref15 ref25 ref9">9, 15, 25</xref>
          ]. With the advent of neural language models, research in NLP has increasingly
attempted to discover if (and how) word embeddings capture aspects of metap2h0o, r3i0c,ity [
          <xref ref-type="bibr" rid="ref21 ref22">21, 22</xref>
          ]. More recently, work has been done on explicitly probing pre-trained language
models to inspect how metaphorical knowledge is enco1d]eadn[d creating shared task datasets
[
          <xref ref-type="bibr" rid="ref30">30</xref>
          ]. In the context of LLMs specifically, GPT-3 has been shown to perform reasonably well
at predicting source domains for English metaphors, although it performs significantly worse
for Spanish 3[
          <xref ref-type="bibr" rid="ref1">1</xref>
          ]. These are promising developments and point to the increasing awareness of
the practical and theoretical challenges underlying the nature of figurative language in NLP.
To date, though, we are unaware of existing research which seeks to operationalize existing
metaphor identification procedures for LLMs.
For each word in the following text, determine if it has a
*more* basic contemporary meaning in other contexts than the
one in the given context. For our purposes, basic meanings
tend to be:
- more concrete; what they evoke is easier to imagine, see,
        </p>
        <p>here, feel, smell, and taste
- related to bodily action
- more precise (as opposed to vague)
- historically older
Basic meanings are not necessarily the most frequent meanings
of the word. Remember to only report YES for words whose use
in this context is not their most basic meaning.</p>
        <p>Examples:
“I demolished his argument.”
- I: NO (pronoun, basic meaning is the same)
- demolished: YES (verb, more basic meaning refers to</p>
        <p>destroying a building)
- his: NO (possessive pronoun, basic meaning is the same)
- argument: NO (noun, basic meaning is the same)
“They write about their family.”
- They: NO (pronoun, basic meaning is the same)
- write: NO (verb, basic meaning is the same)
- about: NO (preposition, basic meaning is the same)
- their: NO (possessive pronoun, basic meaning is the same)
- family: NO (noun, basic meaning is the same)</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Methods</title>
      <p>We are interested in studying whether LLMs are capable of leveraging the contextual and
cultural knowledge necessary to perform difÏcult metaphor identification tasks, such as
recognizing conceptual metaphors. To do this, we operationalize the Metaphor Identification Procedure
(MIP), a set of annotation guidelines introduced by the Pragglejaz Group1i4n].2T00h7e [steps
of MIP are detailed in Figu1r. eMIP and its variants are among the most commonly used
procedures for metaphor annotation in work in corpus linguistics and stylistics; we therefore take
it as a reasonable proxy for the metaphor identification process in humans.</p>
      <p>Examining MIP, the first few steps have clear parallels in transformer-based LLMs. Step 1
is performed by the attention mechanism. Creating the representation of each word, in
traditional attention mechanisms, requires ‘looking’ at every other word in the text, thus allowing
the model to “establish a general understanding of [its] meaning.” Step 2, determining the
lexical units in a text, is directly paralleled by tokenization. Finally, Step 3a is achieved via
contextual embeddings, whose explicit purpose is to represent the specific meaning of a word
in context. However, Step 3b of MIP has no clear counterpart in transformer-based models.
This is because the ‘basic meaning’ of a lexical unit, as defined by MIP, is not obviously stored
or represented within these models.</p>
      <p>
        We want to know if state-of-the-art LLMs are capable of replicating Step 3b of the MIP
procedure despite these barriers, so we transform it into a prompt for instruction-tuned models
(Figure2). In creating the prompt, we leave the text of Step 3b almost entirely intact. “Lexical
units” are replaced by “words” in order to encourage consistent outputs. Emphasis is added in
several locations to only annotating words whichmhoarveebaasic meaning than their usage
in the input text. Finally, two examples are appended to the prompt. These examples are meant
to facilitate in-context learning and provide a structure for the models to mimic in their outputs.
They instruct the model to provide a parenthetical expression next to each word consisting of
the word’s part of speech and its more basic meaning, if one exists. These expressions were
introduced by thgept-4o model itself during initial exploration and were then included in the
prompt because they force the model to provide explanations for its annotations. This allows
for further evaluation of the outputs and ultimately means they play a similar role to Chain of
Thought prompts, which have been shown to improve model performance across a variety of
tasks [
        <xref ref-type="bibr" rid="ref32">32</xref>
        ].
      </p>
      <p>We use this prompt to query three models from OpenAI’s GPT family, chosen because of
their size, popularity, and high performance on a wide range of tasks. Specifically, we look at
gpt-3.5-turbo,1 gpt-4-turbo,2 andgpt-4o.3 We access each model using OpenAI’s Chat
Completion API. A system prompt“Y(ou are a helpful assistant. You have extensive linguistic
knowledge.”) is provided each time the model is queried. Then, the main prompt (F2i)giusre
passed as a user message, with the current text of interest appended in quotations at the end. All
parameters except nuclear samplitnogp_(p) are left at their default values during prompting
for all models. We steotp_p to 0.1, which means that the models only select from the tokens
which make up the top 10% of the probability mass when choosing each next token. We adjust
the top_p value because we are not prompting for diverse or creative text, but instead for
reliable and accurate outputs, which we hypothesize will be more likely produced from
highprobability tokens. The cost of all experiments using the OpenAI A∼PI$1w0a0s.</p>
    </sec>
    <sec id="sec-4">
      <title>4. Data</title>
      <p>We evaluate each model on two datasets. First, we wish to study the models’ basic ability
to use MIP to correctly identify metaphorical words. To this end, we use the Trope Finder
(TroFi) dataset3][, which consists of sentences drawn from Wall Street Journal issues published
between 1987 and 1989. Each sentence contains one of a list of fity words whose usage is
annotated as literal or non-literal. We only consider sentences from the dataset that have been
labeled by humans and take their annotations as ground truth, excluding all sentences labeled
only by the TroFi clustering system. This leaves us with 3,736 sentences. Each model is then
prompted as described above for each sentence and evaluated only on the words annotated in
TroFi. A response is marked as correct if the model says “YES” there is a more basic meaning
for the word of interest and the label from TroFi is “literal” and vice versa. Occasionally, the
models provide labels for a phrase in the input sentence instead of individual words; the label
provided for the entire phrase is then applied to the word of interest if necessary. If a model
1gpt-3.5-turbo-0125
2gpt-4-turbo-2024-04-09
3gpt-4o-2024-05-13
does not provide a label for a word, we act as if the incorrect label was given.</p>
      <p>However, the TroFi dataset does not permit for a very fine-grained analysis of the models’
capabilities. In order to facilitate such an analysis, we create a new dataset consisting of
example sentences pulled from Lakof and Johnson’s original work on C1M6]T. W[e collect all full
sentence examples from the text that we believe can be understood as metaphorical without
context, leaving us with 544 sentences, which we refer toMasWtLhBedataset. The dataset is
made publicly available, along with all model outputs and human ann4otations.</p>
      <p>Each model is again prompted to provide word-level annotations for every sentence. Then,
we perform detailed qualitative analysis on each model’s output for a subset of 100 sentences.
Specifically, we evaluate each output on five binary categories:
• L&amp;J Metaphor(s) – Identified : Whether the model has correctly identiafielld
metaphors highlighted by Lakof and Johnson. For metaphorical phrases, the model’s
response is considered correct if it has identified at at least one key word as having a
more basic meaning.
• L&amp;J Metaphor(s) – Correct Basic Meanings: Whether the model has provided a
correct, or plausible, more basic meaningafllomretaphorical words identified in the
category above. A label of 1 is applied only if all metaphors from Lakof and Johnson have
been correctly identified.
• Additional Annotations: Whether the model has labeled any words not highlighted
by Lakof and Johnson as having a more basic meaning.
• Additional Annotations – Metaphorical: Whetherall the additional words labeled
by the model are plausibly metaphorical or may have a more basic meaning.
• Additional Annotations – Correct Basic Meanings: Whether the model has
provided a correct, or plausible, basic meaningallfoardditional labeled words. It is
possible for models to give the correct basic meaning for a word even if it is not plausibly
metaphorical.</p>
      <p>All annotations are performed by one of the authors, using MIP as a guideline. Confusing
or difÏcult annotations are then discussed between both authors till a consensus is reached.
These annotations are necessarily subjective, and we do not claim that they represent a ground
truth. Instead, since these evaluations are informed by the same procedure used to create the
4https://github.com/rmatouschekh/science-is-exploration</p>
      <p>TroFi Dataset  Model Confusion Matrices
4-turbo
outputs, we claim that they allow us to studpylatuhsiebility of the models’ performance, if
not strictly thaecircuracy. In future work, further evaluation of model outputs with multiple
trained annotators may allow for a more concrete analysis of model performance.</p>
    </sec>
    <sec id="sec-5">
      <title>5. Results and Discussion</title>
      <p>On the TroFi dataset, all models demonstrate an ability to distinguish between literal and
nonliteral word usage with generally better than chance performan1c).eT(Thaubs,ltehey seem to
be able to apply the MIP procedure for identifying metaphors and determine computationally
whether a word has a more basic meaning. This is surprising since, as discussed above, basic
word meanings are not necessarily the most frequent. Nothing about the language modeling
approach to pre-training models guarantees that a model would be able to diferentiate between
the most frequent and most basic meaning of a word.</p>
      <p>We also find, however, that the models are prone to over-labeling words as metaphorical
(Figure3). Whereas all three models achieve high recall for metaphor identification, they
frequently label words used literally as having a more basic meaning, leading to low recall for
the ‘literal’ category and low precision for the ‘metaphorical’ c3a.t5e-gtourryb.o
particularly struggles with accurately identifying literal word usage. It is worth noting that, although
there are clear cases where the models have failed to properly apply MIP, there may also be a
disconnect between annotations in the TroFi corpus and what is considered metaphorical by
the MIP procedure, as MIP was not used for annotation in TroFi.</p>
      <p>For the annotated subset of the MWLB dataset, all the models found all the Lakof and
Johnson metaphors in over 60% of sentences and further provided a correct basic meaning
for the metaphorical words in over 50% of sentences (F4)i.gu4oreperformed the best on this
task, achieving a remarkable performance given the subtlety and complexity of many of the
metaphors and the cultural knowledge required to identify both their metaphoricity and their
basic meanings.4o’s high performance may be due in part to the fact that it was most likely to
label words as metaphorical overall, having annotated additional words in the largest
percent4o
age of sentences (Figur5e). In only 48% of sentences did these additional annotations identify
only plausibly metaphorical words. Nonetheless, the inclusion of correct basic meanings for
such a large proportion of the Lakof and Johnson metaphors suggests that MIP was overall
being correctly applied and true ‘knowledge’ was being demonstrated, not just random chance.</p>
      <p>3.5-turbo struggled the most at replicating the desired output structure, occasionally
excluding words or truncating sentences, inconsistently formatting lists, and sometimes
dropping the parenthetical explanations. This afected its accuracy, as the model sometimes failed
to provide annotations for words central to the metaphors of interest. A4d-dtiutriobnoally,
was least likely to annotate a word as having a more basic meaning5)(.FiTguhrise meant
it identified fewer Lakof and Johnson metaphors, but a greater proportion of the additional
metaphors it annotated were plausible. The basic meanings provi4d-etdurbbyo were usually
of high quality, and tended to be more accurate and specific than those prod3u.c5e-dtbuyrbo.
Despite the benefits, however, this model’s ‘caution’ led it to perform wor4seo tohvaernall.</p>
      <p>Each model struggled to correctly identify when smaller function words were used
metaphorically, particularly prepositions like ‘in,’ . This made the models worse at identifying so-called
container metaphors, such as life is a container and activities are containers. For
example, in the sentence “Thati n’sthecenter of myfield of vision.”, labeled by Lakof and Johnson
as visual fields are containers, the word ‘in’ was overlooked as metaphorical by all three
models.</p>
      <p>The models were also challenged by metaphors in which an entity is being treated as a
diferent kind of object, like instances of personification, place for institution, or producer for
product metaphors. However, the GPT-4 models had a much greater ability to detect and
provide accurate basic meanings for these. For example, in “Let’s not let Thailand become another
Vietnam.”, both4-turbo and4o correctly identified ‘Vietnam’ as a metaph4oore.xplained that
the more basic meaning of the word “refers to the country, whereas in this context it refers to a
situation similar to the Vietnam War4”-atnudrbo provides a similar response. Likewise, in“I
hate to reaHdeidegger.”, 4o recognized that ‘Heidegger’ is being used metaphorically and stated
that the “more basic meaning refers to the person Martin Heidegger, a German philosopher,
rather than his works.” Identifying and explaining both of these metaphors requires a nuanced
understanding of both the semantics of the sentence and the cultural context surrounding them.
% of Samples with Additional Metaphors Labeled
4o
The models cannot always perform this analysis (all three miss that ‘the Alamo’ is metaphorical
in “Rememberthe Alamo!”), but it is impressive that they are ever able to do so.</p>
      <p>In addition, forcing the models to annotate word-by-word makes it challenging for them
to identify metaphors comprising multi-word units. For example, it is difÏcult to determine
which words should be marked as having more basic meanings in the sentGeentctehe“most
out of life.” Analyzing texts word-by-word can also make identifying and providing the basic
meanings of metaphorical compound words more complex. For example, both ‘underage’ and
‘brainchild’ cause problems f3o.r5-turbo; it does not recognize that utnhdeer of underage
is metaphorical and it says that the more basic meaning of brainchild is “a child conceived
in the mind.” In contras4t-,turbo says that the “more basic meanings of ‘brain’ and ‘child’
are more concrete and related to physical objects or beings” and notes that the “more basic
meaning of ‘under’ is physically beneath something” for unde4roargeec. ognizes both words
as metaphorical, but provides worse basic meanings.</p>
      <p>While the models are clearly fallible, they nevertheless demonstrate an impressive ability to
synthesize semantic, syntactic, and cultural information. They are frequently able to recognize
when a word is being used metaphorically and often provide a correct basic meaning for the
word. This ability sometimes holds even for nuanced and complex metaphors.</p>
    </sec>
    <sec id="sec-6">
      <title>6. Conclusion</title>
      <p>We find that large, generative LMs are capable of applying the classic metaphor annotation
procedure, MIP. In doing so, they demonstrate an ability to discern the “basic meaning” of
words and thus a depth of linguistic understanding that is not obviously gleaned from language
modeling pre-training. Notably, they also demonstrate that LLMs are able to execute
linguistic procedures designed for human annotators. This capacity means that generative LLMs
are a promising tool for large-scale computational research on conceptual metaphors, which
has previously been largely infeasible. In addition, it suggests that further research on how
metaphoricity is learned by models may provide insight into their ability to acquire complex
linguistic knowledge that often relies on people’s embodied experiences. These findings also
suggest several avenues for future research, including studies into where information about
words’ basic meanings or metaphoricity is stored by models, an exploration of open source
models ability to annotate for conceptual metaphor, and research on operationalizing other
linguistic procedures for execution by LLMs.</p>
    </sec>
    <sec id="sec-7">
      <title>Acknowledgments</title>
      <p>We would like to thank Lavinia Cerioni for her help in brainstorming this project and our
colleagues at the Center for Humanities Computing for their support and advice. Additionally,
part of the computation done for this project was performed on the UCloud interactive HPC
system, which is managed by the eScience Center at the University of Southern Denmark.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>E.</given-names>
            <surname>Aghazadeh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Fayyaz</surname>
          </string-name>
          , and
          <string-name>
            <given-names>Y.</given-names>
            <surname>Yaghoobzadeh</surname>
          </string-name>
          . “
          <article-title>Metaphors in Pre-Trained Language Models: Probing and Generalization Across Datasets and LanguagPerosc”.eeIndi:ngs of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)</article-title>
          . Dublin, Ireland: Association for Computational Linguistics,
          <year>2022</year>
          , pp.
          <fpage>2037</fpage>
          -
          <lpage>2050</lpage>
          . doi:
          <volume>10</volume>
          .18653/v1/
          <year>2022</year>
          .
          <article-title>acl-long</article-title>
          .
          <volume>14</volume>
          .4
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>J. A.</given-names>
            <surname>Barnden</surname>
          </string-name>
          . “Metaphor and Artificial Intelligence:
          <article-title>Why They Matter to Each Other”</article-title>
          .
          <source>In: The Cambridge Handbook of Metaphor and Thought</source>
          . Cambridge University Press,
          <year>2008</year>
          , pp.
          <fpage>311</fpage>
          -
          <lpage>338</lpage>
          . doi:
          <volume>10</volume>
          .1017/cbo9780511816802.
          <fpage>020</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>J.</given-names>
            <surname>Birke</surname>
          </string-name>
          and
          <string-name>
            <given-names>A.</given-names>
            <surname>Sarkar</surname>
          </string-name>
          .
          <article-title>“A Clustering Approach for Nearly Unsupervised Recognition of Nonliteral Language”. 1In1t:h Conference of the European Chapter of the Association for Computational Linguistics</article-title>
          . Trento, Italy: Association for Computational Linguistics,
          <year>2006</year>
          , pp.
          <fpage>329</fpage>
          -
          <lpage>336</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>L.</given-names>
            <surname>Boroditsky</surname>
          </string-name>
          . “Metaphoric Structuring:
          <article-title>Understanding Time Through Spatial Metaphors”</article-title>
          .
          <source>InC:ognition 75.1</source>
          (
          <issue>2000</issue>
          ), pp.
          <fpage>1</fpage>
          -
          <lpage>28</lpage>
          . doi:
          <volume>10</volume>
          .1016/s0010-
          <volume>0277</volume>
          (
          <issue>99</issue>
          )
          <fpage>00073</fpage>
          -
          <lpage>6</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <surname>J. G</surname>
          </string-name>
          . Carbonell. “Metaphor:
          <article-title>An Inescapable Phenomenon in Natural Language Comprehension”</article-title>
          . In:
          <article-title>Strategies for Natural Language Processing</article-title>
          . Pittsburgh, PA: Psychology Press,
          <year>1982</year>
          , pp.
          <fpage>415</fpage>
          -
          <lpage>434</lpage>
          . doi:
          <volume>10</volume>
          .4324/9781315802671.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>D.</given-names>
            <surname>Casasanto</surname>
          </string-name>
          and
          <string-name>
            <given-names>L.</given-names>
            <surname>Boroditsky</surname>
          </string-name>
          . “
          <article-title>Time in the Mind: Using Space to Think About Time”</article-title>
          .
          <source>In: Cognition 106.2</source>
          (
          <issue>2008</issue>
          ), pp.
          <fpage>579</fpage>
          -
          <lpage>593</lpage>
          . doi:
          <volume>10</volume>
          .1016/j.cognition.
          <year>2007</year>
          .
          <volume>03</volume>
          .00.4
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>W.</given-names>
            <surname>Croft</surname>
          </string-name>
          and
          <string-name>
            <given-names>D. A. Cruse. Cognitive</given-names>
            <surname>Linguistics</surname>
          </string-name>
          . Cambridge, UK: Cambridge University Press,
          <year>2004</year>
          . doi:
          <volume>10</volume>
          .1017/cbo9780511803864.
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>B.</given-names>
            <surname>Dancygier</surname>
          </string-name>
          and E. SweetseMr. ental Spaces in Grammar: Conditional Constructions. Cambridge Studies in Linguistics. Cambridge, UK: Cambridge University Press,
          <year>2005</year>
          . doi:
          <volume>10</volume>
          .1017/cbo9780511486760.
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>A.</given-names>
            <surname>Deignan</surname>
          </string-name>
          and
          <string-name>
            <surname>E. Semino.</surname>
          </string-name>
          “
          <article-title>Corpus Techniques for Metaphor AnalysisM”</article-title>
          .
          <source>eItna:phor Analysis. London, UK</source>
          ,
          <year>2010</year>
          , pp.
          <fpage>161</fpage>
          -
          <lpage>179</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>M.</given-names>
            <surname>Eppe</surname>
          </string-name>
          , E. Maclean,
          <string-name>
            <given-names>R.</given-names>
            <surname>Confalonieri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Kutz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Schorlemmer</surname>
          </string-name>
          , E. Plaza,
          <article-title>and</article-title>
          K.-U. Kühnberger.
          <article-title>“A Computational Framework for Conceptual BlendiAnrgt”i</article-title>
          .
          <source>ficiIanl:Intelligence</source>
          <volume>256</volume>
          (
          <year>2018</year>
          ), pp.
          <fpage>105</fpage>
          -
          <lpage>129</lpage>
          . doi:
          <volume>10</volume>
          .1016/j.artint.
          <year>2017</year>
          .
          <volume>11</volume>
          .00. 5
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>G.</given-names>
            <surname>FauconnierM.ental Spaces</surname>
          </string-name>
          :
          <article-title>Aspects of Meaning Construction in Natural Language</article-title>
          . Cambridge, UK: Cambridge University Press,
          <year>1994</year>
          . d1o0i.:
          <volume>1017</volume>
          /cbo9780511624582.
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>G.</given-names>
            <surname>Fauconnier and M. TurneTrh. e Way We</surname>
          </string-name>
          <article-title>Think: Conceptual Blending and the Mind's Hidden Complexities</article-title>
          . New York City, NY: Basic Books,
          <year>2002</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>R. W.</given-names>
            <surname>Gibbs</surname>
          </string-name>
          . “
          <string-name>
            <surname>Why Do Some People Dislike Conceptual Metaphor Theory?C”</surname>
          </string-name>
          <article-title>oIgnn: itive Semiotics 5</article-title>
          .1 (
          <issue>2009</issue>
          ), pp.
          <fpage>14</fpage>
          -
          <lpage>36</lpage>
          . doi:
          <volume>10</volume>
          .1515/cogsem.
          <year>2013</year>
          .
          <volume>5</volume>
          .12.14.
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>P.</given-names>
            <surname>Group</surname>
          </string-name>
          . “
          <article-title>MIP: A Method for Identifying Metaphorically Used Words in Discourse”</article-title>
          .
          <source>In: Metaphor and Symbol 22.1</source>
          (
          <issue>2007</issue>
          ), pp.
          <fpage>1</fpage>
          -
          <lpage>39</lpage>
          . doi:
          <volume>10</volume>
          .1207/s15327868ms2201\_
          <fpage>1</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>T.</given-names>
            <surname>Krennmayr</surname>
          </string-name>
          . “
          <article-title>What Corpus Linguistics Can Tell Us About Metaphor Use In Newspaper Texts”</article-title>
          .
          <source>In:Journalism Studies 16.4</source>
          (
          <issue>2015</issue>
          ), pp.
          <fpage>530</fpage>
          -
          <lpage>546</lpage>
          . doi:
          <volume>10</volume>
          .1080/1461670x.
          <year>2014</year>
          .
          <volume>9371</volume>
          55.
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>G.</given-names>
            <surname>Lakof</surname>
          </string-name>
          and
          <string-name>
            <given-names>M.</given-names>
            <surname>Johnson</surname>
          </string-name>
          . Metaphors We Live By. Chicago: The University of Chicago Press,
          <year>1980</year>
          . doi:
          <volume>10</volume>
          .7208/chicago/9780226470993.001.0001.
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>R. W.</given-names>
            <surname>Langacker</surname>
          </string-name>
          .
          <source>Foundations of Cognitive Grammar: Volume I: Theoretical Prerequisites</source>
          . Stanford, CA: Stanford University Press,
          <year>1987</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>R. W.</given-names>
            <surname>Langacker</surname>
          </string-name>
          .
          <source>Foundations of Cognitive Grammar:</source>
          Volume II:
          <article-title>Descriptive Application</article-title>
          . Stanford, CA: Stanford University Press,
          <year>1991</year>
          .
          <year>1d0</year>
          .
          <year>o1i5</year>
          :
          <volume>15</volume>
          /9780804764469.
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <surname>P. de Man</surname>
          </string-name>
          . “
          <article-title>Semiology and Rhetoric”</article-title>
          .
          <source>DIni:acritics 3</source>
          .3 (
          <issue>1973</issue>
          ), pp.
          <fpage>27</fpage>
          -
          <lpage>33</lpage>
          . doi:
          <volume>10</volume>
          .2307/46 4524.
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>R.</given-names>
            <surname>Mao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Lin</surname>
          </string-name>
          , and
          <string-name>
            <given-names>F.</given-names>
            <surname>Guerin</surname>
          </string-name>
          . “
          <article-title>Word Embedding and WordNet Based Metaphor Identification and Interpretation”P. rIonc:eedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)</article-title>
          . Melbourne, Australia: Association for Computational Linguistics,
          <year>2018</year>
          , pp.
          <fpage>1222</fpage>
          -
          <lpage>1231</lpage>
          .
          <year>d1o0</year>
          .
          <source>i1:8653/v1/p18-1113.</source>
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>P. V.</given-names>
            <surname>Panicheva</surname>
          </string-name>
          ,
          <string-name>
            <given-names>I. D.</given-names>
            <surname>Mamaev</surname>
          </string-name>
          , and
          <string-name>
            <given-names>T. A.</given-names>
            <surname>Litvinova</surname>
          </string-name>
          . “
          <article-title>Towards Automatic Conceptual Metaphor Detection for Psychological TasksIn”</article-title>
          .
          <source>fIonrm:ation Processing &amp; Management 60.2</source>
          (
          <issue>2023</issue>
          ), pp.
          <fpage>1</fpage>
          -
          <lpage>17</lpage>
          . doi:
          <volume>10</volume>
          .1016/j.ipm.
          <year>2022</year>
          .
          <volume>103191</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>M.</given-names>
            <surname>Ptiček</surname>
          </string-name>
          and
          <string-name>
            <given-names>J.</given-names>
            <surname>Dobša</surname>
          </string-name>
          . “
          <article-title>Methods of Annotating and Identifying Metaphors in the Field of Natural Language Processing”</article-title>
          .
          <source>FIuntu:re Internet 15.6</source>
          (
          <issue>2023</issue>
          ), p.
          <fpage>201</fpage>
          . doi:
          <volume>10</volume>
          .3390/fi1 5060201.
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <given-names>S. W.</given-names>
            <surname>Russell</surname>
          </string-name>
          . “
          <article-title>Computer Understanding of Metaphorically Used VerAbsm”</article-title>
          .
          <source>erInic:an Journal of Computational Linguistics 3.3</source>
          (
          <issue>1976</issue>
          ), pp.
          <fpage>1</fpage>
          -
          <lpage>74</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [24]
          <string-name>
            <given-names>S. W.</given-names>
            <surname>Russell</surname>
          </string-name>
          . “
          <article-title>Information and Experience in Metaphor: A Perspective From Computer Analysis”</article-title>
          .
          <source>In:Metaphor and Symbolic Activity</source>
          <volume>1</volume>
          .4 (
          <issue>1986</issue>
          ), pp.
          <fpage>227</fpage>
          -
          <lpage>270</lpage>
          . doi:
          <volume>10</volume>
          .1207/s153 27868ms0104\_
          <fpage>1</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          [25]
          <string-name>
            <given-names>E.</given-names>
            <surname>Semino</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Demjén</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Hardie</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Payne</surname>
          </string-name>
          , and P. RaysoMne.taphor,
          <article-title>Cancer and the End of Life: A Corpus-Based Study</article-title>
          . Vol.
          <volume>19</volume>
          . Routledge Advances in Corpus Linguistics. Oxford, UK: Routledge,
          <year>2018</year>
          . doi1:
          <fpage>0</fpage>
          .4324/9781315629834.
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          [26]
          <string-name>
            <surname>E. Shutova.</surname>
          </string-name>
          “
          <article-title>Design and Evaluation of Metaphor Processing SystemCos”m</article-title>
          .
          <source>Ipnu:tational Linguistics 41.4</source>
          (
          <issue>2015</issue>
          ), pp.
          <fpage>579</fpage>
          -
          <lpage>623</lpage>
          . doi:
          <volume>10</volume>
          .1162/coli\_a\_
          <volume>00233</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          [27]
          <string-name>
            <given-names>L.</given-names>
            <surname>Talmy</surname>
          </string-name>
          .
          <source>Toward a Cognitive Semantics</source>
          , Volume
          <volume>1</volume>
          :
          <string-name>
            <given-names>Concept</given-names>
            <surname>Structuring Systems</surname>
          </string-name>
          . Cambridge, MA: The MIT Press,
          <year>2000</year>
          . doi1:
          <fpage>0</fpage>
          .7551/mitpress/6847.001.0001.
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          [28]
          <string-name>
            <given-names>L.</given-names>
            <surname>Talmy</surname>
          </string-name>
          .
          <source>Toward a Cognitive Semantics</source>
          , Volume
          <volume>2</volume>
          :
          <article-title>Typology and Process in Concept Structuring</article-title>
          . Cambridge, MA: The MIT Press,
          <year>2000</year>
          . doi1:
          <fpage>0</fpage>
          .7551/mitpress/6848.001.0001.
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          [29]
          <string-name>
            <given-names>P. H.</given-names>
            <surname>Thibodeau</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. K.</given-names>
            <surname>Hendricks</surname>
          </string-name>
          , and
          <string-name>
            <given-names>L.</given-names>
            <surname>Boroditsky</surname>
          </string-name>
          . “
          <article-title>How Linguistic Metaphor Scafolds Reasoning”</article-title>
          .
          <source>In:Trends in Cognitive Sciences 21.11</source>
          (
          <year>2017</year>
          ), pp.
          <fpage>852</fpage>
          -
          <lpage>863</lpage>
          . doi:
          <volume>10</volume>
          .1016/j.tics .
          <year>2017</year>
          .
          <volume>07</volume>
          .001.
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          [30]
          <string-name>
            <given-names>X.</given-names>
            <surname>Tong</surname>
          </string-name>
          , E. Shutova, and
          <string-name>
            <given-names>M.</given-names>
            <surname>Lewis</surname>
          </string-name>
          . “
          <source>Recent Advances in Neural Metaphor Processing: A Linguistic</source>
          ,
          <article-title>Cognitive and Social PerspectiveP”.roInce:edings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies</article-title>
          .
          <source>Online: Association for Computational Linguistics</source>
          ,
          <year>2021</year>
          , pp.
          <fpage>4673</fpage>
          -
          <lpage>4686</lpage>
          . doi:
          <volume>10</volume>
          .18653/v1/
          <year>2021</year>
          .naacl-main.
          <volume>37</volume>
          .2
        </mixed-citation>
      </ref>
      <ref id="ref31">
        <mixed-citation>
          [31]
          <string-name>
            <given-names>L.</given-names>
            <surname>Wachowiak</surname>
          </string-name>
          and
          <string-name>
            <given-names>D.</given-names>
            <surname>Gromann</surname>
          </string-name>
          . “
          <article-title>Does GPT-3 Grasp Metaphors? Identifying Metaphor Mappings with Generative Language Models”</article-title>
          .
          <article-title>PIrnoc:eedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)</article-title>
          . Toronto, Canada: Association for Computational Linguistics,
          <year>2023</year>
          , pp.
          <fpage>1018</fpage>
          -
          <lpage>1032</lpage>
          .
          <year>10d</year>
          .
          <source>o1i8:653/v1/2023 .acl-long.5.8</source>
        </mixed-citation>
      </ref>
      <ref id="ref32">
        <mixed-citation>
          [32]
          <string-name>
            <given-names>J.</given-names>
            <surname>Wei</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Schuurmans</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Bosma</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Ichter</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Xia</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E. H.</given-names>
            <surname>Chi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q. V.</given-names>
            <surname>Le</surname>
          </string-name>
          , and
          <string-name>
            <given-names>D.</given-names>
            <surname>Zhou</surname>
          </string-name>
          . “
          <article-title>Chain-of-Thought Prompting Elicits Reasoning in Large Language Models”</article-title>
          .
          <source>In: Proceedings of the 36th International Conference on Neural Information Processing Systems</source>
          . New Orleans, LA: Curran Associates Inc.,
          <year>2024</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>14</lpage>
          .
          <year>d1o0i</year>
          .
          <volume>4</volume>
          :
          <issue>8550</issue>
          /arXiv.2201.11903.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>