<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">A is the B of C: (Semi)-Automatic Creation of Vossian Antonomasias</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Johanna</forename><surname>Rockstroh</surname></persName>
							<email>rockstro@uni-bremen.de</email>
							<affiliation key="aff0">
								<orgName type="institution">University of Bremen</orgName>
							</affiliation>
							<affiliation key="aff7">
								<orgName type="department">Team &quot;Mordor&quot; at the International Semantic Web Summer School</orgName>
								<address>
									<postCode>2023</postCode>
									<settlement>Bertinoro</settlement>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Giada</forename><surname>D'ippolito</surname></persName>
							<email>giadadippolito30@gmail.com</email>
							<affiliation key="aff1">
								<orgName type="institution">University of Genova</orgName>
							</affiliation>
							<affiliation key="aff7">
								<orgName type="department">Team &quot;Mordor&quot; at the International Semantic Web Summer School</orgName>
								<address>
									<postCode>2023</postCode>
									<settlement>Bertinoro</settlement>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Nicolas</forename><surname>Lazzari</surname></persName>
							<email>nicolas.lazzari3@unibo.it</email>
							<affiliation key="aff2">
								<orgName type="institution">University of Bologna</orgName>
							</affiliation>
							<affiliation key="aff7">
								<orgName type="department">Team &quot;Mordor&quot; at the International Semantic Web Summer School</orgName>
								<address>
									<postCode>2023</postCode>
									<settlement>Bertinoro</settlement>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Anouk</forename><forename type="middle">M</forename><surname>Oudshoorn</surname></persName>
							<email>anouk.oudshoorn@tuwien.ac.at</email>
							<affiliation key="aff3">
								<orgName type="institution">Technical University of Vienna</orgName>
							</affiliation>
							<affiliation key="aff7">
								<orgName type="department">Team &quot;Mordor&quot; at the International Semantic Web Summer School</orgName>
								<address>
									<postCode>2023</postCode>
									<settlement>Bertinoro</settlement>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Disha</forename><surname>Purohit</surname></persName>
							<email>d.purohit@stud.uni-hannover.de</email>
							<affiliation key="aff4">
								<orgName type="institution">Leibniz University Hannover</orgName>
							</affiliation>
							<affiliation key="aff7">
								<orgName type="department">Team &quot;Mordor&quot; at the International Semantic Web Summer School</orgName>
								<address>
									<postCode>2023</postCode>
									<settlement>Bertinoro</settlement>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Ensiyeh</forename><surname>Raoufi</surname></persName>
							<email>ensiyeh.raoufi@lirmm.fr</email>
							<affiliation key="aff5">
								<orgName type="institution" key="instit1">University of Montpellier</orgName>
								<orgName type="institution" key="instit2">LIRMM</orgName>
							</affiliation>
							<affiliation key="aff7">
								<orgName type="department">Team &quot;Mordor&quot; at the International Semantic Web Summer School</orgName>
								<address>
									<postCode>2023</postCode>
									<settlement>Bertinoro</settlement>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Sebastian</forename><surname>Rudolph</surname></persName>
							<email>sebastian.rudolph@tu-dresden.de</email>
							<affiliation key="aff6">
								<orgName type="institution">Technische Universität Dresden</orgName>
							</affiliation>
							<affiliation key="aff7">
								<orgName type="department">Team &quot;Mordor&quot; at the International Semantic Web Summer School</orgName>
								<address>
									<postCode>2023</postCode>
									<settlement>Bertinoro</settlement>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">A is the B of C: (Semi)-Automatic Creation of Vossian Antonomasias</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">2A2E774088208747B0F425ACFC698DA3</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T19:20+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>A Vossian Antonomasia (VA) is a stylistic device used to describe a person (or, more generally, an entity) in terms of a well-known person and a modifying context. For instance, the Norwegian chess world champion Magnus Carlsen was described as "the Mozart of chess" <ref type="bibr" target="#b0">[1]</ref>. All VAs follow the pattern where a source (e.g., "Mozart"), is used to describe a target, (e.g., "Magnus Carlsen"), and the transfer of meaning is "channeled" through the use of the modifier "of chess". Although this rhetorical figure is well-known, there has not yet been a dedicated study of targeted automatic or semi-automatic methods to generate and judge the appropriateness of VAs using large Knowledge Graphs (KGs) such as Wikidata. In our work, we propose the use of vector space embeddings -both KG-based and text-based -for producing VAs. For comparison, we contrast our findings with a purely LLM-based approach, wherein VAs are obtained from ChatGPT using a reasonably engineered prompt. We provide a publicly available GitHub repository 1 for the implementation of our method and a website 2 that allows testing the proposed methods.</p><p>1 https://github.com/MordorISWS23/antonomasia 2 The website is available at https://antonomasia.informatik.uni-bremen.de/. Note that for efficiency reasons, a restricted set of entities is available to users.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>The question of whether computational methods can be used as creative devices can be traced back to the beginning of computers when Ada Lovelace wondered about the endless possibilities of automatic calculators <ref type="bibr" target="#b1">[2]</ref>. Even though Artificial Intelligence techniques have largely been used in creative applications <ref type="bibr" target="#b2">[3]</ref>, the evaluation of such creative outputs remains problematic <ref type="bibr" target="#b3">[4]</ref>. In this work, we propose the generation of Vossian Antonomasias (VAs) as a benchmark for exploring the creativity of AI methods.</p><p>VAs are a popular stylistic device for describing one entity by referring to another, typically in a witty and resourceful manner. A VA consists of three parts: a target entity 𝐴, a source entity 𝐵, and a modifier 𝐶, and is generally expressed as 𝐴 is the 𝐵 of 𝐶.</p><p>A meaningful VA requires a non-trivial degree of creativity and extensive knowledge of the specifics of the target entity. One has to identify a set of salient characteristics of 𝐴 that is similarly, or even more prominently realised by 𝐵. It is fundamental, however, that 𝐴 and 𝐵 differ when compared using the modifier 𝐶. For instance, in the sentence "Nacho Figueras is the Brad Pitt of polo players", Ignacio Figueras, among the most famous polo players in the world, is compared to the actor Brad Pitt due to his appearance. 1  VAs are often used in many journalistic genres and frequently appear in headings, as they can be both informative, enigmatic, and entertaining. In general, 𝐵 is a well-known, widely recognized entity. Through its popularity, the writer encourages readers to classify 𝐴 as similar to 𝐵, despite the difference 𝐶.</p><p>In this work, we present a method to automatically generate VAs by exploiting the latent semantic capabilities of vector space embeddings. We extract potential candidates for 𝐵 by using SPARQL queries over Wikidata. We rely on a heuristic method to select entities that can be classified as popular. By relying on publicly available Knowledge Graph Embeddings (KGE) trained on Wikidata, we compute the vector representations for an arbitrary 𝐴, the identified set of 𝐵 and a restricted set of 𝐶. We experiment with different operations between vectors to select the best 𝐵 candidate.</p><p>In order to investigate the efficacy of each experiment, we compare the use of KGE with word embeddings obtained from large corpora of text <ref type="bibr" target="#b4">[5]</ref>. Additionally, given the recent surge of Large Language Models (LLM) to mine creative analogies <ref type="bibr" target="#b5">[6]</ref>, we rely on ChatGPT <ref type="bibr" target="#b6">[7]</ref> as an additional baseline. We evaluate each method through a user-evaluation study.</p><p>The paper can be summarised as:</p><p>1. Identification of a suitable pool of candidates that can serve as 𝐵 elements; 2. Proposal of a novel method to automatically generate Vossian Antonomasias; 3. Evaluation of the proposed method using a user evaluation study;</p><p>The paper is organised as follows: in Section 2 we describe related work, which is followed by the presentation of the implemented method in Section 3. In Section 4, we discuss the results produced by the methods of Section 3 and present the outcomes of the user evaluation in Section 5. We finish by drawing conclusions and providing an outlook in Section 6.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Related Work</head><p>There has been limited research on automatically detecting Vossian Antonomasias in written text. The authors of <ref type="bibr" target="#b0">[1]</ref> demonstrated that by using Wikidata, they were able to overcome the shortcomings of available Named Entity Recognition (NER) tools and confirmed that VA is a linguistic and cultural phenomenon. Through quantitative VA explorations, they were able to capture the phenomenon as a whole, encompassing the source, target, and, when available, modifier. Their approach involves searching for a network of individuals interconnected by diverse modifiers, where the nodes can function as either sources or targets. This network aids in understanding hidden patterns of role models, revealing how they vary across countries and languages. However, a limitation they acknowledged is their reliance on the most prevalent pattern of VA, namely, "the...of" which resulted in the omission of numerous expressions extracted from the New York Times (NYT) corpus. For instance, notable phrases such as "the American Oscar Wilde" and "Harlem's Mozart" were overlooked despite their significance. Another approach for the extraction of VAs is presented by <ref type="bibr" target="#b7">[8]</ref>. The focus is on the extraction of the target by using coreference resolution and visualising the connections between the source and target entities extracted in the VAs in form of a web demo. The authors of <ref type="bibr" target="#b8">[9]</ref> use neural networks for the end-to-end detection of VAs resulting in two models: one for binary sentence classification and another for sequence tagging of all parts of the VA on the word level.</p><p>As opposed to the work described above, our approach focuses on the generation of VAs rather than their detection. Similarly to the approach of <ref type="bibr" target="#b0">[1]</ref>, our method only focuses on the pattern "the...of" by exploiting the latent semantic space of Knowledge Graph Embeddings and word embedding methods. Knowledge Graph Embedding compute a vectorial approximation of the originating Knowledge Graph through the use of various geometrical intuitions. In TransE <ref type="bibr" target="#b9">[10]</ref> predicates and entities are modelled as translations in the vector space. Given a triple (ℎ, 𝑟, 𝑡), the vector embedding of head entity ℎ, predicate or relation 𝑟, and tail entity 𝑡 are computed to minimise the quantity |ℎ + 𝑟 − 𝑡| -i.e. 𝑡 should be close to the ℎ + 𝑟. Increasingly complex methods have been presented in literature <ref type="bibr" target="#b10">[11]</ref>.</p><p>Based on the distributional hypothesis, word embeddings <ref type="bibr" target="#b11">[12]</ref> are used to compute vectors based on the distribution of words in large corpora of text, such as GloVe <ref type="bibr" target="#b12">[13]</ref>, where the representation is obtained using word co-occurrence statistics or word2Vec <ref type="bibr" target="#b13">[14]</ref>, where words with similar contextual distribution are approximated as similar vectors.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Methodology</head><p>This section details our proposed approach to automatically generate VAs, as outlined in Section 1. Figure <ref type="figure" target="#fig_0">1</ref> provides a high-level summary of our approach.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1.">Wikidata as a Knowledge Resource</head><p>We rely on Wikidata to identify a set of candidates that can serve as 𝐵 entities. This allows us to benefit from a large number of triples <ref type="foot" target="#foot_0">2</ref>   which enables us to select a sufficient sample of candidates for every component of the Vossian Antonomasias. Additionally, Wikidata provides a more structured and consistent data model when compared to similar resources, such as DBPedia <ref type="bibr" target="#b15">[16]</ref>. Moreover, we can leverage the language-independent design of Wikidata as opposed to DBPedia <ref type="bibr" target="#b15">[16]</ref> to ensure that the targets are widely popular.</p><p>We first extract the entities that will be used as 𝐵 candidates by means of SPARQL queries. We retrieve popular fictional characters and popular humans with the query of Listing 1. Given the large number of triples retrieved by both queries, they are executed on the Semantic Builders<ref type="foot" target="#foot_1">3</ref> SPARQL endpoint rather than on the regular Wikidata endpoint <ref type="foot" target="#foot_2">4</ref> . This allows us to overcome the querying timeout imposed by Wikidata and extract 3815 entities. The extracted entities are used as the set of candidates for 𝐵 in a VA. We rely on a heuristic method to compute the popularity of an entity: the number of worldwide available Wikipedia articles in distinct languages for an entity as a proxy for its popularity. Given 𝑡 the number of translations of one entity, we found 𝑡 ≥ 70 for real-world individuals and 𝑡 &gt; 30 for fictional characters to be a good estimate.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.">VA Generation using Vector Representations</head><p>We generate VA sentences by using geometrical transformations on the latent space provided by vector embeddings. Given an arbitrary 𝐴, we constrain the modifier 𝐶 to be the occupation of the entity 𝐴. The underlying assumption of the proposed model is that, despite their different occupations, 𝐴 and 𝐵 need to be similar with respect to their salient features. For this reason, given a particular 𝐴, all those entities 𝑏 ∈ 𝐵 that share the same modifier (i.e. the same occupation) are excluded from the pool of candidates. This brings us closer to ensuring the accurate selection of 'B' in accordance with the conditions specified in section 1. ) the embedding vector of the predicate 𝐶 that denotes the occupation of an entity (i.e. P106 in Wikidata), the translationbased method first disregards (i.e., "subtracts") A's occupation C obtaining:</p><formula xml:id="formula_0">⃗ 𝑎 ′ = ⃗ 𝑎 − ⃗ 𝑐<label>(1)</label></formula><p>Then, we define the fitness 𝑓 (where smaller is better) of a candidate ⃗ 𝑏 ′ ∈ ℬ as the Euclidean distance between ⃗ 𝑎 ′ and ⃗ 𝑏 ′ , i.e.,</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>𝑓 (⃗ 𝑎, ⃗ 𝑏</head><formula xml:id="formula_1">′ ) = | ⃗ 𝑎 ′ − ⃗ 𝑏 ′ | = √ 𝑑 ∑ 𝑖=0 (𝑎 ′ 𝑖 − 𝑏 ′ 𝑖 ) 2 (2)</formula><p>where 𝑑 is the dimension of the vector embedding space. The projection-based method relies on a different assumption. Informally, we would like to compute the fitness 𝑓 (⃗ 𝑎, ⃗ 𝑏 ′ ) on a subspace of the whole embedding space where every information related to the occupation of the entities is ignored. We compute the projection of ⃗ 𝑎 and ⃗ 𝑏 ′ to this subspace, which is a hyperplane perpendicular to ⃗ 𝑐, using</p><formula xml:id="formula_2">𝜋 ⃗ 𝑐 ( ⃗ 𝑥) = ⃗ 𝑥 − ( ⃗ 𝑥 ∘ ⃗ 𝑐 ⃗ 𝑐 ∘ ⃗ 𝑐 ) ⃗ 𝑐<label>(3)</label></formula><p>where ∘ denotes the inner product and ⃗ 𝑥 is either ⃗ 𝑎 or ⃗ 𝑏 ′ . The fitness function 𝑓 is hence adjusted to the cosine distance between the projections of ⃗ 𝑎 and ⃗ 𝑏 ′ , formally   <ref type="figure" target="#fig_2">2a</ref>) and the projection-based method (Figure <ref type="figure" target="#fig_2">2b</ref>).</p><formula xml:id="formula_3">𝑓 (⃗ 𝑎, ⃗ 𝑏 ′ ) = 𝜋 ⃗ 𝑐 (⃗ 𝑎) ∘ 𝜋 ⃗ 𝑐 ( ⃗ 𝑏 ′ ) |𝜋 ⃗ 𝑐 (⃗ 𝑎)||𝜋 ⃗ 𝑐 ( ⃗ 𝑏 ′ )|<label>(4)</label></formula><p>The fitness functions in Equation ( <ref type="formula">2</ref>) and Equation ( <ref type="formula" target="#formula_3">4</ref>) can be seen as similarity functions between two entities. A suitable 𝐵 can be extracted by taking the entity that minimises such distance. Intuitively, in the translation-based approach, the vector representing 𝐵 is supposed to be similar to the one of 𝐴 after we "translate away" or "subtract" the characteristics pertaining to 𝐶 using Equation (1), while in projection-based method by applying Equation (3), we "project away" such characteristics.</p><p>As addressed in Section 1, a good VA needs to be a creative sentence. While it is difficult to assess creativity in an objective manner, it has been argued that among the many characteristics, a creative output needs to display novelty when compared to others <ref type="bibr" target="#b16">[17]</ref>. Given a set of entities B that minimises the fitness function 𝑓, we propose to further rank such entities by using their 𝐿 1 norm. The intuition is that among all the candidates in B , the ones with a greater distance from the origin of the vector space are the most "extremal" ones.</p><p>Figure <ref type="figure" target="#fig_2">2</ref> depicts a simplified illustration of Equation 2 and Equation 4. Using t-SNE <ref type="bibr" target="#b17">[18]</ref>, we reduce the dimensionality of embedding vector ⃗ 𝑎 of a sample entity A, a set ℬ of embedding vectors of each candidate entity for 𝐵, and ⃗ 𝑏 ∈ B vectors.</p><p>Given a particular entity 𝐴 and the corresponding entity 𝐵, selected with one of the proposed methods, we generate an assertional VA following template: {A} {verb} the {B} of {C}. If the entity 𝐴 has an entry in Wikidata that certifies its death, we set verb to was, else we use is.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.3.">Purely LMM-based Baseline via ChatGPT</head><p>Finally, we use ChatGPT <ref type="bibr" target="#b6">[7]</ref>, a Large Language Model, as a baseline for VA generation. Through extensive experimentation, we found the prompt that obtains the best result to be Following the discussion of Section 1, we argue that an effective prompt for VA needs to display the following properties:</p><p>• 𝐵 should not share characteristics with 𝐶;</p><p>• 𝐴 and 𝐵 should share at least one salient characteristic; • 𝐴 and 𝐵 should be popular enough to draw the analogy in the context of 𝐶 Provide 10 Vossian Antonomasias for &lt;Name of A&gt;, where she is equated with another person. Each of the phrases should have the structure "&lt;Name of A&gt; is the [person name] of [profession]", where [profession] must not characterize [person name]. Provide a very short justification for each example.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Table 1</head><p>ChatGPT prompt used to obtain a VA.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Experimental Setting</head><p>As briefly addressed in Section 1, we experiment with two different methods to obtain vector representations of an entity: Knowledge Graph Embeddings (KGE) and Word Embeddings (WE).</p><p>We employ TransE <ref type="bibr" target="#b9">[10]</ref> as KGE method. We directly reuse the publicly available model shared by GraphVite <ref type="bibr" target="#b18">[19]</ref> and trained on the Wikidata-5M dataset <ref type="bibr" target="#b19">[20]</ref>. For the WE method, we employ word2vec <ref type="bibr" target="#b13">[14]</ref> and GloVe <ref type="bibr" target="#b12">[13]</ref> provided by gensim <ref type="bibr" target="#b4">[5]</ref>.</p><p>Finally, we leverage the use of meta-embedding techniques, i.e. combining different embedding methods together <ref type="bibr" target="#b20">[21]</ref>, to exploit the main advantages of both methods. We combine KGE and WE by means of concatenation and averaging. When averaging two vectors with different dimensionality, we apply zero-padding <ref type="bibr" target="#b20">[21]</ref>. Note that, even though they are supposed to converge to a similar semantic, the latent space represented by a method might differ drastically from other methods. To prevent a drastically higher influence of one method over the other, we normalise both methods by their 𝐿 2 norm before combining them.</p><p>To allow interested readers to try out the different methods themselves, we set up a demonstrator website available at https://antonomasia.informatik.uni-bremen.de/.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Results and Evaluation</head><p>Due to the highly diverse nature of VAs, we decided to test the quality of the output with human evaluation. A small selection of examples generated by the methods described in Section 3 is presented in Table <ref type="table">2</ref>.</p><p>The selection on Table <ref type="table">2</ref> shows that ChatGPT, while being very creative when it comes to the description of the domain, does not perform well in identifying a proper 𝐵 that does not share characteristics with the modifier 𝐶, such as in the sentence "Bill Gates is the Einstein of Societal Transformation". This phenomenon particularly occurs with politicians or writers. Despite the explicit request in the prompt of Table <ref type="table">1</ref>, ChatGPT did not manage to adapt the chosen entity. The results generated by the KGE, WE and their combination mostly meet the mentioned criteria, even though some exceptions occur, such as in the sentence "Angela Merkel is the Eva Braun of politics. ".</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.1.">User Evaluation</head><p>The method described in Section 3 allows combining several different techniques to generate a VA. For the user evaluation, after manual experimentation, we restricted the set of techniques to the ones described in Table <ref type="table" target="#tab_3">3</ref>. The presented selection allows us to evaluate the importance of different assumptions, such as whether the methods based on the distributional hypothesis can complement content-based methods. Moreover, we are able to assess whether the presented methods can overcome the issues of ChatGPT, namely the difficulty of selecting 𝐴 and 𝐵 from a different domain dictated by 𝐶. We identify six individuals that will be used as 𝐴: Nelson Mandela, Angela Merkel, Mark Twain, Albert Einstein, Bill Gates and Ronald Reagan. Those individuals are both part of the entities extracted using the query in Listing 1 and the real-world samples from the New York Times <ref type="bibr" target="#b0">[1]</ref> and Der Umblätterer<ref type="foot" target="#foot_3">5</ref> corpora. As mentioned in Section 3, we contain 𝐶 to the profession of the entity.</p><p>To recruit participants for our study, we distributed a flyer as advertisement and shared it with colleagues and friends who themselves distributed this further. The study can be done online without any supervision. Due to the specificity of Vossian Antonomasia, we present a definition on the front page of the study:</p><p>Vossian antonomasias refer to someone by a special characteristic instead of their name.For example, calling Bill Gates "the Henry Ford of the computer age" highlights his influence as entrepeneur and his effect on the development of technology. It is a way to describe someone by an important quality they possess.</p><p>The study was open for one week. We provide each participant with 21 sentences. We randomly sample 3 𝐴s from the set described above and provide 7 VAs, one for every generation method of Table <ref type="table" target="#tab_3">3</ref>. The selected VAs are randomly sampled from the top 10 sentences identified by the method used. The participant is asked to judge each VA on three aspects: how well the description fits, how understandable it is and how original the VA is. These three aspects can be ranked on a Likert Scale ranging form 1 to 5. After that, the knowledge about the source and the target will be inquired in the form of questions: we ask how well the participant knows the individuals. The possible answers are I know who that person is, I have heard of the name but I cannot relate it to anything, and I have never heard of that name before. This ensures a distinction between a negative rating caused by ignorance of the output's components and the lack of a proper connection between the source, the target and the modifier.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.2.">Results</head><p>Through the user-evaluation test, we obtain a set of 207 human evaluations on automatically generated VAs, provided by 29 unique annotators. The sentences presented to the participants   <ref type="table" target="#tab_3">3</ref> with 1 being the best and 5 being the worst rating.</p><p>are repeated to avoid a random evaluation. The inter-annotator agreement, computed using Cohen's Kappa score <ref type="bibr" target="#b21">[22]</ref>, is 0.0491 on average. Such a low score highlights the difficulty in evaluating VAs since they greatly depend on the reader's knowledge, cultural reference and degree of familiarity with the selected subject.</p><p>In Figure <ref type="figure" target="#fig_5">5</ref> the distribution of ratings among the methods reported in Table <ref type="table">5</ref> is reported. Intuitively, whose distribution is skewed towards low ratings (represented in green) should be considered the best-performing method. Using the translation-based meta-embedding method outperforms all the other methods. Interestingly, the baseline provided by ChatGPT performs worse than any other method. While this might be reconducted to the lack of explicit VA-related knowledge that the underlying LLM is based on, it can also be argued that the prompt that we propose to use is not perfectly suited for this task. We will further address this aspect in Section 6. At first glance, the best method turns out to be word2vec using the translation technique, which results in 74 VAs rated with a score of 1. It needs to be argued, however, that using this method results also in a high variance in the results. Indeed, the same method is also the one that obtains the highest amount of low-rated VAs. When taking into account the whole distribution, the projection-based method TransE achieved the best performance: the lowest number of "bad" VAs is obtained with such a method. Indeed, this is the method that achieves the best overall performances, as can be seen in Table <ref type="table" target="#tab_4">4</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Table 5</head><p>Overview of the mean rating for each user question. The mean rating for each confidence rating (𝜇 Table <ref type="table">5</ref> provides an overview of the mean rating 𝜇 𝑖 for each method, where 𝑖 represents the set of sentences for which an evaluator expressed specific confidence in the knowledge of 𝐴 and 𝐵. Those results are complementary to the ones of Figure <ref type="figure" target="#fig_5">5</ref>. Interestingly, the best overall method, projection-based TransE, does not classify as the best in any specific user question. Depending on the target task, one model can be considered better than the other, even though projection-based TransE guarantees the most consistent results.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.">Conclusions and Discussion</head><p>We looked at generating Vossian Antononmasias by using embeddings and LLM as a way to characterize the creativity of AI. Our approach has resulted in creative examples of VAs, which proves that both methods are suitable for solving such tasks. Since the lack of a clear definition of creativity prevents a quantitative evaluation of the results, we conducted a manual qualitative analysis of the results which highlighted several different weaknesses. The information bias that is inherent to Wikidata results in VAs that are mostly focused on Western culture. While this might be tampered by penalising some entities, we argue it would only partially solve the issue. A different approach, which takes into account the semantic representation of each entity, can help overcome such issues, making the creation process transparent and explainable.</p><p>The human evaluation described in Section 5 meaningful insights into the effectiveness of our methods. Firstly, they generally perform better than the ChatGPT baseline, which fails in the generation of original and understandable VAs. Moreover, the results of Figure <ref type="figure" target="#fig_5">5</ref> and Table <ref type="table">5</ref> show how the use of Knowledge Graph Embeddings is generally to be preferred over the methods, such as word embeddings and meta-embeddings. However, the low inter-annotator agreement shows that rating Vossian Antonomasias is highly subjective and most probably depends on not only the knowledge of an entity but also knowledge of the domain the entity refers to. This could be addressed by filtering the human annotators into groups according to their domain knowledge before rating the sentences.</p><p>The focus on the occupation as a similarity measure resulted in several complications, such as the use of semantically similar occupations like television actor and actor in the generated sentences. Additionally, since some entities hold multiple occupations, a more accurate estimation of their primary occupation needs to be investigated. A possible solution is to aggregate the representation vectors of all their occupations instead of selecting a single occupation. Similarly, fictional characters are sometimes compared to their real-life actors. A possible solution is to impose a minimum distance between vectors that are too close. An orthogonal solution is to consider other criteria when comparing entities, such as achievement or awards. Apart from famous people, famous locations or events along with their appropriate modifiers could be added to increase the sample size and achieve a greater variability in the results.</p><p>The mentioned limitations of the evaluation and sampling of entities show that the generation of Vossian Antonomasia with an open-domain approach proves to be rather difficult. Instead, we suggest focusing on specific domains, thereby using fine-tuning of the embeddings or different embedding methods to overcome the mentioned shortcomings.</p><p>Additionally, we envision integrating an LLM in the pre-evaluation step by evaluating the VAs that have been generated by combining our proposed methods. The idea is to list the salient similarities and differences between 𝐴 and 𝐵 for a given VA by looking at their characteristic properties. Following recent LLM-prompting studies <ref type="bibr" target="#b22">[23,</ref><ref type="bibr" target="#b23">24]</ref>, we plan on performing additional manual or automatic prompt engineering <ref type="bibr" target="#b24">[25,</ref><ref type="bibr" target="#b25">26]</ref>. This can lead to more effective results, since any change in the prompts may affect significantly the quality of the output.</p><p>Moreover, an interesting approach is to perform knowledge injection <ref type="bibr" target="#b26">[27,</ref><ref type="bibr" target="#b27">28]</ref> into a LLM, following a neuro-symbolic approach to use the language model's potential and control for the criteria defining Vossian Antonomasia. The knowledge available to an LLM like ChatGPT is currently limited regarding whether an entity has recently died or is a fictional character based on the time it was trained. The injection of structured knowledge can help overcoming this issue.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: Vossian Antonomasia (VA) methodology: the approach we followed in our research work.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head></head><label></label><figDesc>(a) Equation 2 (translationbased) illustrated. (b) Equation 4 (projectionbased) illustrated.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>Figure 2 :</head><label>2</label><figDesc>Figure 2: Illustration of the fitness function 𝑓 for the translation-based method (Figure2a) and the projection-based method (Figure2b).</figDesc><graphic coords="6,171.38,109.36,125.00,99.83" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_3"><head>Figure 3 : 2</head><label>32</label><figDesc>Figure 3: Screenshot of the demonstrator website</figDesc><graphic coords="8,198.80,84.19,197.67,301.95" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_4"><head>Figure 4 :</head><label>4</label><figDesc>Figure 4: Frontend of the user study for rating the VAs.</figDesc><graphic coords="10,151.76,84.19,291.75,161.25" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_5"><head>Figure 5 :</head><label>5</label><figDesc>Figure 5: Distribution of the mean rating among the methods of Table3with 1 being the best and 5 being the worst rating.</figDesc><graphic coords="10,127.66,284.72,339.96,155.28" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>with broad coverage of encyclopedic knowledge, Evaluate Extract Use Pre-Trained Models</head><label></label><figDesc></figDesc><table><row><cell></cell><cell cols="2">Embeddings</cell><cell>Exploration</cell></row><row><cell></cell><cell>Knowledge Graph</cell><cell>Text-based</cell></row><row><cell></cell><cell>Embeddings (KGE)</cell><cell>Embeddings</cell></row><row><cell></cell><cell cols="2">Mapping B's into</cell><cell>Applying similarity</cell></row><row><cell>Extracting B's from</cell><cell cols="2">embedding space</cell><cell>measures to identify</cell></row><row><cell>Wikidata using</cell><cell></cell><cell></cell><cell>appropriate B</cell></row><row><cell>SPARQL queries</cell><cell></cell><cell></cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>'s Target (A) is the source (B) of modifier (C)</head><label></label><figDesc></figDesc><table /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head></head><label></label><figDesc>PREFIX wdt: &lt;http://www.wikidata.org/prop/direct/&gt; PREFIX wd: &lt;http://www.wikidata.org/entity/&gt; SELECT ?item ?itemLabel ?occupation ?sitelinks WHERE { SPARQL Query to extract candidate entities for 𝐵. &lt;type&gt; and &lt;threhsold&gt; are replaced by wd:Q15632617 and 30 for fictional characters and wd:Q5 and 70 for humans.We propose two different methods for the selection of the best 𝐵 candidate: a translation-based approach and a projection-based approach. The translation-based approach follows the intuition of TransE and word2vec: given ⃗ 𝑎 1 , … , 𝑎 𝑑 ) the embedding vector of an arbitrary target entity A, ℬ the set of embedding vectors of each candidate entity for 𝐵, and ⃗ 𝑐 = (𝑐 1 , … , 𝑐 𝑑</figDesc><table><row><cell>?item wdt:P31 &lt;type&gt;;</cell></row><row><cell>wdt:P106 ?occupation;</cell></row><row><cell>wikibase:sitelinks ?sitelinks .</cell></row><row><cell>FILTER(&lt;threshold&gt; &lt; ?sitelinks).</cell></row><row><cell>SERVICE wikibase:label {bd:serviceParam wikibase:language</cell></row><row><cell>"[AUTO_LANGUAGE],en".}}</cell></row><row><cell>𝑥, ⃗ 𝑦 the vector representations of two arbitrary entities and</cell></row><row><cell>given ⃗ 𝑐 the vector that represents a predicate holding between ⃗ 𝑥 and ⃗ 𝑦, it has been observed</cell></row><row><cell>that ⃗ 𝑥 + ⃗ 𝑐 ≈ ⃗ 𝑦.</cell></row><row><cell>Given ⃗ 𝑎 = (</cell></row></table><note>↪Listing 1:</note></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_3"><head>Table 3</head><label>3</label><figDesc>Methods employed in the user evaluation study. The meta embedding obtained by concatenating TransE and word2vec is written as TransE ⊕ word2vec.</figDesc><table><row><cell>Method</cell><cell>Fitness function</cell></row><row><cell>ChatGPT</cell><cell></cell></row><row><cell cols="2">TransE ⊕ word2vec Project (Equation (4))</cell></row><row><cell cols="2">TransE ⊕ word2vec Translate (Equation (2)</cell></row><row><cell>TransE</cell><cell>Translate (Equation (2))</cell></row><row><cell>TransE</cell><cell>Project (Equation (4))</cell></row><row><cell>word2vec</cell><cell>Translate (Equation (2))</cell></row><row><cell>word2vec</cell><cell>Project (Equation (4))</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_4"><head>Table 4</head><label>4</label><figDesc>Average rating for each method. Lower is better. Best result is represented in bold.</figDesc><table><row><cell></cell><cell>Method</cell><cell>Average rating</cell></row><row><cell></cell><cell>ChatGPT</cell><cell>2.69</cell></row><row><cell>Project</cell><cell cols="2">TransE ⊕ word2vec 2.64 TransE 2.71 word2vec 2.66</cell></row><row><cell>Translate</cell><cell cols="2">TransE ⊕ word2vec 2.57 TransE 2.70 word2vec 2.59</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_5"><head></head><label></label><figDesc>1 , 𝜇 2 , and 𝜇 3 ) is reported alongside the overall mean rate 𝜇 1∪2∪3 for each question. The best results for each criterion are highlighted in bold. Lower values indicate better performance.</figDesc><table><row><cell></cell><cell>Method</cell><cell></cell><cell>Fit</cell><cell></cell><cell></cell><cell></cell><cell cols="2">Understand</cell><cell></cell><cell></cell><cell cols="2">Original</cell></row><row><cell></cell><cell></cell><cell>𝜇 1</cell><cell>𝜇 2</cell><cell>𝜇 3</cell><cell>𝜇</cell><cell>𝜇 1</cell><cell>𝜇 2</cell><cell>𝜇 3</cell><cell>𝜇</cell><cell>𝜇 1</cell><cell>𝜇 2</cell><cell>𝜇 3</cell><cell>𝜇</cell></row><row><cell></cell><cell>ChatGPT</cell><cell cols="3">2.21 2.09 2.81</cell><cell cols="4">2.51 2.42 2.27 3.22</cell><cell cols="4">2.80 2.63 3.27 2.73</cell><cell>2.77</cell></row><row><cell>Project</cell><cell cols="12">TransE ⊕ word2vec 2.26 2.07 2.67 TransE 1.93 2.38 2.67 word2vec 1.88 2.08 2.63 2.30 2.00 2.62 3.34 2.42 2.16 2.14 3.15 2.64 2.47 3.14 2.97 2.45 2.14 2.00 3.41 2.84 2.50 2.46 3.15 2.83 2.50 2.77 3.08</cell><cell>2.88 2.84 2.86</cell></row><row><cell>Translate</cell><cell cols="4">TransE ⊕ word2vec 2.17 2.60 2.33 TransE 1.70 2.71 2.55 word2vec 1.89 2.07 3.00</cell><cell cols="4">2.35 2.22 2.87 2.89 2.38 1.85 3.29 3.23 2.48 2.11 1.93 3.31</cell><cell cols="4">2.71 2.39 2.67 2.81 2.87 2.40 3.00 3.10 2.65 2.11 2.93 2.88 2.65 2.67 2.87</cell></row></table></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_0">According to Wikidata's statistics the knowledge graph currently contains 104,204,236 items.<ref type="bibr" target="#b14">[15]</ref> </note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_1">https://semantic.builders/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="4" xml:id="foot_2">https://query.wikidata.org/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="5" xml:id="foot_3">https://www.umblaetterer.de/datenzentrum/vossianische-antonomasien.html</note>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgements</head><p>The project leading to this application has received funding from the European Union's Horizon 2020 research and innovation programme under grant agreement No 101034440. This work was partially funded by the Klaus Tschira Foundation, grant number 40300928 and the French National Research Agency ANR DACE-DL project, grant number ANR-21-CE23-0019.</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">The Michael Jordan of Greatness&apos;-Extracting Vossian Antonomasia from Two Decades of The New York Times</title>
		<author>
			<persName><forename type="first">F</forename><surname>Fischer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Jäschke</surname></persName>
		</author>
		<idno type="DOI">10.1093/llc/fqy087</idno>
		<ptr target="https://doi.org/10.1093/llc/fqy087.doi:10.1093/llc/fqy087" />
	</analytic>
	<monogr>
		<title level="m">Digital Scholarship in the Humanities</title>
				<imprint>
			<date type="published" when="1987">1987-2007. 2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Sketch of the Analytical Engine invented by Charles Babbage</title>
		<author>
			<persName><forename type="first">L</forename><forename type="middle">F</forename><surname>Menabrea</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Ada&apos;s Legacy: Cultures of Computing from the Victorian to the Digital Age</title>
				<meeting><address><addrLine>Esq.</addrLine></address></meeting>
		<imprint>
			<date type="published" when="1843">1843</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Artificial Intelligence in the Creative Industries: A Review</title>
		<author>
			<persName><forename type="first">N</forename><surname>Anantrasirichai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">R</forename><surname>Bull</surname></persName>
		</author>
		<idno type="DOI">10.1007/s10462-021-10039-7</idno>
		<ptr target="https://doi.org/10.1007/s10462-021-10039-7.doi:10.1007/s10462-021-10039-7" />
	</analytic>
	<monogr>
		<title level="j">Artif. Intell. Rev</title>
		<imprint>
			<biblScope unit="volume">55</biblScope>
			<biblScope unit="page" from="589" to="656" />
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Redefining Creativity in the Era of AI? Perspectives of Computer Scientists and New Media Artists</title>
		<author>
			<persName><forename type="first">R</forename><surname>Wingström</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Hautala</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Lundman</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Creativity Research Journal</title>
		<imprint>
			<biblScope unit="page" from="1" to="17" />
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Software Framework for Topic Modelling with Large Corpora</title>
		<author>
			<persName><forename type="first">R</forename><surname>Řehůřek</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Sojka</surname></persName>
		</author>
		<ptr target="http://is.muni.cz/publication/884893/en" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the LREC 2010 Workshop on New Challenges for NLP Frameworks</title>
				<meeting>the LREC 2010 Workshop on New Challenges for NLP Frameworks<address><addrLine>Valletta, Malta</addrLine></address></meeting>
		<imprint>
			<publisher>ELRA</publisher>
			<date type="published" when="2010">2010</date>
			<biblScope unit="page" from="45" to="50" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Cam: A Large Language Model-based Creative Analogy Mining Framework</title>
		<author>
			<persName><forename type="first">B</forename><surname>Bhavya</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Xiong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Zhai</surname></persName>
		</author>
		<idno type="DOI">10.1145/3543507.3587431</idno>
		<idno>doi:10.1145/3543507.3587431</idno>
		<ptr target="https://doi.org/10.1145/3543507.3587431" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the ACM Web Conference 2023</title>
				<meeting>the ACM Web Conference 2023<address><addrLine>New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>Association for Computing Machinery</publisher>
			<date type="published" when="2023">2023</date>
			<biblScope unit="page" from="3903" to="3914" />
		</imprint>
	</monogr>
	<note>WWW &apos;23</note>
</biblStruct>

<biblStruct xml:id="b6">
	<monogr>
		<title/>
		<author>
			<persName><forename type="first">Chatgpt</forename><surname>Openai</surname></persName>
		</author>
		<ptr target="VersionGPT-3.5" />
		<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Who is the Madonna of Italian-American Literature?&quot;: Extracting and Analyzing Target Entities of Vossian Antonomasia</title>
		<author>
			<persName><forename type="first">M</forename><surname>Schwab</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Jäschke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Fischer</surname></persName>
		</author>
		<ptr target="https://sighum.files.wordpress.com/2023/03/latech-clfl-2023-unofficial-proceedings.pdf" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 7th Joint SIGHUM Workshop on Computational Linguistics for Cultural Heritage, Social Sciences, Humanities and Literature</title>
				<meeting>the 7th Joint SIGHUM Workshop on Computational Linguistics for Cultural Heritage, Social Sciences, Humanities and Literature</meeting>
		<imprint>
			<publisher>Association for Computational Linguistics</publisher>
			<date type="published" when="2023">2023</date>
			<biblScope unit="page" from="110" to="115" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">The Rodney Dangerfield of Stylistic Devices&quot;: End-to-End Detection and Extraction of Vossian Antonomasia Using Neural Networks</title>
		<author>
			<persName><forename type="first">M</forename><surname>Schwab</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Jäschke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Fischer</surname></persName>
		</author>
		<idno type="DOI">10.3389/frai.2022.868249</idno>
		<ptr target="https://doi.org/10.3389/frai.2022.868249.doi:10.3389/frai.2022.868249" />
	</analytic>
	<monogr>
		<title level="j">Frontiers in Artificial Intelligence</title>
		<imprint>
			<biblScope unit="volume">5</biblScope>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Translating Embeddings for Modeling Multi-relational Data</title>
		<author>
			<persName><forename type="first">A</forename><surname>Bordes</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Usunier</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>García-Durán</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Weston</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Yakhnenko</surname></persName>
		</author>
		<ptr target="https://proceedings.neurips.cc/paper/2013/hash/1cecc7a77928ca8133fa24680a88d2f9-Abstract.html" />
	</analytic>
	<monogr>
		<title level="m">Advances in Neural Information Processing Systems 26: 27th Annual Conference on Neural Information Processing Systems 2013. Proceedings of a meeting held</title>
				<editor>
			<persName><forename type="first">C</forename><forename type="middle">J C</forename><surname>Burges</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">L</forename><surname>Bottou</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">Z</forename><surname>Ghahramani</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">K</forename><forename type="middle">Q</forename><surname>Weinberger</surname></persName>
		</editor>
		<meeting><address><addrLine>Lake Tahoe, Nevada, United States</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2013">December 5-8, 2013. 2013</date>
			<biblScope unit="page" from="2787" to="2795" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<monogr>
		<title level="m" type="main">A Survey of Knowledge Graph Embedding and Their Applications</title>
		<author>
			<persName><forename type="first">S</forename><surname>Choudhary</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Luthra</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Mittal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Singh</surname></persName>
		</author>
		<idno>CoRR abs/2107.07842</idno>
		<ptr target="https://arxiv.org/abs/2107.07842.arXiv:2107.07842" />
		<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<monogr>
		<author>
			<persName><forename type="first">F</forename><surname>Almeida</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Xexéo</surname></persName>
		</author>
		<idno>CoRR abs/1901.09069</idno>
		<ptr target="http://arxiv.org/abs/1901.09069.arXiv:1901.09069" />
		<title level="m">Word embeddings: A survey</title>
				<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Glove: Global Vectors for Word Representation</title>
		<author>
			<persName><forename type="first">J</forename><surname>Pennington</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Socher</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">D</forename><surname>Manning</surname></persName>
		</author>
		<idno type="DOI">10.3115/v1/d14-1162</idno>
		<ptr target="https://doi.org/10.3115/v1/d14-1162.doi:10.3115/v1/d14-1162" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing, EMNLP 2014</title>
				<editor>
			<persName><forename type="first">A</forename><surname>Moschitti</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">B</forename><surname>Pang</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">W</forename><surname>Daelemans</surname></persName>
		</editor>
		<meeting>the 2014 Conference on Empirical Methods in Natural Language Processing, EMNLP 2014<address><addrLine>Doha, Qatar</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2014">October 25-29, 2014. 2014</date>
			<biblScope unit="page" from="1532" to="1543" />
		</imprint>
	</monogr>
	<note>, A meeting of SIGDAT, a Special Interest Group of the ACL, ACL</note>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">Efficient Estimation of Word Representations in Vector Space</title>
		<author>
			<persName><forename type="first">T</forename><surname>Mikolov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Corrado</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Dean</surname></persName>
		</author>
		<ptr target="http://arxiv.org/abs/1301.3781" />
	</analytic>
	<monogr>
		<title level="m">1st International Conference on Learning Representations, ICLR 2013</title>
				<editor>
			<persName><forename type="first">Y</forename><surname>Bengio</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">Y</forename><surname>Lecun</surname></persName>
		</editor>
		<meeting><address><addrLine>Scottsdale, Arizona, USA</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2013">May 2-4, 2013. 2013</date>
		</imprint>
	</monogr>
	<note>Workshop Track Proceedings</note>
</biblStruct>

<biblStruct xml:id="b14">
	<monogr>
		<ptr target="https://www.wikidata.org/wiki/Wikidata:Statistics" />
		<title level="m">Wikidata statistics</title>
				<imprint>
			<date type="published" when="2023-06-16">2023. 2023-06-16</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">Wikidata and DBpedia: A Comparative Study</title>
		<author>
			<persName><forename type="first">D</forename><surname>Abián</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Guerra</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Martínez-Romanos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">T</forename><surname>Lado</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">International KEYSTONE Conference</title>
				<imprint>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">Assessing Creativity</title>
		<author>
			<persName><forename type="first">G</forename><surname>Ritchie</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. of AISB&apos;01 Symposium</title>
				<meeting>of AISB&apos;01 Symposium</meeting>
		<imprint>
			<date type="published" when="2001">2001</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">Visualizing Data using t-SNE</title>
		<author>
			<persName><forename type="first">L</forename><surname>Van Der Maaten</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Hinton</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Machine Learning Research</title>
		<imprint>
			<biblScope unit="volume">9</biblScope>
			<date type="published" when="2008">2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<analytic>
		<title level="a" type="main">Graphvite: A High-Performance CPU-GPU Hybrid System for Node Embedding</title>
		<author>
			<persName><forename type="first">Z</forename><surname>Zhu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Xu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Qu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Tang</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">The World Wide Web Conference</title>
				<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2019">2019</date>
			<biblScope unit="page" from="2494" to="2504" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<analytic>
		<title level="a" type="main">KEPLER: A Unified Model for Knowledge Embedding and Pre-trained Language Representation</title>
		<author>
			<persName><forename type="first">X</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Gao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Zhu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Tang</surname></persName>
		</author>
		<idno type="DOI">10.1162/tacl_a_00360</idno>
		<ptr target="https://doi.org/10.1162/tacl_a_00360.doi:10.1162/tacl\_a\_00360" />
	</analytic>
	<monogr>
		<title level="j">Trans. Assoc. Comput. Linguistics</title>
		<imprint>
			<biblScope unit="volume">9</biblScope>
			<biblScope unit="page" from="176" to="194" />
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<analytic>
		<title level="a" type="main">A Survey on Word Meta-Embedding Learning</title>
		<author>
			<persName><forename type="first">D</forename><surname>Bollegala</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>O'neill</surname></persName>
		</author>
		<idno type="DOI">10.24963/ijcai.2022/758</idno>
		<ptr target="https://doi.org/10.24963/ijcai.2022/758.doi:10.24963/ijcai.2022/758" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Thirty-First International Joint Conference on Artificial Intelligence, IJCAI 2022</title>
				<editor>
			<persName><forename type="first">L</forename><forename type="middle">D</forename><surname>Raedt</surname></persName>
		</editor>
		<meeting>the Thirty-First International Joint Conference on Artificial Intelligence, IJCAI 2022<address><addrLine>Vienna, Austria</addrLine></address></meeting>
		<imprint>
			<publisher>ijcai</publisher>
			<date type="published" when="2022-07-29">23-29 July 2022. 2022</date>
			<biblScope unit="page" from="5402" to="5409" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b21">
	<analytic>
		<title level="a" type="main">A Coefficient of AgI see you&apos;re on the paper, I updated the table and will proofread everything now, do you think that are particulreement for Nominal Scales</title>
		<author>
			<persName><forename type="first">J</forename><surname>Cohen</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Educational and Psychological Measurement</title>
		<imprint>
			<biblScope unit="volume">20</biblScope>
			<biblScope unit="page" from="37" to="46" />
			<date type="published" when="1960">1960</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b22">
	<monogr>
		<author>
			<persName><forename type="first">X</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Zheng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Du</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Ding</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Qian</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Yang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Tang</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2103.10385</idno>
		<title level="m">GPT Understands, Too</title>
				<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b23">
	<monogr>
		<author>
			<persName><forename type="first">J</forename><surname>White</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><surname>Fu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Hays</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Sandborn</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Olea</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Gilbert</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Elnashar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Spencer-Smith</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">C</forename><surname>Schmidt</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2302.11382</idno>
		<title level="m">A Prompt Pattern Catalog to Enhance Prompt Engineering with ChatGPT</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b24">
	<analytic>
		<title level="a" type="main">Prompt Programming for Large Language Models: Beyond the Few-Shot Paradigm</title>
		<author>
			<persName><forename type="first">L</forename><surname>Reynolds</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Mcdonell</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Extended Abstracts of the 2021 CHI Conference on Human Factors in Computing Systems</title>
				<imprint>
			<date type="published" when="2021">2021</date>
			<biblScope unit="page" from="1" to="7" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b25">
	<monogr>
		<author>
			<persName><forename type="first">Y</forename><surname>Zhou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">I</forename><surname>Muresanu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Han</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Paster</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Pitis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Chan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Ba</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2211.01910</idno>
		<title level="m">Large language Models Are Human-Level Prompt Engineers</title>
				<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b26">
	<monogr>
		<title level="m" type="main">Unifying Large Language Models and Knowledge Graphs: A Roadmap</title>
		<author>
			<persName><forename type="first">S</forename><surname>Pan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Luo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Wu</surname></persName>
		</author>
		<idno type="DOI">10.48550/ARXIV.2306.08302</idno>
		<ptr target="https://arxiv.org/abs/2306.08302.doi:10.48550/ARXIV.2306.08302" />
		<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b27">
	<monogr>
		<title level="m" type="main">ChatGPT is not Enough: Enhancing Large Language Models with Knowledge Graphs for Fact-aware Language Modeling</title>
		<author>
			<persName><forename type="first">L</forename><surname>Yang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Ding</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Wu</surname></persName>
		</author>
		<idno type="DOI">10.48550/ARXIV.2306.11489</idno>
		<ptr target="https://arxiv.org/abs/2306.11489.doi:10.48550/ARXIV.2306.11489" />
		<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
