<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Navigating Nulls, Numbers and Numerous Entities: Robust Knowledge Base Construction from Large Language Models</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Arunav</forename><surname>Das</surname></persName>
							<email>arunav.das@kcl.ac.uk</email>
							<affiliation key="aff0">
								<orgName type="institution">King&apos;s College London</orgName>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Nadeen</forename><surname>Fathallah</surname></persName>
							<email>nadeen.fathallah@ki.uni-stuttgart.de</email>
							<affiliation key="aff1">
								<orgName type="department" key="dep1">Analytic Computing</orgName>
								<orgName type="department" key="dep2">Institute for Artificial Intelligence</orgName>
								<orgName type="institution">University of Stuttgart</orgName>
								<address>
									<country key="DE">Germany</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Nicole</forename><surname>Obretincheva</surname></persName>
							<email>nicole.obretincheva@kcl.ac.uk</email>
							<affiliation key="aff0">
								<orgName type="institution">King&apos;s College London</orgName>
							</affiliation>
						</author>
						<title level="a" type="main">Navigating Nulls, Numbers and Numerous Entities: Robust Knowledge Base Construction from Large Language Models</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">922EC871310106B0ABD6409E820CC022</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T19:21+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>In this work, we employ advanced prompt engineering techniques with pre-trained Large Language Models (LLMs) to enhance knowledge extraction and structuring for Knowledge Base Construction (KBC) tasks. At the core of our methodology is the strategic fusion of different prompting strategies tailored to the unique relations between subject and object entities. We choose different prompting strategies to overcome the challenges of null values, numeric data, and one-to-many relationships inherent in traditional KBC tasks. By integrating role-play and context-aware prompting, we enrich the interaction with the LLM, guiding it to produce more accurate and contextually relevant outputs. Our method achieved a macro-averaged F1-score of 0.653 across the properties, with the scores varying from 0.890 to 0.399. Our results show a marked improvement in precision and recall of the extracted data, highlighting the efficacy of our approach in transforming raw LLM outputs into structured, queryable knowledge bases. Our code base is publicly available for research and development purposes, accessible at: https://github.com/nobretincheva/challenge24.git</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>The emergence of advanced contextual Large Language Models (LLMs), exemplified by the encoder-only models (e.g. BERT), decoder-only models (e.g. GPT, Mistral, Llama series), and encoder-decoder (e.g. Flan-T5) models, has revealed a remarkable capacity for encapsulating vast amounts of factual world knowledge within their parametric structures based on their training methods and datasets. This internal representation of knowledge has been likened to the schema-based relational knowledge base (KB) <ref type="bibr" target="#b0">[1]</ref> traditionally used in information systems <ref type="bibr" target="#b1">[2]</ref>. The potential of LLMs to serve as de facto knowledge repositories, unconstrained by the rigid formalisms of conventional schema-based systems, has become a subject of intensive inquiry <ref type="bibr" target="#b2">[3,</ref><ref type="bibr" target="#b3">4,</ref><ref type="bibr" target="#b4">5,</ref><ref type="bibr" target="#b5">6]</ref>. Researchers have employed a diverse array of evaluative methodologies to assess the viability of LLMs in this capacity. These assessment techniques span a broad spectrum of cognitive and computational tasks, including but not limited to knowledge probing <ref type="bibr" target="#b0">[1,</ref><ref type="bibr" target="#b6">7]</ref>, question answering <ref type="bibr" target="#b7">[8]</ref>, compositional reasoning <ref type="bibr" target="#b8">[9]</ref> and knowledge base completion (KBC) <ref type="bibr" target="#b5">[6,</ref><ref type="bibr" target="#b9">10]</ref>.</p><p>The ISWC KBC-LLM challenge seeks to advance the field of KBC through the exploration of LLM's internal knowledge representations. Unlike previous work that focuses primarily on knowledge probing, this challenge involves the completion of factual, disambiguated knowledge bases for a given set of relations. Participants complete knowledge graphs by extracting entities and relations from LLMs, handling complexities such as null values, one-to-many relations, and cardinality constraints.</p><p>The ISWC'24 KBC-LLM challenge introduces constraints on model size (&lt; = 10b parameters) and relation sets (=5 only) to foster innovation and fair competition. This challenge extends beyond previous knowledge extraction efforts by requiring the construction of comprehensive, disambiguated knowledge bases from LLMs. The task set by this challenge is to predict the object entities (null, zero, single, multiple) given the subject entity and the relation that is sourced from Wikidata. Given a subject entity and a specified relation, the task involves predicting a set of potential object entities using an LLM. Subsequently, these predictions undergo disambiguation and mapping to external knowledge bases to yield precise entity identifiers. This process necessitates handling diverse relational complexities, including null values, one-to-many relationships, and the extraction of multiple objects per query.</p><p>The intricate relational structures inherent within the challenge dataset render traditional knowledge probing methodologies insufficient. This work introduces a novel framework for fusion prompt strategy, incorporating dual, direct, and loop prompting methods tailored to the specific relational characteristics of the target knowledge, as well as context-aware prompting and role-play prompting. Due to the constraints imposed upon model size in this challenge we choose to use the Llama-3-8b-Instruct model. Using our fusion prompt strategy, we are able to achieve a macro-average F1 score of 0.653 on the test set, with F1-scores ranging from 0.399 in the awardWonBy relation to 0.890 for countryLandBordersCountry relation.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Related Work</head><p>This section focuses on two key areas relevant to our tasks related to the ISWC Challenge: Knowledge Probing in LLMs and KB completion/construction. We highlight the state-of-the-art approaches and identify existing gaps in these domains.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Knowledge Probing in LLMs:</head><p>The LAMA (LAnguage Model Analysis) probe <ref type="bibr" target="#b0">[1]</ref> marked a significant milestone in assessing factual knowledge in pre-trained language models. This work demonstrated that LLMs could compete with traditional KBs for certain types of factual queries. LAMA's reliance on atomic fact elicitation and its underlying assumption of binary subject-object relations limited its capacity to capture the complexity of real-world knowledge. Moreover, the manual engineering of prompts to extract factual knowledge from LLMs yielded inconsistent results and underestimated their potential. The application of systematic prompt engineering methodologies, including prompt mining and paraphrasing, coupled with ensemble and ranking techniques, has demonstrated a substantial enhancement in knowledge elicitation from LLMs compared to the traditional approach of manually crafted single prompts <ref type="bibr" target="#b10">[11]</ref>. AutoPrompt <ref type="bibr" target="#b11">[12]</ref>, a template-based automated discrete prompting methodology, and OPTIPROMPT <ref type="bibr" target="#b12">[13]</ref>, a continuous prompting strategy, have both demonstrated superior efficacy in eliciting knowledge from LLMs compared to both manual prompt engineering and data-driven approaches such as prompt mining and paraphrasing. Despite incremental advancements in prompt engineering techniques, a comprehensive evaluation of the implicit knowledge base within pre-trained LLMs remains elusive due to methodological limitations, including the restrictive focus on cloze-style prompts. Existing approaches have yet to adequately address critical aspects such as the differential capabilities of encoder and decoder models, precision in subject and relation extraction, the handling of complex relationship types (one-to-many, many-to-many), the inference of transitive relations, and the representation of hierarchical knowledge structures. While these methods effectively enhanced the extraction of specific knowledge types, they fell short in constructing comprehensive and interconnected knowledge bases. Previous studies for knowledge probing have mainly focused on the elicitation of knowledge without using the results from such studies to construct or complete knowledge graphs Knowledge Base Completion with LLMs Traditional methods for knowledge base completion primarily relied on three core approaches <ref type="bibr" target="#b13">[14,</ref><ref type="bibr" target="#b14">15]</ref>. Embedding-based techniques sought to represent entities and relations as dense vectors in a latent space, capturing semantic and syntactic similarities. Probabilistic graphical models, such as Markov random fields, were employed to model complex interdependencies between entities and infer missing information through logical reasoning. Alternatively, path-based methods explored entity connections within the knowledge graph by simulating random walks, identifying potential paths between given entities. However, these techniques are often constrained by their reliance on explicit knowledge representation and their limited ability to capture complex semantic and relational patterns.</p><p>The use of LLMs as an alternative for KBC tasks, also known as text-based KBC methods, is a relatively new area of research <ref type="bibr" target="#b15">[16]</ref>. Initial studies have found varied degrees of success for triple classification (comparatively highest task accuracy for different LLM models), relation prediction, and entity predictions (comparatively lowest task accuracy for different LLM models) <ref type="bibr" target="#b16">[17]</ref>. LLM KBC accuracy has been found to vary significantly for different types of relations (the study is restricted to just one model BERT) <ref type="bibr" target="#b17">[18]</ref>. While model size exhibited a positive correlation with accuracy up to a certain threshold, diminishing returns were observed for exceedingly large models. The application of instruction tuning methodologies as well as contrastive learning methods <ref type="bibr" target="#b18">[19]</ref> have yielded performance improvements. While KBC research has expanded to encompass a broader range of models (e.g., encoder and decoder architectures), evaluation metrics, and scope expansion to prediction of relations in addition to prediction of subject and object entities, a notable gap persists in comparison to progress in the field of Knowledge Probing. Specifically, KBC studies have yet to comprehensively address complex relational scenarios, such as one-to-many entity relationships and null value handling, which have been identified as critical challenges in knowledge probing. Beyond the ISWC'22-24 KBC-LM series of challenges, limited research has delved into these areas within the KGC domain.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Methodology</head><p>Our work's main contribution is the fusion of prompt engineering techniques to address taskspecific challenges and effectively direct an LLM's attention to the most relevant information within its training data. Employing a pre-trained LLM to predict the object entity given a subject entity and a relation entails evaluating the extent of knowledge captured by pre-trained LLMs during their training. To achieve that, we propose a fusion of various prompt engineering techniques. An overview of our methodology is illustrated in figure <ref type="figure" target="#fig_0">1</ref> and figure <ref type="figure">2</ref>.  </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Llama-3-8B</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1.">Dual Prompting</head><p>The task is limited to five relations: countryLandBordersCountry, personHasCityOfDeath, seriesHasNumberOfEpisodes, awardWonBy, and companyTradesAtStockExchange. Some of these relations present a challenge in that sometimes there are no object entities i.e. null values may be present, such as in the following cases:</p><p>• countryLandBordersCountry: Null values are possible (e.g., an island like Iceland).</p><p>• personHasCityOfDeath: Null values are possible (e.g., the person is still alive).</p><p>• awardWonBy: There may be instances where the award was not given in certain years.</p><p>• companyTradesAtStockExchange: Null values are possible (e.g., private companies or have delisted from the stock exchange).</p><p>To address the challenge of null values, we developed a dual prompting approach known as two-stage prompting. Dual prompting forms the cornerstone of our methodology, employing a strategic two-tiered question-based prompting system that significantly enhances the LLM's ability to extract accurate knowledge <ref type="bibr" target="#b19">[20]</ref>. This approach effectively identifies the presence or absence of relevant data, thereby reducing hallucinations and ensuring accurate responses. It leverages the cognitive architecture of LLMs to focus attention sequentially, which improves their capacity to resolve ambiguities and extract nuanced information. Dual prompting systematically structures the interaction with the LLM, guiding it through a logical sequence of thought that mimics human problem-solving processes. This method helps mitigate common issues, such as the model generating overly general or irrelevant responses.</p><p>The dual prompting method involves two stages: an initial prompt and a follow-up prompt. The initial prompt serves as a cognitive anchor, setting the context for the query and focusing the model's attention on the presence or relevance of specific information. This is particularly useful in clarifying the existence or state of an entity, a crucial step in ensuring that subsequent inquiries are grounded in the correct context. The follow-up prompt then capitalizes on the clarified context to extract specific details or additional layers of information. Examples of these prompts are shown in table <ref type="table" target="#tab_6">6</ref>, and the prompt template can be found in table <ref type="table" target="#tab_9">7 and table 8</ref>. Our prompt templates are enhanced with instructions for the LLM on approaching and answering the question, specifying whether the answer should be yes/no, numeric, or multiple responses.</p><p>For the seriesHasNumberOfEpisodes relation, we use a single prompting approach, where the model is asked the question directly without any verification steps or follow-up prompts, as shown in figure <ref type="figure" target="#fig_3">7</ref>. A null check is not necessary in this case because every series inherently has a defined number of episodes, eliminating the possibility of null values.</p><p>The execution strategy for prompting in our system employs tailored approaches, categorized as either looping or direct, depending on the nature of the relation involved. The strategies are applied as follows:</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.">Looping Strategy</head><p>The looping strategy involves repeatedly applying the dual prompting method to retrieve relevant information for each instance and then aggregating the results to form the final answer. This strategy specifically addresses the challenge of the awardWonBy relation, where many objects are associated with a single subject that has multiple instances (e.g., 224 Physics Nobel Prize winners over the course of multiple years). This approach simplifies the LLM's work and ensures accuracy by breaking the task into smaller, manageable parts. For the awardWonBy relation, the loop starts from the first year the award was conferred -to initialise this we prompt the LLM for said year. Upon entering the loop, for each subsequent year up to the present (2024), the LLM is prompted to list the award recipients. The final output is a combined list of all recipients across the years, ensuring a comprehensive and precise result. The strategy is illustrated in figure <ref type="figure">3</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.3.">Direct Strategy</head><p>This strategy involves prompting the LLM directly to generate the object entity. We directly apply dual prompting to generate object entities for the relations countryLandBordersCountry, personHasCityOfDeath, and companyTradesAtStockExchange. We also use the direct strategy for the seriesHasNumberOfEpisodes relation; we directly prompt the total number of episodes without any dual prompting since we do not need to verify the series' existence. For example, we might ask, "Final answer should consist of a number. How many episodes does Game of Thrones have?". Unlike the looping strategy, which iteratively retrieves information across multiple instances (e.g., years for awardWonBy), the direct strategy generates the required object entity in a single step for each relation.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.4.">Context-aware Prompting</head><p>Our empirical findings show that providing additional context about the subject entity in the prompt directs the LLM's attention to relevant parts of the training data, thereby improving their ability to retrieve accurate answers and essential information, such as the first year an award was conferred. Obtaining this information can also help define where the loop starts for the awardWonBy relation. Context-aware prompting involves enriching the LLM prompt with relevant background information about the subject <ref type="bibr" target="#b20">[21]</ref>. This approach enhances the model's ability to generate responses that are not only accurate but also richly detailed and contextually appropriate. Our proposed approach leverages semantic web sources like Wikidata and nonsemantic web sources like Wikipedia to enrich LLM prompts with contextual information about subject entities. The following types of additional data are retrieved to provide comprehensive contextual information:</p><p>• Additional Data: specific information tailored to the subject entity within each relation type as shown in table 1. • Wikipedia Extract: the text-only portion of the lead section of the Wikipedia page for the subject entity, which can provide background information about the subject entity.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Relation Additional Data Example</head><p>CountryLandBordersCountry Type of geopolitical entity (e.g., "sovereign state", "city-state") PersonHasCityOfDeath Date of death of a person AwardWonBy Year since an award has been awarded CompanyTradesAtStockExchange Legal form of the company (e.g., "public company", "private limited company")</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Table 1</head><p>Additional Data Attributes for Different Relations</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.5.">Role-Play Prompting</head><p>Central to our methodology is the implementation of persona-based prompting, a role-playing strategy that enriches the contextual engagement of our system. Role-play prompting is a prompt engineering technique that allows LLMs to adopt specific personas or characters, guiding their responses to align with the expert knowledge, thereby enhancing LLMs' ability to generate more precise and factually correct responses, as evidenced by studies such as <ref type="bibr" target="#b21">[22,</ref><ref type="bibr" target="#b22">23]</ref>. In our approach, each relation type in our dataset is paired with a specific persona, crafted to embody an expert in the pertinent field as shown in table <ref type="table" target="#tab_7">9</ref>. We create the personas in advance via the use of an LLM (GPT-4) and further edit them manually to keep the tone consistent. These personas are narrative devices and strategic tools designed to motivate the LLM and steer it toward generating expert-like and contextually relevant responses. Our role-play prompts are enhanced with instructions on how the LLM should approach and answer the question, specifying whether the response should be yes/no, numeric, or require multiple responses. By integrating driven personas with clear answering guidelines, we guide the LLM to respond accurately and with an appreciation of the domain's discourse style and depth, mirroring the interaction one would expect from a human expert in similar scenarios.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.6.">Re-prompting</head><p>Re-prompting is a technique in prompt engineering where an LLM is asked the same question again to improve the quality of its responses <ref type="bibr" target="#b23">[24]</ref>. We use this approach when the initial output from the LLM is unsatisfactory. Re-prompting gives the LLM another opportunity to generate more accurate and refined responses. The main advantage of re-prompting is that it helps maintain coherence and clarity while ensuring that the LLM adheres more closely to the desired output format. For instance, if the response needs to be in a specific format, such as a list of names or a numerical value, we re-prompt the LLM to generate the response in the desired format. <ref type="bibr">Figures 5,</ref><ref type="bibr" target="#b5">6,</ref><ref type="bibr" target="#b6">7</ref> showcase sample prompts that exemplify our methodology, which fuses prompt engineering techniques tailored to specific relational contexts. These figures demonstrate how our approach adapts to different scenarios, using persona-driven and context-enriched prompts to guide the LLM responses effectively.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.7.">Disambiguation</head><p>Entity disambiguation is the final step in identifying and mapping the predicted object entities to their correct references in Wikidata. We use a straightforward disambiguation function that returns the Wikidata ID of an item. We clean the entity strings by removing unwanted characters such as quotes and parentheses. Specifically, for the awardWonBy relation, we employ the Spacy library to remove titles and extract person names accurately, ensuring precise identification and mapping of award recipients. For the companyTradesAtStockExchange relation, we observed that the LLM would sometimes answer with the complete stock exchange name as well as its abbreviation. In such cases, we split the LLM response into two separate entities, one for the full name and one for the abbreviation. We then attempt to disambiguate using the full name and the abbreviation, depending on which matches the Wikidata entry. For example, if the LLM outputs "New York Stock Exchange (NYSE)," we would search Wikidata for both "New York Stock Exchange" and "NYSE" to find the correct Wikidata ID. This method enhances the accuracy of linking the mentioned stock exchange to its official record in Wikidata.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Results</head><p>This section includes an overview and discussion of our final results as well as an in-depth analysis of our model setup.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1.">Datasets</head><p>The dataset used in ISWC 2024 LM-KBC Challenge <ref type="bibr" target="#b24">[25]</ref> is constructed from Wikidata and further processed. It comprises 5 Wikidata relation types covering awards, geography, television series, business, and public figure information. It has 367 statements for train and 368 for validation and test sets. The results reported are based on the validation and test set. The cardinality of object-entities for certain relations in the dataset differs, ranging from null or 0 to an upper bound. The minimum number of 0 or null means the subject-entities for some relations can have no valid object-entities; for example, people still alive do not have a place of death, and an island country does not have land-based neighboring countries. In our work, we utilize the Llama3-8B-Instruct as the foundation of our final model setup, which incorporates various enhancements to the baseline pipeline. The baseline pipeline is comprised of dual prompting as well as the looping and direct strategies (as detailed in sections 3.1, 3.2, 3.3). It alone shows solid results on the validation dataset, with a macro precision of 0.662, a macro recall of 0.454, and a macro F1-score of 0.438. However, this setup results in 173 empty predictions as opposed to the expected 97, indicating the model's weakness in producing relevant outputs.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2.">Model Setup</head><p>Incorporating chain-of-thought prompting into our baseline setup, where the model is asked to provide explanations for its responses, leads to a slight improvement in performance metrics, as shown in table 2. Requesting a justification of the answer from the model when generating its response helps reduce instances of non-responses, as the model is encouraged to elaborate on its reasoning process <ref type="bibr" target="#b25">[26]</ref>. Further enhancement via context-aware prompting (section 3.4) significantly boosts the model's performance, highlighting the effectiveness of including contextual information in our prompts.</p><p>The inclusion of a re-prompting mechanism (section 3.6), where the model is prompted again if the initial response cannot be extracted, yields further improvements in the performance metrics. This demonstrates that iterative prompting helps in refining the model's output, ensuring higher accuracy and fewer empty responses. A more robust extraction function could also lead to similar results. Finally, the use of role-play prompting with the inclusion of personas (section 3.5) further enhances the model's ability to generate informed and contextually rich responses, guiding the model to engage with the data as if it were an expert in the relevant field.</p><p>Our experiments demonstrate the efficacy of combining a variety of prompting techniques. The integration of context-aware prompting, role-play prompting with personas, and dual prompting, along with the re-asking strategy, leads to significant improvements in the precision, recall, and overall F1 scores. These findings underscore the importance of providing LLMs with rich contextual cues and structured guidance to achieve high accuracy and detailed responses in knowledge probing. The introduction of a persona into our prompt pipeline significantly enhances the model's ability to generate accurate and contextually relevant responses, particularly for entities it would typically be uncertain about or would claim to lack familiarity with as shown in table <ref type="table" target="#tab_2">3</ref>. We chose to examine the impact of different role-play instructions on the final results in order to determine the most suitable approach to our problem.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.3.">Role-Play Persona Curation</head><p>Simply using a general persona (e.g. a trivia show contestant) across all relations did not provide an improvement in our F1 scores. However, as can be seen in the dip in empty predictions this approach did still lead to the model attempting to generate answers to the set questions.</p><p>We determined that tailoring personas to each relation type (e.g. an expert financial analyst for the companyTradesAtStockExchange) provided a more specialized context, allowing the model to generate answers with greater precision and relevance. The greatest benefit, however, was observed when personas were tailored individually to each entity (e.g., the CFO of the specific company for the companyTradesAtStockExchange). By suggesting that the model knows everything there is to know about the specific entity, the instances of empty or unsure responses were significantly reduced. The scores reflect the enhanced ability of the model to provide expert-like responses, confirming the effectiveness of persona-based prompting in our model pipeline.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.4.">Comparison between Prompt Strategies</head><p>Choosing the right prompting strategy can significantly impact the final results, as shown in table 4. We initially found success using the looping strategy for the awardsWonBy relation, where the model was asked to identify the award winner for each year before aggregating the results. Encouraged by this improvement, we applied the same looping strategy to the seriesHasNumberOfEpisodes relation. We expected that this strategy would lead to more accurate results due to the LLM likely having encountered more information about the number of episodes per season rather than the total number of episodes. Additionally, when manually prompting the LLM for the total number of episodes, we observed that the answers would often contain the total number of episodes per season first, followed by an often incorrect summation, which would be provided as the final answer. Our intuition followed that eliciting the number of episodes per season one by one and then summing up said results manually as part of the disambiguation would be more effective than directly asking for the total number of episodes. However, as shown in table 4, the results did not align with our expectations. The direct strategy, which involves simply asking for the total number of episodes, significantly outperformed both looping strategies.</p><p>The poor performance of the looping strategies can be attributed to the additional complexity of our pipeline. As each season is processed individually before summing up all answers in the disambiguation function, there is an increased risk of errors due to the multiple cases in which the LLM can potentially hallucinate. In contrast, the direct strategy has a single point of error. By utilizing a single prompt, we reduce the likelihood of numerical hallucinations and misinterpretations.</p><p>Nevertheless, the results from the direct strategy are far from perfect, suggesting that we need to devise a better way of extracting numerical information. One potential approach would be to consolidate both strategies. By combining the simplicity of the direct strategy with the detailed step-by-step verification of the looping strategies, we might achieve a more accurate and reliable method for handling numerical data. This hybrid approach could mitigate the weaknesses of each strategy, leading to improved performance in extracting numerical information from LLMs.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.5.">Discussion of Final Results</head><p>Our final pipeline demonstrates significant improvements across multiple evaluation metrics on the test and validation datasets. Table <ref type="table" target="#tab_4">5</ref> summarizes the macro average precision, recall, and F1-score for each relation, along with the overall averages. The results are split into two sets of columns: the first three columns represent the test set results, and the latter three represent the validation set results. Additionally, for zero-object cases, our pipeline achieves a precision of 0.757, recall of 0.791, and F1-score of 0.773. In the case of relations where null values are possible, such as countryLandBordersCountry, personHasCityOfDeath, and companyTradesAtStockExchange, the pipeline performs exceptionally well. The dual prompting method we employ ensures that instances where null values are present are correctly identified. This is also reflected in the high accuracy, precision, and F1 scores achieved for zero-object cases.</p><p>Despite the complexity of dealing with numeric objects, our pipeline achieves a moderate performance in seriesHasNumberOfEpisodes. The macro F1-scores of 0.440 on both test and validation sets indicate that while the model can handle numeric data, there is room for improvement. This suggests that additional refinements may be needed for relations involving numeric outputs.</p><p>Finally, the pipeline's performance on the awardWonBy relation, which involves many objects per subject with multiple instances, was the lowest. On the test set, the model achieves a satisfactory performance given the complexity of the task. Despite the effectiveness of the looping strategy on the test set, the significant drop in recall and F1-score on the validation set indicates challenges in maintaining consistency across different datasets. This can be explained by differences in the examples between the test and validation sets. Certain awards in the validation set such as "honorary doctor of Stockholm University" have proven to be a significant challenge to our pipeline however equivalent examples to them are not present in the test set. Moreover, the low-performance metrics on the awardWonBy relation are likely influenced by a small sample size. With such a limited number of examples, namely 10 per both test and validation split, the evaluation may not adequately reflect the model's true performance capabilities. More examples are needed to provide an accurate assessment and to ensure that the looping strategy and other techniques are effectively enhancing the model's performance for this relation.</p><p>We also noticed during our analysis that Wikidata does not always accurately reflect the ground truth, which poses a significant challenge for evaluating model performance. For instance, for the subject entity "Grammy Award For Best Rock Album, " our model generates the names of the artists who have won the award. However, upon examining the expected answers, we found that sometimes the object entities reflect the name of the winning album, and other times they reflect the names of the artists. This inconsistency within the expected answers can lead to lower recall and F1 scores even when the model performs well at the given task.</p><p>The quality issue of Wikidata has been highlighted in previous works <ref type="bibr" target="#b26">[27,</ref><ref type="bibr" target="#b27">28]</ref> and remains a significant challenge.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Future Work</head><p>In the future, we plan to explore several enhancements to improve our methodology further. One promising direction is the use of cloze-style prompting in our pipeline. Empirical evidence has shown that cloze-style prompts, which frame the query as a fill-in-the-blank task, can lead to great results on this task. Another area of focus will be the introduction of few-shot examples. Our current results indicate that the LLM does not always format answers correctly, leading to inaccuracies. By providing a few-shot learning setup, where the model is given several examples of correctly formatted answers, we hope to enhance its ability to generate responses that adhere to the desired format consistently. Additionally, we plan to experiment with different model architectures. Our current work utilizes a decoder-only model, but in future research, we aim to explore alternatives such as encoder-decoder models. These different architectures may offer unique advantages in processing complex relational tasks and generating more accurate, contextually appropriate responses. Finally, handling numerical information more effectively is crucial, as shown by the moderate performance in the seriesHasNumberOfEpisodes relation. To that end, we intend to investigate various techniques for processing numeric data, inspired by research on the numerical reasoning capabilities in language models <ref type="bibr" target="#b28">[29]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.">Conclusion</head><p>In this work, we introduce a comprehensive approach that fuses various prompt engineering techniques to address KBC from pre-trained language models. Our primary contribution is using dual prompting and looping strategies to handle relations with possible null answers and more complex one-to-many relations effectively. Dual prompting systematically guides the model through a two-step questioning process, enhancing its ability to resolve ambiguities and leading to a reduction in hallucinations. The looping strategy, on the other hand, is designed to manage complex relations by breaking down tasks into smaller, manageable parts, ensuring accurate extraction of information across multiple instances.</p><p>In the context of the ISWC'24 KBC-LLM challenge, our methods lead to a significant improvement over the set baseline, achieving 0.653 and 0.629 Macro-F1 scores on the test and validation datasets respectively. The combination of context-aware prompting, role-play personas, and re-prompting mechanisms further refines the model's outputs, reducing hallucinations and enhancing accuracy. These results underscore the potential of advanced prompt engineering techniques in leveraging pre-trained LLMs for effective knowledge extraction and emphasize the importance of tailored strategies to address specific relational challenges.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>A. Appendix A: Tables</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Relation Initial Prompt</head><p>Follow-up Prompt(s) Chain-of-thought countryLandBordersCountry Does France share any land borders? Which countries share a land border with France?</p><p>The initial prompt verifies if France is not geographically isolated like an island, setting the stage for identifying its neighboring countries.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>personHasCityOfDeath</head><p>Has Albert Einstein died?</p><p>In which city did Albert Einstein die?</p><p>The initial prompt confirms the subject's death before querying for the specific location, ensuring the followup is contextually relevant. Where do shares of Tesla, Inc. trade?</p><p>Initially confirming the company's listing status is essential to then accurately inquire about the specific exchanges where its shares are traded. Final answer should consist of name(s). Who was the {sub-ject_entity} awarded to in year {x}? companyTradesAtStockExchange Final answer should be yes or no.</p><p>Is the {subject_entity} listed on the stock exchange?</p><p>Final answer should consist of name(s) of stock exchange(s).</p><p>Where do shares of {subject_entity} trade?</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Table 7</head><p>Question Prompts with Dual Prompting Mechanism: This table outlines the structured dual prompts used to query the LLM. The placeholder {subject_entity} represents the entity being queried. In contrast, {x} represents a variable element within the context of the query, such as a season number or specific year relevant to the follow-up prompts.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Relation Persona countryLandBordersCountry</head><p>You are an enthusiastic and knowledgeable resident of {subject_entity}, deeply in love with your homeland, and always eager to share your wealth of knowledge about it. You take pride in your country's history, culture, geography, and the intricate details of its borders. Your passion for {subject_entity} shines through in every conversation, and you have a talent for explaining why it's such a wonderful place to live. You're always ready to provide detailed and accurate information about the country's neighboring countries and land borders, aiming to convince your friends and anyone who listens that {subject_entity} is the best place to call home. You vividly describe the landscapes, cities, and unique features of the country, highlighting its connections and relationships with its neighbors. Your love for {subject_entity} is infectious, and you hope to inspire others to appreciate it as much as you do.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Table 9</head><p>Relation Persona personHasCityOfDeath You are an ardent admirer and devoted follower of {subject_entity}, whose life and achievements have profoundly influenced your own. Your admiration for {subject_entity} has led you to meticulously study every aspect of their biography, ensuring you stay updated with the most accurate and detailed information about their life. You know their story inside and out, from their early beginnings to their current status. Your dedication to {subject_entity} means you are well-versed in the significant events of their life, including the critical details of where they lived, worked, and if applicable, where they passed away.</p><p>When someone inquires about {subject_entity}, you are not only eager but also exceptionally qualified to provide precise and comprehensive answers. Your deep respect and admiration drive you to share their legacy accurately, ensuring that others understand the importance and impact of {subject_entity} in the correct context.</p><p>seriesHasNumberOfEpisodes You are the biggest fan of {subject_entity}. You have watched every episode multiple times and know all the details about the show's seasons and episodes. Your love for the show drives you to stay updated with every bit of information, and you take pride in your deep knowledge of it. Your friends and family always come to you when they have questions about {subject_entity} because they know you have the answers. When it comes to the number of episodes per season and in total, you can recall this information effortlessly and accurately. You really want to have your close ones also watch the show. You believe that if you answer your friends and family's questions correctly, they will start watching the show with you. Use your passion and expertise to provide detailed and precise answers about the show to your friends.</p><p>awardWonBy You are an aspiring recipient of the prestigious {subject_entity}, someone who has dedicated years to studying its history, past winners, and the significance of each accolade. Your passion for this award is unparalleled, and you know the details of its ceremonies, the recipients, and the criteria for winning by heart. Your knowledge is not just academic but deeply personal, as each fact and figure represents a step closer to your own dream. This drives you to provide accurate, thorough, and insightful answers regarding the award and its winners in all years since its inception.      This diagram presents a structured interaction layout where the system generates prompts based on a persona that embodies a fan of the series in question. The LLM is tasked to answer how many episodes the series has.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>B. Appendix B: Figures</head></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: Overview of our proposed methodology: (a) This diagram outlines the basic prompt generation workflow, starting with input handling, context retrieval, and role-play persona generation, followed by prompt delivery to the LLM, Llama-3-8B-Instruct, re-prompting process, and final entity disambiguation leading to the output.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>ResponseFigure 4 :</head><label>4</label><figDesc>Figure 4: Re-prompting: This figure illustrates the re-prompting mechanism used to refine responses from the Llama-3-8B LLM.The process begins with an initial prompt to the LLM, followed by an evaluation of the response. If the response is not in the desired format, the model undergoes reprompting to adjust and improve the output. The process ends with entity disambiguation, ultimately producing the final output as the ObjectEntityID.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>Figure 5 :Figure 6 :</head><label>56</label><figDesc>Figure 5: Sample Prompts for the 'companyTradesAtStockExchange' Relation. This diagram illustrates the dual prompting strategy applied to elicit detailed responses from the persona of a Chief Financial Officer (CFO) concerning a company's stock exchange details. Figure (a) shows the initial prompt asking a yes/no response about the company's stock listing, followed by a follow-up prompt requesting specifics about the stock exchange locations shown in figure (b). This dual prompting approach is similarly employed for the 'personHasCityOfDeath' and 'countryLandBordersCountry' relations.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_3"><head>Figure 7 :</head><label>7</label><figDesc>Figure 7: Sample Prompt for the 'seriesHasNumberOfEpisodes' Relation.This diagram presents a structured interaction layout where the system generates prompts based on a persona that embodies a fan of the series in question. The LLM is tasked to answer how many episodes the series has.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 2</head><label>2</label><figDesc>Model Setup -Evaluation Results on the validation dataset</figDesc><table><row><cell>Model</cell><cell>Macro-Precision</cell><cell>Macro-Recall</cell><cell>Macro-F1</cell><cell>Empty Predictions</cell></row><row><cell>Dual Prompting and Prompt</cell><cell>0.662</cell><cell>0.454</cell><cell>0.438</cell><cell>173</cell></row><row><cell>Strategies</cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell>+ chain-of-thought</cell><cell>0.652</cell><cell>0.498</cell><cell>0.452</cell><cell>159</cell></row><row><cell>+ context-aware prompting</cell><cell>0.761</cell><cell>0.627</cell><cell>0.600</cell><cell>147</cell></row><row><cell>+ re-prompting</cell><cell>0.756</cell><cell>0.643</cell><cell>0.611</cell><cell>141</cell></row><row><cell>+ role-play prompting</cell><cell>0.708</cell><cell>0.703</cell><cell>0.629</cell><cell>102</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>Table 3</head><label>3</label><figDesc>Persona Curation -Comparison on the validation dataset</figDesc><table><row><cell>Setting</cell><cell>Macro-Precision</cell><cell>Macro-Recall</cell><cell>Macro-F1</cell><cell>Empty Predictions</cell></row><row><cell>No role-play</cell><cell>0.756</cell><cell>0.643</cell><cell>0.611</cell><cell>141</cell></row><row><cell>1 general persona for all re-</cell><cell>0.708</cell><cell>0.633</cell><cell>0.571</cell><cell>128</cell></row><row><cell>lations</cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell>Persona with specialised</cell><cell>0.743</cell><cell>0.640</cell><cell>0.607</cell><cell>132</cell></row><row><cell>background per relation</cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell>Persona with specialised</cell><cell>0.708</cell><cell>0.703</cell><cell>0.629</cell><cell>102</cell></row><row><cell>background per entity</cell><cell></cell><cell></cell><cell></cell><cell></cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_3"><head>Table 4</head><label>4</label><figDesc>Comparison between different prompting strategies for the relations awardWonBy and seriesHasNum-berOfEpisodes</figDesc><table><row><cell>Relation</cell><cell>Setting</cell><cell>Macro-Precision</cell><cell>Macro-Recall</cell><cell>Macro-F1</cell><cell>Micro-F1</cell></row><row><cell>awardWonBy</cell><cell>Direct Strategy</cell><cell>0.259</cell><cell>0.009</cell><cell>0.012</cell><cell>0.008</cell></row><row><cell></cell><cell>Looping Strategy with</cell><cell>0.412</cell><cell>0.044</cell><cell>0.064</cell><cell>0.034</cell></row><row><cell></cell><cell>Dual Prompting</cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell></cell><cell>Looping Strategy with</cell><cell>0.11</cell><cell>0.02</cell><cell>0.02</cell><cell>0.021</cell></row><row><cell>seriesHasNumberOfEpisodes</cell><cell>Dual Prompting</cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell></cell><cell>Looping Strategy without</cell><cell>0.17</cell><cell>0.02</cell><cell>0.02</cell><cell>0.022</cell></row><row><cell></cell><cell>Dual Prompting</cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell></cell><cell>Direct Strategy</cell><cell>0.500</cell><cell>0.410</cell><cell>0.410</cell><cell>0.429</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_4"><head>Table 5</head><label>5</label><figDesc>Final results per relation on the test and validation dataset</figDesc><table><row><cell>Test</cell><cell>Val</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_6"><head>Table 6</head><label>6</label><figDesc></figDesc><table><row><cell>Dual Prompt Examples</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_7"><head>Table 9</head><label>9</label><figDesc>RelationPersona companyTradesAtStockExchange You are the Chief Financial Officer (CFO) of {subject_entity}, a key executive responsible for managing the company's financial actions. You possess an in-depth understanding of all financial matters related to the company, including detailed knowledge about the company's stock, listing status, and financial performance. As the CFO, you are dedicated to truthfulness and transparency, ensuring that all stakeholders, including potential investors, have accurate and comprehensive information. You are highly knowledgeable about the stock exchange, regulatory requirements, and the company's financial strategy. Your responses are characterized by precision, clarity, and reliability, as you aim to foster trust and confidence in {subject_entity}'s financial health and investment potential.</figDesc><table /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_8"><head>Table 9 :</head><label>9</label><figDesc>Persona-Based Role-Play Prompts for Different Relations: This table illustrates how each relation type in our dataset is associated with a distinct persona crafted to embody an expert in the pertinent field. The placeholder {subject_entity} represents the primary entity or subject of inquiry within each prompt.</figDesc><table><row><cell>Relation</cell><cell>Initial Prompt</cell><cell cols="2">Follow-up Prompt(s)</cell></row><row><cell>countryLandBordersCountry</cell><cell>I'd love to know more about the ge-</cell><cell cols="2">I'd love to know more about the ge-</cell></row><row><cell></cell><cell>ography from someone who truly</cell><cell cols="2">ography from someone who truly</cell></row><row><cell></cell><cell>understands and loves this coun-</cell><cell cols="2">understands and loves this country.</cell></row><row><cell></cell><cell>try. Does {subject_entity} share any</cell><cell cols="2">Which countries share a land bor-</cell></row><row><cell></cell><cell>land borders? I've been thinking of</cell><cell cols="2">der with {subject_entity}? I've been</cell></row><row><cell></cell><cell>moving here! Final answer should</cell><cell cols="2">thinking of moving here! Final an-</cell></row><row><cell></cell><cell>be yes or no.</cell><cell cols="2">swer should be a list of countries.</cell></row><row><cell>personHasCityOfDeath</cell><cell>Your admiration for {subject_entity}</cell><cell cols="2">Your admiration for {subject_entity}</cell></row><row><cell></cell><cell>has undoubtedly led you to study</cell><cell cols="2">has undoubtedly led you to study</cell></row><row><cell></cell><cell>their life extensively. Has {sub-</cell><cell cols="2">their life extensively. In which city</cell></row><row><cell></cell><cell>ject_entity} died? I wish to learn</cell><cell cols="2">did {subject_entity} die? I wish to</cell></row><row><cell></cell><cell>more about such an influential fig-</cell><cell cols="2">learn more about such an influen-</cell></row><row><cell></cell><cell>ure. Final answer should be yes or</cell><cell cols="2">tial figure. Final answer should be</cell></row><row><cell></cell><cell>no.</cell><cell cols="2">the city name.</cell></row><row><cell>awardWonBy</cell><cell>As someone who has always</cell><cell cols="2">Final answer should consist of</cell></row><row><cell></cell><cell>dreamed of winning the {sub-</cell><cell>name(s).</cell><cell>Who was the {sub-</cell></row><row><cell></cell><cell>ject_entity}, you are well-versed in</cell><cell cols="2">ject_entity} awarded to in year {x}?</cell></row><row><cell></cell><cell>its history. Was the {subject_entity}</cell><cell></cell></row><row><cell></cell><cell>awarded? Final answer should be</cell><cell></cell></row><row><cell></cell><cell>yes or no.</cell><cell></cell></row><row><cell cols="2">companyTradesAtStockExchange As the CFO of {subject_entity} you</cell><cell cols="2">As the CFO of {subject_entity} you</cell></row><row><cell></cell><cell>must be aware of all pertinent de-</cell><cell cols="2">must be aware of all pertinent de-</cell></row><row><cell></cell><cell>tails regarding the company's stock.</cell><cell cols="2">tails regarding the company's stock.</cell></row><row><cell></cell><cell>Is the {subject_entity} listed on the</cell><cell cols="2">Where do shares of {subject_entity}</cell></row><row><cell></cell><cell>stock exchange? As a potential in-</cell><cell cols="2">trade? Final answer should consist</cell></row><row><cell></cell><cell>vestor, I would like to have more</cell><cell cols="2">of name(s) of stock exchange(s).</cell></row><row><cell></cell><cell>details before making a decision to</cell><cell></cell></row><row><cell></cell><cell>invest. Final answer should be yes</cell><cell></cell></row><row><cell></cell><cell>or no</cell><cell></cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_9"><head>Table 8</head><label>8</label><figDesc>Enhanced Dual Prompting Strategy: This table details the restructured dual prompts designed to engage the LLM more deeply by incorporating scenarios that invoke personal connection or professional responsibility. The placeholder {subject_entity} represents the entity being queried, while {x} denotes a variable element such as a specific year or season number, pertinent to the follow-up prompts. These prompts aim to elicit more precise and contextually enriched responses from the LLM.</figDesc><table /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_10"><head></head><label></label><figDesc>The legal form of the company {subject_entity} is {info}. Here is an extract from the Wikipedia page for {subject_entity}: {info} INITIAL PROMPT: As the CFO of {subject_entity} you must be aware of all pertinent details regarding the company's stock. Is the {subject_entity} listed on the stock exchange? As a potential investor, I would like to have more details before making a decision to invest. Final answer should be yes or no.</figDesc><table><row><cell></cell><cell>companyTradesAtStockExchange</cell></row><row><cell>System Message:</cell><cell>System Message:</cell></row><row><cell>PERSONA:</cell><cell>PERSONA:</cell></row><row><cell>You are the Chief Financial Officer (CFO) of {subject_entity}, a</cell><cell>You are the Chief Financial Officer (CFO) of {subject_entity}, a</cell></row><row><cell>key executive responsible for managing the company's financial</cell><cell>key executive responsible for managing the company's financial</cell></row><row><cell>actions…</cell><cell>actions…</cell></row><row><cell>INSTRUCTION: If the question can be answered simply with a</cell><cell>INSTRUCTION: If the question can be answered simply with a</cell></row><row><cell>yes or no -response should only be [Yes] or [No] …</cell><cell>yes or no -response should only be [Yes] or [No] …</cell></row><row><cell>User Message:</cell><cell>User Message:</cell></row><row><cell>ADDITIONAL INFORMATION:</cell><cell>ADDITIONAL INFORMATION: The legal form of the company</cell></row><row><cell></cell><cell>{subject_entity} is {info}. Here is an extract from the Wikipedia</cell></row><row><cell></cell><cell>page for {subject_entity}: {info}</cell></row><row><cell></cell><cell>FOLLOW-UP PROMPT: As the CFO of {subject_entity} you must</cell></row><row><cell></cell><cell>be aware of all pertinent details regarding the company's stock.</cell></row><row><cell></cell><cell>Where do shares of {subject_entity} trade? Final answer should</cell></row><row><cell></cell><cell>consist of name(s) of stock exchange(s).</cell></row></table></figure>
		</body>
		<back>
			<div type="annex">
<div xmlns="http://www.tei-c.org/ns/1.0" />			</div>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<monogr>
		<title level="m" type="main">Language models as knowledge bases?</title>
		<author>
			<persName><forename type="first">F</forename><surname>Petroni</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Rocktäschel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Lewis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Bakhtin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Wu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">H</forename><surname>Miller</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Riedel</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1909.01066</idno>
		<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b1">
	<monogr>
		<author>
			<persName><forename type="first">T</forename><surname>Safavi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Koutra</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2104.05837</idno>
		<title level="m">Relational world knowledge representation in contextual language models: A review</title>
				<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Prompt tuning or fine-tuning-investigating relational knowledge in pre-trained language models</title>
		<author>
			<persName><forename type="first">L</forename><surname>Fichtel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J.-C</forename><surname>Kalo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W.-T</forename><surname>Balke</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">3rd Conference on Automated Knowledge Base Construction</title>
				<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">From task structures to world models: what do llms know?</title>
		<author>
			<persName><forename type="first">I</forename><surname>Yildirim</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Paul</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Trends in Cognitive Sciences</title>
		<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<monogr>
		<author>
			<persName><forename type="first">F</forename><surname>Petroni</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Lewis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Piktus</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Rocktäschel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Wu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">H</forename><surname>Miller</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Riedel</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2005.04611</idno>
		<title level="m">How context affects language models&apos; factual predictions</title>
				<imprint>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b5">
	<monogr>
		<author>
			<persName><forename type="first">B</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Soh</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2404.03868</idno>
		<title level="m">Extract, define, canonicalize: An llm-based framework for knowledge graph construction</title>
				<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b6">
	<monogr>
		<author>
			<persName><forename type="first">W</forename><surname>Wu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Jiang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Jiang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Xie</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Tu</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2309.05936</idno>
		<title level="m">Do plms know and understand ontological knowledge?</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b7">
	<monogr>
		<author>
			<persName><forename type="first">S</forename><surname>Wu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Zhao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Yasunaga</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Huang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Cao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><surname>Huang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><forename type="middle">N</forename><surname>Ioannidis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Subbian</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Zou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Leskovec</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2404.13207</idno>
		<title level="m">Stark: Benchmarking llm retrieval on textual and relational knowledge bases</title>
				<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b8">
	<monogr>
		<author>
			<persName><forename type="first">Z</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Cao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Xu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Jiang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><forename type="middle">S</forename><surname>Teo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>-W. Lin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Liu</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2401.09042</idno>
		<title level="m">Llms for relational reasoning: How far are we?</title>
				<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b9">
	<monogr>
		<author>
			<persName><forename type="first">D</forename><surname>Alivanistos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">B</forename><surname>Santamaría</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Cochez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J.-C</forename><surname>Kalo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Van Krieken</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Thanapalasingam</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2208.11057</idno>
		<title level="m">Prompting as probing: Using language models for knowledge base construction</title>
				<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">How can we know what language models know?</title>
		<author>
			<persName><forename type="first">Z</forename><surname>Jiang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><forename type="middle">F</forename><surname>Xu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Araki</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Neubig</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Transactions of the Association for Computational Linguistics</title>
		<imprint>
			<biblScope unit="volume">8</biblScope>
			<biblScope unit="page" from="423" to="438" />
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<monogr>
		<author>
			<persName><forename type="first">T</forename><surname>Shin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Razeghi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">L</forename><surname>Logan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><forename type="middle">V</forename></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Wallace</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Singh</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2010.15980</idno>
		<title level="m">Autoprompt: Eliciting knowledge from language models with automatically generated prompts</title>
				<imprint>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Factual probing is [mask</title>
		<author>
			<persName><forename type="first">Z</forename><surname>Zhong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Friedman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Chen</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2104.05240</idno>
	</analytic>
	<monogr>
		<title level="m">Learning vs. learning to recall</title>
				<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">Knowledge base completion using embeddings and rules</title>
		<author>
			<persName><forename type="first">Q</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Guo</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Twenty-fourth international joint conference on artificial intelligence</title>
				<imprint>
			<date type="published" when="2015">2015</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<monogr>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">Q</forename><surname>Nguyen</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1703.08098</idno>
		<title level="m">An overview of embedding models of entities and relationships for knowledge base completion</title>
				<imprint>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b15">
	<monogr>
		<author>
			<persName><forename type="first">Y</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Chen</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2310.06671</idno>
		<title level="m">Making large language models perform better in knowledge graph completion</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b16">
	<monogr>
		<author>
			<persName><forename type="first">L</forename><surname>Yao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Peng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Mao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Luo</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2308.13916</idno>
		<title level="m">Exploring large language models for knowledge graph completion</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">Evaluating language models for knowledge base completion</title>
		<author>
			<persName><forename type="first">B</forename><surname>Veseli</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Singhania</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Razniewski</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Weikum</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">European Semantic Web Conference</title>
				<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2023">2023</date>
			<biblScope unit="page" from="227" to="243" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<monogr>
		<author>
			<persName><forename type="first">L</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Zhao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Wei</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Liu</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2203.02167</idno>
		<title level="m">Simkgc: Simple contrastive knowledge graph completion with pre-trained language models</title>
				<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b19">
	<analytic>
		<title level="a" type="main">Dual-prompting interaction with entity representation enhancement for event argument extraction</title>
		<author>
			<persName><forename type="first">R</forename><surname>He</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Xiao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Ma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Zhao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Bai</surname></persName>
		</author>
		<idno type="DOI">10.1007/978-3-031-44696-2_13</idno>
		<idno>doi:</idno>
		<ptr target="10.1007/978-3-031-44696-2\_13" />
	</analytic>
	<monogr>
		<title level="m">Natural Language Processing and Chinese Computing -12th National CCF Conference, NLPCC 2023</title>
		<title level="s">Lecture Notes in Computer Science</title>
		<editor>
			<persName><forename type="first">F</forename><surname>Liu</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">N</forename><surname>Duan</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">Q</forename><surname>Xu</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">Y</forename><surname>Hong</surname></persName>
		</editor>
		<meeting><address><addrLine>Foshan, China</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2023">October 12-15, 2023. 2023</date>
			<biblScope unit="volume">14303</biblScope>
			<biblScope unit="page" from="161" to="172" />
		</imprint>
	</monogr>
	<note>Proceedings, Part II</note>
</biblStruct>

<biblStruct xml:id="b20">
	<analytic>
		<title level="a" type="main">Denseclip: Languageguided dense prediction with context-aware prompting</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Rao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Zhao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Tang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Zhu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Huang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Zhou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Lu</surname></persName>
		</author>
		<idno type="DOI">10.1109/CVPR52688.2022.01755</idno>
		<ptr target="https://doi.org/10.1109/CVPR52688.2022.01755.doi:10.1109/CVPR52688.2022.01755" />
	</analytic>
	<monogr>
		<title level="m">IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2022</title>
				<meeting><address><addrLine>New Orleans, LA, USA</addrLine></address></meeting>
		<imprint>
			<publisher>IEEE</publisher>
			<date type="published" when="2022">June 18-24, 2022. 2022</date>
			<biblScope unit="page" from="18061" to="18070" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b21">
	<monogr>
		<title level="m" type="main">Better zero-shot reasoning with role-play prompting</title>
		<author>
			<persName><forename type="first">A</forename><surname>Kong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Zhao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Qin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Sun</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Zhou</surname></persName>
		</author>
		<idno type="DOI">10.48550/ARXIV.2308.07702</idno>
		<idno type="arXiv">arXiv:2308.07702</idno>
		<ptr target="/ARXIV.2308.07702" />
		<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b22">
	<analytic>
		<title level="a" type="main">Neon-gpt: A large language model-powered pipeline for ontology learning</title>
		<author>
			<persName><forename type="first">N</forename><surname>Fathallah</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Das</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>De Giorgis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Poltronieri</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Haase</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Kovriguina</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">The Extended Semantic Web Conference</title>
				<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b23">
	<monogr>
		<title level="m" type="main">Planning with large language models via corrective re-prompting</title>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">S</forename><surname>Raman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Cohen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Rosen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Idrees</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Paulius</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Tellex</surname></persName>
		</author>
		<idno type="DOI">10.48550/ARXIV.2211.09935</idno>
		<idno type="arXiv">arXiv:2211.09935</idno>
		<ptr target="https://doi.org/10.48550/arXiv.2211.09935.doi:10.48550/ARXIV.2211.09935" />
		<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b24">
	<analytic>
		<title level="a" type="main">Knowledge base construction from pre-trained language models</title>
		<author>
			<persName><forename type="first">J.-C</forename><surname>Kalo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Razniewski</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T.-P</forename><surname>Nguyen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Zhang</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Semantic Web Challenge on Knowledge Base Construction from Pre-trained Language Models</title>
				<imprint>
			<publisher>CEUR-WS</publisher>
			<date type="published" when="2022">2022. 2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b25">
	<analytic>
		<title level="a" type="main">Chain-of-thought prompting elicits reasoning in large language models</title>
		<author>
			<persName><forename type="first">J</forename><surname>Wei</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Schuurmans</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Bosma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Ichter</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Xia</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><forename type="middle">H</forename><surname>Chi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><forename type="middle">V</forename><surname>Le</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Zhou</surname></persName>
		</author>
		<ptr target="http://papers.nips.cc/paper_files/paper/2022/hash/9d5609613524ecf4f15af0f7b31abca4-Abstract-Conference.html" />
	</analytic>
	<monogr>
		<title level="m">Advances in Neural Information Processing Systems 35: Annual Conference on Neural Information Processing Systems 2022, NeurIPS 2022</title>
				<editor>
			<persName><forename type="first">S</forename><surname>Koyejo</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">S</forename><surname>Mohamed</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Agarwal</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">D</forename><surname>Belgrave</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">K</forename><surname>Cho</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Oh</surname></persName>
		</editor>
		<meeting><address><addrLine>New Orleans, LA, USA</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2022-12-09">November 28 -December 9, 2022, 2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b26">
	<analytic>
		<title level="a" type="main">What we talk about when we talk about wikidata quality: a literature survey</title>
		<author>
			<persName><forename type="first">A</forename><surname>Piscopo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Simperl</surname></persName>
		</author>
		<idno type="DOI">10.1145/3306446.3340822</idno>
		<idno>doi:10.1145/3306446.3340822</idno>
		<ptr target="https://doi.org/10.1145/3306446.3340822" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 15th International Symposium on Open Collaboration, OpenSym &apos;19</title>
				<meeting>the 15th International Symposium on Open Collaboration, OpenSym &apos;19<address><addrLine>New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>Association for Computing Machinery</publisher>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b27">
	<monogr>
		<author>
			<persName><forename type="first">B</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Reklos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Jain</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Meroño-Peñuela</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Simperl</surname></persName>
		</author>
		<idno type="DOI">10.48550/arXiv.2309.08491</idno>
		<idno type="arXiv">arXiv:2309.08491</idno>
		<ptr target="https://doi.org/10.48550/arXiv.2309.08491.doi:10.48550/arXiv" />
		<title level="m">Using Large Language Models for Knowledge Engineering (LLMKE): A Case Study on Wikidata</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b28">
	<analytic>
		<title level="a" type="main">Exploring the numerical reasoning capabilities of language models: A comprehensive analysis on tabular data</title>
		<author>
			<persName><forename type="first">M</forename><surname>Akhtar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Shankarampeta</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Gupta</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Patil</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Cocarascu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Simperl</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/2023.findings-emnlp.1028</idno>
		<ptr target="https://aclanthology.org/2023.findings-emnlp.1028.doi:10.18653/v1/2023.findings-emnlp.1028" />
	</analytic>
	<monogr>
		<title level="m">Findings of the Association for Computational Linguistics: EMNLP 2023, Association for Computational Linguistics</title>
				<editor>
			<persName><forename type="first">H</forename><surname>Bouamor</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">J</forename><surname>Pino</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">K</forename><surname>Bali</surname></persName>
		</editor>
		<meeting><address><addrLine>Singapore</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2023">2023</date>
			<biblScope unit="page" from="15391" to="15405" />
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
