<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Column Vocabulary Association (CVA): Semantic Interpretation of Dataless Tables</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Margherita</forename><surname>Martorana</surname></persName>
							<email>m.martorana@vu.nl</email>
							<affiliation key="aff0">
								<orgName type="department">Department of Computer Science</orgName>
								<orgName type="institution">Vrije Universiteit Amsterdam</orgName>
								<address>
									<addrLine>De Boelelaan 1105</addrLine>
									<settlement>Amsterdam</settlement>
									<country key="NL">The Netherlands</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Xueli</forename><surname>Pan</surname></persName>
							<email>x.pan2@vu.nl</email>
							<affiliation key="aff0">
								<orgName type="department">Department of Computer Science</orgName>
								<orgName type="institution">Vrije Universiteit Amsterdam</orgName>
								<address>
									<addrLine>De Boelelaan 1105</addrLine>
									<settlement>Amsterdam</settlement>
									<country key="NL">The Netherlands</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Benno</forename><surname>Kruit</surname></persName>
							<email>b.b.kruit@vu.nl</email>
							<affiliation key="aff0">
								<orgName type="department">Department of Computer Science</orgName>
								<orgName type="institution">Vrije Universiteit Amsterdam</orgName>
								<address>
									<addrLine>De Boelelaan 1105</addrLine>
									<settlement>Amsterdam</settlement>
									<country key="NL">The Netherlands</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Tobias</forename><surname>Kuhn</surname></persName>
							<email>t.kuhn@vu.nl</email>
							<affiliation key="aff0">
								<orgName type="department">Department of Computer Science</orgName>
								<orgName type="institution">Vrije Universiteit Amsterdam</orgName>
								<address>
									<addrLine>De Boelelaan 1105</addrLine>
									<settlement>Amsterdam</settlement>
									<country key="NL">The Netherlands</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Jacco</forename><surname>Van Ossenbruggen</surname></persName>
							<email>ossenbruggen@vu.nl</email>
							<affiliation key="aff0">
								<orgName type="department">Department of Computer Science</orgName>
								<orgName type="institution">Vrije Universiteit Amsterdam</orgName>
								<address>
									<addrLine>De Boelelaan 1105</addrLine>
									<settlement>Amsterdam</settlement>
									<country key="NL">The Netherlands</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Column Vocabulary Association (CVA): Semantic Interpretation of Dataless Tables</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">A35F2B33EA5DD0B6A0D8D5AABB12A404</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T19:03+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Large Language Models</term>
					<term>Metadata Enrichment</term>
					<term>Retrieval Augmented Generation</term>
					<term>Semantic Table Interpretation</term>
					<term>Semantic Web</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Traditional Semantic Table <ref type="table">Interpretation</ref> (STI) methods rely primarily on the underlying table data to create semantic annotations. This year's SemTab challenge introduced the "Metadata to KG" track, which focuses on performing STI by using only metadata information, without access to the underlying data. In response, we introduce a new term: Column Vocabulary Association (CVA). This term refers to the task of semantic annotation of column headers solely based on metadata information. This study evaluates several methods for the CVA task, including a Large Language Models (LLMs) approach combined with Retrieval Augmented Generation (RAG), using three commercial GPT models and four open-source models, along with temperature setting variations. We also evaluate a traditional similarity approach using SentenceBERT. Our experiments operate in a zero-shot setting, without fine-tuning or examples for the LLMs, to maintain a generalized and domain-agnostic application.</p><p>Initial findings indicate that LLMs generally perform well at temperatures below 1.0, achieving an accuracy of 100% on the challenge test set. Traditional methods outperforms several LLMs, instead, when metadata and glossary are closely related. However, interim results on the full data set show that our approaches reach an accuracy of 70%, suggesting possible discrepancies in test representativeness, though further investigation is needed.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>Tabular data is the most common format used for data storage and sharing <ref type="bibr" target="#b0">[1]</ref>. However, tabular data often lacks semantic annotations and can contain inaccurate or missing information. Semantic Table <ref type="table">Interpretation</ref> (STI) aims to find semantic annotations for table cells and columns, as well as column relationships, using existing Knowledge Graphs (KGs). Semantic annotations are particularly important when used to enrich and augment metadata. In fact, several studies <ref type="bibr" target="#b1">[2,</ref><ref type="bibr" target="#b2">3,</ref><ref type="bibr" target="#b3">4]</ref> have shown that high-quality metadata supports data Findability, Accessibility, Interoperability and Reusability (FAIR Guiding Principles) <ref type="bibr" target="#b4">[5]</ref>. Rich metadata plays a critical role when dealing with confidential data, as the underlying data is generally not open and freely accessible. Enhancing the FAIRness for this type of data has gained more attention in recent years, and in previous work we have shown that high-quality and rich metadata improves the discovery and reuse of these resources <ref type="bibr" target="#b5">[6]</ref>.</p><p>Nevertheless, the automatic enrichment of metadata -when only the metadata is availablepresents a significant challenge. In such cases, much of the contextual information is lacking, and the underlying data cannot be used to identify the most suitable annotations. Large Language Models (LLMs) and Retrieval Augmented Generation (RAG) <ref type="bibr" target="#b6">[7]</ref> can offer promising approaches to address these challenges. LLMs' training data can be leveraged as background knowledge, while RAG systems can further integrate external resources -such as knowledge graphs, controlled vocabularies, and glossaries -to extend the LLMs' adaptability across various domains. Also, tuning the LLM's temperature parameter allows for adjusting the model's creativity or determinism, which could be used to balance between precision and flexibility when identifying relevant annotations.</p><p>In this year's SemTab challenge, participants in the "Metadata to KG" track aim to annotate tables using only table metadata (e.g. column and table names) without accessing the underlying data. This approach tests the ability to enrich metadata effectively under similar conditions imposed by restricted access data. To guide our investigation we have formulated the following research questions:</p><p>• How do traditional semantic similarity methods compare to newer methods using Large Language Models (LLMs) in the semantic annotation of table metadata when the underlying data is not available? • How does the temperature setting of LLMs impact their performance in this task? • How do different combinations of metadata information in traditional methods affect their performance? • How does the nature of the input data and glossary influence the results?</p><p>In the pages that follow, we further describe the importance of metadata, especially in settings where the underlying data is not available. We also introduce the term "Column Vocabulary Association", before discussing our main methodology and results.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Background</head><p>In the next section, we present the key background concepts relevant to this research. First, we introduce the concept of "Dataless Tables", referring to tables where the data is confidential and cannot be accessed. Next, we provide an overview of LLMs and RAG. Finally, we introduce the term "Column-Vocabulary Association", which relates directly to the specific task set by this year's SemTab Challenge.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1.">Dataless tables</head><p>There has been a recent rise in solutions for sharing confidential or restricted-access data. For example, multiple online Open Government Data (OGD) portals, such as the Central Bureau for Statistics Netherlands (CBS)<ref type="foot" target="#foot_0">1</ref> , the U.S. Government's Open Data<ref type="foot" target="#foot_1">2</ref> , and Canada's Open Government Portal<ref type="foot" target="#foot_2">3</ref> , have been developed to enhance innovation and research by allowing users to explore population data. These portals typically provide aggregated statistics, giving the general public and researchers access to data for use in fields such as journalism, software development, and research <ref type="bibr" target="#b7">[8]</ref>. However, much of the population data remains inaccessible due to confidentiality concerns, including sensitive data like patient records, individual-level statistics, and other data containing Personally Identifiable Information (PII).</p><p>Various solutions have been proposed to facilitate the reuse of restricted-access data. For instance, the Personal Health Train <ref type="bibr" target="#b8">[9]</ref> enables users to send their algorithms to where the data is stored, allowing analysis without needing direct access to the data. However, users still need to know that the data exists and its structural details. Detailed metadata descriptions play a crucial role in addressing this challenge. In previous work, we introduced the DataSet Variable Ontology (DSV) <ref type="bibr" target="#b9">[10]</ref>, a metadata schema designed to capture information at both the dataset and variable levels. This demonstrated that high-quality metadata can enable the discovery of restricted access data by annotating non-confidential information, such as column descriptions, dataset structure, and summary statistics.</p><p>In this context, we introduce the concept of "Dataless Tables". These tables allow for the description of structural elements, summary statistics, and metadata, like column descriptions, while keeping the actual data confidential and inaccessible. Although the raw data is unavailable, such tables retain important features of the dataset, which can be annotated using frameworks like DSV, making them valuable for data discovery and analysis under restricted conditions. We refer to them as "dataless" because, while the data itself is not directly available, the tables still carry structural and contextual information useful for various applications.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2.">LLMs and RAG</head><p>Large Language Models (LLMs) have brought significant advances in the field of Natural Language Processing (NLP). LLMs are trained on a vast amount of data, which allows them to handle a wide range of tasks, including those they weren't explicitly trained for <ref type="bibr" target="#b10">[11,</ref><ref type="bibr" target="#b11">12]</ref>. Studies have shown that ChatGPT has outperformed crowd-workers in tweet classification <ref type="bibr" target="#b12">[13]</ref>, and the open-source model SOTAB in Column Property Annotation tasks <ref type="bibr" target="#b13">[14]</ref>. Additionally, it has been shown that incorporating semantic technologies and Knowledge Graphs (KGs) can further enhance the accuracy of text classification <ref type="bibr" target="#b14">[15]</ref>. However, the LLM's training data often lacks real-time updates, resulting in outdated or incomplete information <ref type="bibr" target="#b15">[16]</ref>, which can lead to factual inaccuracies or irrelevant content, a phenomenon commonly referred to as "hallucinations" <ref type="bibr" target="#b16">[17,</ref><ref type="bibr" target="#b17">18,</ref><ref type="bibr" target="#b18">19,</ref><ref type="bibr" target="#b19">20]</ref>. Furthermore, LLMs have limited domain-specific expertise, which can impact their reliability in specialized tasks <ref type="bibr" target="#b20">[21,</ref><ref type="bibr" target="#b21">22]</ref>.</p><p>To address these challenges, Retrieval-Augmented Generation (RAG) <ref type="bibr" target="#b6">[7]</ref> systems have emerged as a promising solution to these challenges <ref type="bibr" target="#b22">[23,</ref><ref type="bibr" target="#b23">24,</ref><ref type="bibr" target="#b24">25,</ref><ref type="bibr" target="#b25">26]</ref>. By combining the generative capabilities of LLMs with external, high-quality information sources, RAG systems improve the accuracy and reliability of information retrieval. These systems have been shown to enhance performance in various applications, including code generation <ref type="bibr" target="#b26">[27]</ref> and both domain-agnostic and domain-specific question answering <ref type="bibr" target="#b27">[28,</ref><ref type="bibr" target="#b28">29,</ref><ref type="bibr" target="#b29">30]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.3.">Column Vocabulary Association (CVA)</head><p>In the domain of Semantic Table Interpretation (STI), there are some well known challenges, including the Column Type Annotation (CTA), the Column Entity Annotation (CEA), and the Column Property Annotation (CPA) tasks. The CTA task involves identifying the semantic type (e.g. dates or geographical locations) of each column in the table. The CEA task, instead, involves linking each cell to an entity in a knowledge graph: for example, the cell containing the string "New York" to be linked to the WikiData entity for New York City (Q60). CPA requires the identification of relationships between columns of a table: for example, recognising that the columns with headers "Mayor's Name" and "City" are related to each other by the property eg:isMayorOf.</p><p>In this work we introduce a new term: the "Column Vocabulary Association" (CVA) task. This task differs significantly from the previous ones because it does not rely on any information from the underlying data within the table. Instead, it aims to associate column headers with entries in controlled vocabularies purely based on semantic similarities. The distinction between the word association and annotation is also important in this context. Annotation typically refers to the labeling of data with tags or categories. In contrast, with the term association we refer more on the conceptual linkage between the textual information in a column header, and an external knowledge repository. This approach emphasizes understanding and leveraging the semantic meaning of the column headers themselves, without using any underlying data. By focusing on semantic similarities, we aim to create a method for interpreting and integrating restricted access datasets, to facilitate metadata enrichment and data discovery.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">SemTab Challenge</head><p>Since its start in 2019, the Semantic Web Challenge on Tabular Data to Knowledge Graph Matching (SemTab) has focused on benchmarking systems and approaches that support and enhance Semantic Table <ref type="table">Interpretation</ref> (STI). The SemTab challenge typically consists of two main tracks: the "Accuracy Track", where participants annotate tables with tasks like CTA and CEA, and the "Dataset Track", focused on submitting new datasets and benchmarks across different domains. This year, SemTab introduced the new "Metadata to KG" track, where participants are asked to map table metadata to KGs without having access to the underlying data. This presents a unique challenge due to the limited available context, making traditional STI methods less applicable, as they typically rely on actual data for annotation. To better define this metadata-only task, we introduced the term Column Vocabulary Association (CVA). As previously described in section 2.3, CVA involves annotating columns using only KGs and table metadata, without accessing the underlying data. This approach is especially relevant in scenarios where the data is confidential and inaccessible.</p><p>In a prior research <ref type="bibr" target="#b30">[31]</ref>, we investigated a similar concept, using a Large-Context Window approach to enhance the model's performance, instead of a RAG system. We primarily focused on whether we could use, in principle, LLMs to annotate column headers with an external vocabulary without any use of the underlying data, and whether contextual and hierarchical information had any impact in performance. In that work we proposed three metrics for evaluation. The first -"LLM Internal Consistency" -assesses how consistently LLMs perform the annotation. The second metric -"Inter-LLMs Alignment" -evaluates how multiple LLMs annotate the same column headers. The third metric -"Human-Computer Agreement" -involves comparing LLM-generated annotations against human annotations. Importantly, we did not treat human annotations as the groundtruth or gold standard. Instead, we focus on comparing the variance among human annotations with those produced my the LLMs. In this approach, we highlight that humans can also disagree on finding the most suitable annotations, in the same way LLMs can. Also, all our metrics involved performing the same annotations task multiple times, allowing us to measure the consistency of the LLM's outputs over several iterations rather than just once. The goal of this year SemTab challenge and the "Metadata to KG" track align with our previous research, as creating annotations is often challenging, requiring domain expertise and difficult to automate.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1.">Metadata to KG -Round 1</head><p>Round 1 of the "Metadata to KG" track required participants to map a set of table metadata to DBpedia properties. Participants were provided with tables metadata and DBpedia properties files in both JSONL and OWL formats, all of which were accessible in the following GitHub repository <ref type="foot" target="#foot_3">4</ref> . The tables metadata file included information about 141 columns derived from different tables. For each column, the provided information included the column ID, column label, table ID, table name, and a list of the other column labels within the same table. The DBpedia properties file contained 2,881 properties. For each DBpedia property, the information included the property ID (the actual URI of the property in DBpedia), the property label, and the description. Below, we report examples of a table metadata entry (Listing 1) and of DBpedia property (Listing 2). § ⊵ { "id": "58891288_0_1117541047012405958_Director(s)", "label": "Director(s)", "table_id": "58891288_0_1117541047012405958", "table_name": "Film", "table_columns": ["Rank", "Title", "Year", "Director(s)", "Overall Rank"] } ¦ ¥</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Listing 1: Example of table metadata § ⊵</head><p>{ "id": "http://dbpedia.org/ontology/director", "label": "film director", "desc": "A film director is a person who directs the making of a film." }</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>¦ ¥</head><p>Listing 2: Example of DBpedia property Additionally, the SemTab organizers supplied a sample table metadata file and a sample ground truth (which contains only 9 metadata entries and their corresponding annotations), along with a Python evaluation script. The objective was to develop approaches for mapping each table metadata with up to 5 DBpedia properties for each column, based on semantic similarities and relevance, and then rank the mappings from the most to least accurate. The evaluation script assessed the mappings by calculating two metrics: hit@1, which checks if the first mapping is correct, and hit@5, which checks if the correct mapping is within the top five. Participants tested their systems on sample metadata (containing only 9 columns) and submitted their complete results to the track organisers for evaluation against the overall ground truth.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.">Metadata to KG -Round 2</head><p>Round 2 of the "Metadata to KG" track introduced a level of complexity by using a collection of custom vocabularies for the mapping task. Participants were again provided with tables metadata and custom vocabularies files in JSONL and OWL formats, accessible in the following GitHub repository <ref type="foot" target="#foot_4">5</ref> . In this round, the tables metadata file contained 1181 entries (one entry corresponds to one column) from various datasets, with each column having the same information as in round 1 included: column ID, column label, table ID, table name, and the other columns labels. The custom vocabularies file consisted of 1192 entries, where each entry had, again, the same information as in round 1: ID (in this case not a URI, but a minted ID), label and description. The tables metadata included a very diverse set of topics: including but not limited to: COVID-19 clinical trials, Indian movies ratings and Saudia Arabia stock exchange data. As in Round 1, the SemTab organizers supplied a sample table metadata file and a sample ground truth for validation and testing purposes (containing 11 metadata entries and their corresponding annotations), with the Python evaluation script. As before, the objective was to map each table column metadata to up to 5 relevant custom vocabulary terms based on semantic similarities and relevance, and rank these mappings by accuracy. The evaluation script again used hit@1 and hit@5 metrics to assess the quality of the mappings.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Methods</head><p>Here we outline our methodology, which considers the CVA task as a textual information retrieval challenge. Given that most of the table metadata and glossary information are described in text, the goal is to retrieve the most similar glossary entries from the glossary files based on the table metadata only.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1.">Implementation details</head><p>Our approach combines two main methods: one leveraging LLMs (both open-source and commercial) and the other utilizing a traditional semantic similarity technique. We tested several LLMs, including three GPT models (gpt-3.5-turbo-0125, gpt-4o and gpt-4-turbo), two LLama models (llama3-70b and llama3-8b), a Gemma model (gemma-7b) and a Mixtral model (mixtral-8x7b). To explore how LLMs' creativity impacts performance on our task, we experimented with varying temperature settings (0.5, 0.75, 1.0, 1.25, and 1.5) for each model. In general, temperature is a key hyperparameter in LLMs that controls the randomness of generated outputs, with higher temperature typically producing more and diverse responses, and lower temperatures more deterministic ones. In this study, we only changed the temperature setting to focus on one parameter at the time and explore wheter the creativity or determinism of the LLMs had an impact on task performance. Other parameters, such as top-k and top-p sampling, which control token selection probability and distribution, were kept at default values. While these additional parameters may also affect performance, investigating them was out of scope of this paper and could be explored in future research. Further, we utilized RAG systems to enhance the LLMs' performance. Specifically, we employed two RAG systems. The first is OpenAI's RAG, which uses the OpenAI Assistant API with a built-in file search tool to process and retrieve relevant data. The second is an open-source setup with LlamaIndex, ChromaDB, and Groq, where LlamaIndex integrates the data, ChromaDB serves as a vector database, and Groq accelerates inference with custom hardware.</p><p>For our semantic similarity method, we implemented SentenceBERT <ref type="bibr" target="#b31">[32]</ref> using the model all-MiniLM-L6-v2 to calculate cosine similarity scores between sentence embeddings. Here, we embedded both the metadata and glossary information, computing the cosine similarity for all possible pairs and selecting the glossary term with the highest similarity score for each metadata entry. We explored whether the choice of information embedded (e.g. table name, column headers, glossary term description) impacted similarity results, experimenting with various configurations to determine the most effective approach.</p><p>All experiments with LLMs were conducted in a zero-shot setting, meaning the models received no fine-tuning and were provided no examples through prompts or assistant instructions. This approach is a key feature of our methodology. We chose this strategy because we aim to develop a method that is domain-agnostic. Fine-tuning models with specific examples from particular mappings and vocabularies could bias the reported accuracy towards that specific domain. This is particularly relevant in scenarios where a large amount of datasets requires annotation, or where we are dealing with a new vocabulary without ground truth or suitable examples to provide. While a few-shot approach could also be a reasonable solution in cases where domain experts provide examples or ground truth, our focus in this work is on the more naive zero-shot setting. This allows us to propose a method that can be applied more in general settings, regardless of the data domain or vocabulary.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2.">CVA with LLMs</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Prompt Engineering</head><p>Through trial and error, we developed effective prompts, both user queries and assistant instructions. We found that repeating some information from the assistant instructions within the prompt resulted in more precise results by ensuring the models only used the data we provided, thus minimizing hallucinations. Both the prompt and instructions specified to return the 5 most similar glossary entries for each metadata. Below, we show the instructions given to the assistants and the query template for the user prompt used in Round 1 of the SemTab Challenge. The instructions and prompts for Round 2, which are quite similar, can be found on the GitHub page <ref type="foot" target="#foot_5">6</ref> .</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>ASSISTANT INSTRUCTIONS</head><p>Your task is to match column metadata to DBpedia properties. The full set of DBpedia properties will be provided in the vector. Columns metadata, instead, will be provided by the user and it will contain the following information: column ID, column label, table ID, table name and the labels of the other columns within that table. The matching between the column and the DBpedia properties is to be made based on the semantic similarities between the metadata (i.e. what the column express), and DBpedia properties. You can add multiple properties, but no more 5. Return the results in the following format: 'colID': '00000_0_0000_XXX', 'propID': ['http://dbpedia.org/ontology/PROPERTY_ID', ..., 'http://dbpedia.org/ontology/PROPERTY_ID']. Sort the matched DBpedia in descending order of relevance, starting with the most relevant. Choose ONLY from the DBpedia properties. Return ONLY the results, no other text. Return results for each and every single column metadata.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>QUERY TEMPLATE</head><p>Based on the instruction given to you, find the most relevant DBpedia property, for each of the following metadata in json format: {input_metadata} Each json element is an independent column metadata. The metadata do not have any relationship, so the matching with the DBpedia properties should only be based on the information provided within its own metadata. You can add multiple properties, but no more 5. Return the results in the following format: 'colID': '00000_0_0000_XXX', 'propID': ['http://dbpedia.org/ontology/PROPERTY_ID', ..., 'http://dbpedia.org/ontology/PROPERTY_ID']. Sort the matched DBpedia in descending order of relevance, starting with the most relevant. Choose ONLY from the DBpedia properties provided in the vector. Return ONLY the results, no other text. Return results for each and every single column metadata.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Model-temperature selection</head><p>We ran the queries three times for each LLM and temperature combination, then evaluated the preliminary performance using an evaluation script and groundtruth provided by the organisers. Based on these results, we selected the best-performing LLM-temperature combination to compute results on the full dataset.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>CVA on full metadata set</head><p>In the first round, we built a vector representation of the complete glossary JSON file containing the DBpedia properties through the RAG system. We then processed the metadata entries in batches of 25 when using the OpenAI API. For the open-source LLMs, instead, each metadata entry was added individually. While processing each metadata individually may lead to better performance, we implemented this batching strategy primarily to reduce costs associated with running some of the more expensive OpenAI models (i.e. GPT-4o and GPT-4-turbo). In the second round, given the larger size of the glossary, we split it into smaller, topic-based glossaries. We created 75 smaller glossary files and divided the full metadata set into 75 corresponding files. Each metadata file was then processed one at a time against the vector containing the 75 glossary files.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.3.">Semantic similarity using SentenceBERT</head><p>Our second method involved computing the semantic similarity between table metadata and glossary entries using SentenceBERT <ref type="bibr" target="#b31">[32]</ref>. First, we generated a vector representation for each metadata and glossary entry. Next, we calculated the cosine similarity between the embedding of each table metadata and the glossary entries to identify the top 5 glossary entries with the highest cosine similarity scores.</p><p>The initial steps of this method posed two challenges: first, selecting the appropriate table metadata and glossary information for vector generation, and second, determining the optimal approach for vectorizing the textual content. Specifically, we needed to decide whether to concatenate all textual elements before vectorization or to vectorize each component separately and then combine the vectors to create a final embedding. To address these questions, we experimented with various combinations of textual information, namely metadata column headers and table names, and glossary term labels and descriptions. We computed the cosine similarities for each combination and evaluated the results against the ground truth using the evaluation script provided by the organizers. Based on these analyses, we selected the most effective combinations for application across the complete metadata set.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.4.">Evaluation</head><p>The evaluation was conducted using a script provided by the track organizers. This script computed the accuracy of the generated mappings for a sample metadata file and a sample ground truth. It calculated two metrics: hit@1 and hit@5. To reiterate, users are supposed to generate the 5 most relevant mapping between the table metadata and the glossary, sorted from the most relevant. Hit@1 checks if the first mapping (thus the one considered to be the most relevant) is correct, while hit@5 checks if the correct mapping is among the top five results. Participants were then asked to generate the mappings for the entire table metadata file and submit them to the organizers. The organizers can then run the evaluation script again using the complete ground truth, which has not yet been shared with the participants.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Results</head><p>In the following sections, we present the preliminary challenge's results. These results show the accuracy scores obtained from the evaluation script on the sample metadata and the sample groundtruth provided. At this point, we are not aware on how our methods performed for the full set of metadata file, as the complete groundtruth has not yet been provided to participants.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.1.">CVA with LLMs</head><p>Here we present the results from our initial analysis using different LLMs and temperature settings. We employed three models from OpenAI and four open-source models, testing them at five different temperatures, as detailed in table 1. Table <ref type="table">1</ref> shows the average accuracy results for each model-temperature combination, evaluated using the evaluation script with the sample metadata and sample ground truth. Each query was run three times per model-temperature combination, and accuracy results were then averages. The numbers in bold correspond to the best-performing model-temperature combinations. gpt-4o outperformed other models in both Rounds 1 and 2, specifically at temperatures 0.5, 0.75, and 1.0. We observed that the LLMs did not perform very well in Round 2. In the discussion section6 we explore possible reasons for this outcome. Based on these preliminary results, for Round 1, we used gpt-4o at temperatures 0.5, 0.75, and 1.0 on the full metadata file for final analysis. For Round 2, we used gpt-4o only at temperatures 0.5 and 0.75.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Table 1</head><p>Results of different models for sample data in Round 1 and 2. The cells with the "X" refers to tries where the LLM could not compute the task, and either the API was not returning any results over a long period of time or, in the case of gemma-7b the model was returning "failure" message. In bold, instead, we show the best performing results. In the table h1 and h5 refers to Hit@1 and Hit@5 metrics from the evaluation script. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>LLM</head><formula xml:id="formula_0">X gpt-4-turbo 0 0 X X X X X X X X llama3-8b 0 0 0 0 0 0 0 0 0 0 llama3-70b 0 0 0 0 0 0 0 0 0 0 gemma-7b X X X X X X X X X X mixtral-8x7b 0 0 0 0 0 0 0 0 0 0</formula></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.2.">CVA with SentenceBERT</head><p>Below we show the results from our initial analysis with SentenceBERT. Table <ref type="table" target="#tab_1">2</ref> includes the possible combinations of information from the table metadata and the glossary, and the accuracy results for both Round 1 and 2, which we obtained by running the evaluation script against the ground truth for the sample metadata file. We used these results to find the best performing combinations, which were then applied to the full metadata file.</p><p>In Round 1, we did not have a single combination that performed best for both hit@1 and hit@5. The best hit@1 (0.56) is obtained when we use the column label and/or table name to represent the table metadata embedding, and use the property label for the DBpedia property embedding. The best hit@5 (0.67) is obtained when we use the sum of the vectors of the column label and the table name as the table metadata embedding, and use the vector of the property label as the DBpedia property embedding.</p><p>For Round 2, the best hit@1 and hit@5 are both obtained when we use the sum of the vectors of the column label and the table name as the table metadata embedding, and encode the vocabulary description as the vocabulary embedding. Based on this results, we did perform SentenceBERT on the full data for Round 1. For Round 2, instead, we sent the results from SentenceBERT for final analysis using the setting with the sum of the vectors of the column label and the table name as the table metadata embeddings. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.3.">Interim results on full metadata set</head><p>The results presented above are based on the test metadata provided by the track organizers. These results reflect the performance of our methods solely on the test metadata, which we used for selecting the most effective approaches for application to the complete metadata set. After, the full set of results was submitted and evaluate by the organizers against the complete groundtruth set (which is not accessible to participants). The organizers subsequently provided us with interim results, shown in Table <ref type="table" target="#tab_2">3</ref>.</p><p>In Round 1, the model that performed best on the test data was GPT-4o, with temperature settings of 0.5, 0.75, and 1.0. This model achieved a top accuracy of 0.89 on the test data, as illustrated in Table <ref type="table">1</ref>. However, when applied to the full dataset, our performance dropped to 70%. In Round 2, the best approaches based on the test data were the traditional semantic similarity method (SentenceBERT) and GPT-4o with temperatures of 0.5 and 0.75. The GPT models reached an accuracy of 100% on the test data, while SentenceBERT achieved 91%. Nevertheless, the performance on the full dataset was lower, with SentenceBERT reaching only 68% accuracy and GPT models achieving up to 52%.</p><p>While we have received these interim results, the complete results and ground truth data have not yet been provided to participants. One potential reason for the lower performance on the full dataset could be the quality of the test data, which may include mapping that are "impossible" or particularly challenging even for humans, especially since there is no human performance baseline to compare against. Additionally, the discrepancy between the interim results and test performance may suggest that the test data is not fully representative of the entire dataset, which could further explain the variations in performance. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.">Conclusion</head><p>Our analysis provides several key observations. Firstly, we found that repeating phrases in both the assistant instructions and user prompts improved the LLM's adherence and reduced hallucinations by preventing irrelevant entries. Also, the effectiveness of traditional semantic similarity methods can outperform that of more advanced techniques like LLMs and RAG, depending on the nature of the glossary and metadata. In fact, in Round 1, there was no clear link between the table metadata and the glossary, which consisted of DBpedia properties. However, we saw a strong connection in Round 2, likely because both the metadata and glossary were developed by the same institution. We think that these variations impacted the performance of the methods proposed in this study. LLMs, particularly gpt-4o, performed much better in Round 1, where leveraging the LLM's background knowledge was crucial for identifying the most relevant mappings. In round 2, instead, the more traditional semantic similarity method using SentenceBERT was sufficient and sometimes even outperformed LLMs. This improvement is attributed to the high degree of semantic similarity between the metadata and the glossary, as both likely originated from the same institution and were intentionally designed to be compatible and consistent. Regarding temperature settings for LLMs, we observed that lower temperatures generally led to better performance. Higher temperatures, particularly above 1.25, often resulted in no outputs or errors, especially with models like gpt-4-turbo and gemma-7b. This suggests that lower temperatures may be more effective for tasks requiring precise mapping.</p><p>It is also essential to note that the interim results from the full dataset show lower performance compared to results derived from the test data alone. Currently, we don't have yet the complete ground truth and results for further examination, as these remain with the task organizers. Our observations suggest that the test ground truth and data provided may not be suitable to assess our approaches for generalization to the full dataset. Only a small number of test results were provided in both rounds (9 ground truth metadata entries in Round 1 and 11 in Round 2), and we are uncertain about how the ground truth was generated (whether automatically, through extraction from resources, or manually). The evaluation's aim might need to focus more on comparing performance against human annotations, as parts of metadata annotation using external vocabularies can be challenging even for humans. This difficulty can happen if the vocabulary lacks suitable terms to describe the metadata or if a term is too complicated to explain clearly. In this context, trying to limit hallucinations might not be the best approach. In previous research <ref type="bibr" target="#b30">[31]</ref>, where we labeled column headers with terms from a specific vocabulary (CESSDA) <ref type="foot" target="#foot_6">7</ref> using also a zero-shot LLMs approach, we showed that hallucinations can indicate problems with the vocabulary used: columns about mental health were labeled "Mental Health", even though this topic wasn't included in CESSDA. These findings could help in creating better and more comprehensive vocabularies. Another consideration is that we utilized the same prompt for all LLMs. While we found that commercial models performed generally better, opensource LLMs might require different prompts than commercial ones, and further investigation is needed to address this.</p><p>In conclusion, our study examined methods for mapping column headers to glossaries in a zero-shot setting. This approach allows for broader evaluation across domains. We introduced the concept of "Column Vocabulary Association" (CVA), distinguishing it from other Semantic Table <ref type="table">Interpretation</ref> (STI) tasks. Additionally, we defined the term of "Dataless Tables", referring to tables where the structure and metadata are available but the actual data is confidential and inaccessible. Our findings suggest that LLMs have good performance when there is no clear connection between metadata and glossaries (as seen in Round 1), while traditional methods perform better in cases where there is a strong relationship between them (Round 2).</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 2</head><label>2</label><figDesc>Results of various embedding combinations for sample data in Round 1 and 2 are presented, with best performing results in bold. The table includes Hit@1 (h1) and Hit@5 (h5) metrics from evaluation script.</figDesc><table><row><cell>Metadata</cell><cell>Glossary</cell><cell cols="2">Round 1</cell><cell cols="2">Round 2</cell></row><row><cell>Embeddings</cell><cell>Embeddings</cell><cell>h1</cell><cell>h5</cell><cell>h1</cell><cell>h5</cell></row><row><cell>encode(label)</cell><cell>encode(label)</cell><cell cols="4">0.56 0.56 0.36 0.55</cell></row><row><cell>encode(label)</cell><cell>encode(lable + desc)</cell><cell cols="4">0.22 0.56 0.45 0.82</cell></row><row><cell>encode(label + table_name)</cell><cell>encode(label)</cell><cell cols="4">0.56 0.56 0.09 0.27</cell></row><row><cell>encode(label + table_name)</cell><cell>encode(lable + desc)</cell><cell cols="4">0.33 0.44 0.64 0.73</cell></row><row><cell>encode(label)</cell><cell>encode(desc)</cell><cell cols="4">0.11 0.33 0.45 0.82</cell></row><row><cell>encode(label + table_name)</cell><cell>encode(desc)</cell><cell cols="4">0.22 0.44 0.64 0.73</cell></row><row><cell>encode(label) + encode(table_name)</cell><cell>encode(desc)</cell><cell>0</cell><cell cols="3">0.33 0.64 0.91</cell></row><row><cell cols="6">encode(label) + encode(table_name) encode(desc) + encode(label) 0.22 0.44 0.55 0.91</cell></row><row><cell>encode(label) + encode(table_name)</cell><cell>encode(label)</cell><cell cols="4">0.44 0.67 0.27 0.45</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>Table 3</head><label>3</label><figDesc>Interim results on full data set in Round 1 and 2 are presented, with the best performing results in bold. * The embedding settings for this approach included the following: metadata embeddings (computed as encode(column label) + encode(table name)) and glossary term embeddings (computed as encode(term description)).</figDesc><table><row><cell>Approach</cell><cell>Specifics</cell><cell>Round 1</cell><cell></cell></row><row><cell></cell><cell></cell><cell>h1</cell><cell>h5</cell></row><row><cell>gpt-4o</cell><cell>0.50 (temp)</cell><cell>0.50</cell><cell>0.50</cell></row><row><cell>gpt-4o</cell><cell>0.50 (temp)</cell><cell>0.49</cell><cell>0.49</cell></row><row><cell>gpt-4o</cell><cell>0.75 (temp)</cell><cell>0.55</cell><cell>0.70</cell></row><row><cell>gpt-4o</cell><cell>1.00 (temp)</cell><cell>0.53</cell><cell>0.56</cell></row><row><cell></cell><cell></cell><cell>Round 2</cell><cell></cell></row><row><cell></cell><cell></cell><cell>h1</cell><cell>h5</cell></row><row><cell>gpt-4o</cell><cell>0.50 (temp)</cell><cell>0.49</cell><cell>0.52</cell></row><row><cell>gpt-4o</cell><cell>0.75 (temp)</cell><cell>0.45</cell><cell>0.49</cell></row><row><cell>SentenceBert</cell><cell>*</cell><cell>0.37</cell><cell>0.68</cell></row></table></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_0">https://www.cbs.nl</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_1">https://data.gov</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_2">https://search.open.canada.ca/data/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="4" xml:id="foot_3">https://github.com/sem-tab-challenge/2024/blob/main/data/metadata2kg/round1/README.md</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="5" xml:id="foot_4">https://github.com/sem-tab-challenge/2024/blob/main/data/metadata2kg/round2/README.md</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="6" xml:id="foot_5">https://github.com/sem-tab-challenge/2024/blob/main/data/metadata2kg/round1/README.md</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="7" xml:id="foot_6">https://vocabularies.cessda.eu/vocabulary/TopicClassification</note>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgments</head><p>We acknowledge that ChatGPT was utilized to generate and debug part of the python and latex code utilised in this work. This work is funded by the Netherlands Organisation of Scientific Research (NWO), ODISSEI Roadmap project: 184.035.014.</p><p>Author M.M. led the research by developing the main ideas, conceptualising the research design, participating in the programming tasks, reviewing the literature and drafting most of the manuscript. Author X.P. provided most of the technical expertise, particularly in programming and APIs integration to perform our analysis, and has helped in reviewing the manuscript. Authors B.K., T.K and J.v.O. provided guidance through their supervisory roles and gave feedback to improve the overall study.</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Tabular data: Deep learning is not all you need</title>
		<author>
			<persName><forename type="first">R</forename><surname>Shwartz-Ziv</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Armon</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Information Fusion</title>
		<imprint>
			<biblScope unit="volume">81</biblScope>
			<biblScope unit="page" from="84" to="90" />
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">The fair guiding principles for data stewardship: fair enough?</title>
		<author>
			<persName><forename type="first">M</forename><surname>Boeckhout</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">A</forename><surname>Zielhuis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">L</forename><surname>Bredenoord</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">European journal of human genetics</title>
		<imprint>
			<biblScope unit="volume">26</biblScope>
			<biblScope unit="page" from="931" to="936" />
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<monogr>
		<author>
			<persName><forename type="first">B</forename><surname>Mons</surname></persName>
		</author>
		<title level="m">Data stewardship for open science: Implementing FAIR principles</title>
				<imprint>
			<publisher>Chapman and Hall/CRC</publisher>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Towards fair principles for research software</title>
		<author>
			<persName><forename type="first">A.-L</forename><surname>Lamprecht</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Garcia</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Kuzak</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Martinez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Arcila</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><forename type="middle">Martin</forename><surname>Del Pico</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><forename type="middle">Dominguez</forename><surname>Del Angel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Van De Sandt</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Ison</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><forename type="middle">A</forename><surname>Martinez</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Data Science</title>
		<imprint>
			<biblScope unit="volume">3</biblScope>
			<biblScope unit="page" from="37" to="59" />
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">The fair guiding principles for scientific data management and stewardship</title>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">D</forename><surname>Wilkinson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Dumontier</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><forename type="middle">J</forename><surname>Aalbersberg</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Appleton</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Axton</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Baak</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Blomberg</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J.-W</forename><surname>Boiten</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><forename type="middle">B</forename><surname>Da Silva Santos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><forename type="middle">E</forename><surname>Bourne</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Scientific data</title>
		<imprint>
			<biblScope unit="volume">3</biblScope>
			<biblScope unit="page" from="1" to="9" />
			<date type="published" when="2016">2016</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Aligning restricted access data with fair: a systematic review</title>
		<author>
			<persName><forename type="first">M</forename><surname>Martorana</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Kuhn</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Siebes</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Van Ossenbruggen</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">PeerJ Computer Science</title>
		<imprint>
			<biblScope unit="volume">8</biblScope>
			<biblScope unit="page">e1038</biblScope>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Retrieval-augmented generation for knowledgeintensive nlp tasks</title>
		<author>
			<persName><forename type="first">P</forename><surname>Lewis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Perez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Piktus</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Petroni</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Karpukhin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Goyal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Küttler</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Lewis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>-T. Yih</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Rocktäschel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Riedel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Kiela</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 34th International Conference on Neural Information Processing Systems, NIPS &apos;20</title>
				<meeting>the 34th International Conference on Neural Information Processing Systems, NIPS &apos;20<address><addrLine>Red Hook, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>Curran Associates Inc</publisher>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Open government data portals: Predictors of site engagement among early users of health data ny</title>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">M</forename><surname>Begany</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><forename type="middle">G</forename><surname>Martin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><forename type="middle">J</forename><surname>Yuan</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Government Information Quarterly</title>
		<imprint>
			<biblScope unit="volume">38</biblScope>
			<biblScope unit="page">101614</biblScope>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Distributed learning on 20 000+ lung cancer patientsthe personal health train</title>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">M</forename><surname>Deist</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><forename type="middle">J</forename><surname>Dankers</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Ojha</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">S</forename><surname>Marshall</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Janssen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Faivre-Finn</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Masciocchi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Valentini</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Chen</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Radiotherapy and Oncology</title>
		<imprint>
			<biblScope unit="volume">144</biblScope>
			<biblScope unit="page" from="189" to="200" />
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Advancing data sharing and reusability for restricted access data on the web: introducing the dataset-variable ontology</title>
		<author>
			<persName><forename type="first">M</forename><surname>Martorana</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Kuhn</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Siebes</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Van Ossenbruggen</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 12th Knowledge Capture Conference 2023</title>
				<meeting>the 12th Knowledge Capture Conference 2023</meeting>
		<imprint>
			<date type="published" when="2023">2023</date>
			<biblScope unit="page" from="83" to="91" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Language models are unsupervised multitask learners</title>
		<author>
			<persName><forename type="first">A</forename><surname>Radford</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Wu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Child</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Luan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Amodei</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Sutskever</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">OpenAI blog</title>
		<imprint>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="page">9</biblScope>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Large language models are zero-shot reasoners</title>
		<author>
			<persName><forename type="first">T</forename><surname>Kojima</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">S</forename><surname>Gu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Reid</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Matsuo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Iwasawa</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Advances in neural information processing systems</title>
		<imprint>
			<biblScope unit="volume">35</biblScope>
			<biblScope unit="page" from="22199" to="22213" />
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Chatgpt outperforms crowd workers for text-annotation tasks</title>
		<author>
			<persName><forename type="first">F</forename><surname>Gilardi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Alizadeh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Kubli</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Proceedings of the National Academy of Sciences</title>
		<imprint>
			<biblScope unit="volume">120</biblScope>
			<biblScope unit="page">e2305016120</biblScope>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">Column property annotation using large language models</title>
		<author>
			<persName><forename type="first">K</forename><surname>Korini</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Bizer</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the ESWC Conference</title>
				<meeting>the ESWC Conference</meeting>
		<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">Chatgraph: Interpretable text classification by converting chatgpt knowledge to graphs</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Shi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Ma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Zhong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><surname>Tan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Mai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Huang</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">2023 IEEE International Conference on Data Mining Workshops (ICDMW), IEEE</title>
				<imprint>
			<date type="published" when="2023">2023</date>
			<biblScope unit="page" from="515" to="520" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<monogr>
		<author>
			<persName><forename type="first">H</forename><surname>He</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Roth</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2301.00303</idno>
		<title level="m">Rethinking with retrieval: Faithful large language model inference</title>
				<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b16">
	<monogr>
		<author>
			<persName><forename type="first">G</forename><surname>Marcus</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2002.06177</idno>
		<title level="m">The next decade in ai: four steps towards robust artificial intelligence</title>
				<imprint>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">Factual error correction for abstractive summarization models</title>
		<author>
			<persName><forename type="first">M</forename><surname>Cao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Dong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Wu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">C K</forename><surname>Cheung</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP)</title>
				<meeting>the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP)</meeting>
		<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="6251" to="6258" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<analytic>
		<title level="a" type="main">The curious case of hallucinations in neural machine translation</title>
		<author>
			<persName><forename type="first">V</forename><surname>Raunak</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Menezes</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Junczys-Dowmunt</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/2021.naacl-main.92</idno>
		<ptr target="https://aclanthology.org/2021.naacl-main.92.doi:10.18653/v1/2021.naacl-main.92" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Association for Computational Linguistics</title>
				<editor>
			<persName><forename type="first">K</forename><surname>Toutanova</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Rumshisky</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">L</forename><surname>Zettlemoyer</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">D</forename><surname>Hakkani-Tur</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">I</forename><surname>Beltagy</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">S</forename><surname>Bethard</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">R</forename><surname>Cotterell</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">T</forename><surname>Chakraborty</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">Y</forename><surname>Zhou</surname></persName>
		</editor>
		<meeting>the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Association for Computational Linguistics</meeting>
		<imprint>
			<date type="published" when="2021">2021</date>
			<biblScope unit="page" from="1172" to="1183" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<analytic>
		<title level="a" type="main">Survey of hallucination in natural language generation</title>
		<author>
			<persName><forename type="first">Z</forename><surname>Ji</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Lee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Frieske</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Yu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Su</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Xu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Ishii</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><forename type="middle">J</forename><surname>Bang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Madotto</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Fung</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">ACM Computing Surveys</title>
		<imprint>
			<biblScope unit="volume">55</biblScope>
			<biblScope unit="page" from="1" to="38" />
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<analytic>
		<title level="a" type="main">Are ChatGPT and GPT-4 general-purpose solvers for financial text analytics? a study on several typical tasks</title>
		<author>
			<persName><forename type="first">X</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Chan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Zhu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Pei</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Ma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Shah</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/2023.emnlp-industry.39</idno>
		<ptr target="https://aclanthology.org/2023.emnlp-industry.39.doi:10.18653/v1/2023.emnlp-industry.39" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing: Industry Track, Association for Computational Linguistics</title>
				<editor>
			<persName><forename type="first">M</forename><surname>Wang</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">I</forename><surname>Zitouni</surname></persName>
		</editor>
		<meeting>the 2023 Conference on Empirical Methods in Natural Language Processing: Industry Track, Association for Computational Linguistics<address><addrLine>Singapore</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2023">2023</date>
			<biblScope unit="page" from="408" to="422" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b21">
	<monogr>
		<author>
			<persName><forename type="first">X</forename><surname>Shen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Backes</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Zhang</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2304.08979</idno>
		<title level="m">In chatgpt we trust? measuring and characterizing the reliability of chatgpt</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b22">
	<monogr>
		<author>
			<persName><forename type="first">Y</forename><surname>Gao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Xiong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Gao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Jia</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Pan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Bi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Dai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Sun</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Wang</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2312.10997</idno>
		<title level="m">Retrieval-augmented generation for large language models: A survey</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b23">
	<analytic>
		<title level="a" type="main">Active retrieval augmented generation</title>
		<author>
			<persName><forename type="first">Z</forename><surname>Jiang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><forename type="middle">F</forename><surname>Xu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Gao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Sun</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Dwivedi-Yu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Yang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Callan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Neubig</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing</title>
				<meeting>the 2023 Conference on Empirical Methods in Natural Language Processing</meeting>
		<imprint>
			<date type="published" when="2023">2023</date>
			<biblScope unit="page" from="7969" to="7992" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b24">
	<analytic>
		<title level="a" type="main">Retrieval augmented language model pre-training</title>
		<author>
			<persName><forename type="first">K</forename><surname>Guu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Lee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Tung</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Pasupat</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Chang</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">International conference on machine learning</title>
				<meeting><address><addrLine>PMLR</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="3929" to="3938" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b25">
	<analytic>
		<title level="a" type="main">Improving language models by retrieving from trillions of tokens</title>
		<author>
			<persName><forename type="first">S</forename><surname>Borgeaud</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Mensch</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Hoffmann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Cai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Rutherford</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Millican</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">B</forename><surname>Van Den Driessche</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J.-B</forename><surname>Lespiau</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Damoc</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Clark</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">International conference on machine learning</title>
				<meeting><address><addrLine>PMLR</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2022">2022</date>
			<biblScope unit="page" from="2206" to="2240" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b26">
	<analytic>
		<title level="a" type="main">Docprompting: Generating code by retrieving the docs</title>
		<author>
			<persName><forename type="first">S</forename><surname>Zhou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">U</forename><surname>Alon</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><forename type="middle">F</forename><surname>Xu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Jiang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Neubig</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">The Eleventh International Conference on Learning Representations</title>
				<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b27">
	<monogr>
		<author>
			<persName><forename type="first">B</forename><surname>Peng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Galley</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>He</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Cheng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Xie</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Hu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><surname>Huang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Liden</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Yu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Chen</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2302.12813</idno>
		<title level="m">Check your facts and try again: Improving large language models with external knowledge and automated feedback</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b28">
	<analytic>
		<title level="a" type="main">Leveraging passage retrieval with generative models for open domain question answering</title>
		<author>
			<persName><forename type="first">G</forename><surname>Izacard</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Grave</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/2021.eacl-main.74</idno>
		<ptr target="https://aclanthology.org/2021.eacl-main.74.doi:10.18653/v1/2021.eacl-main.74" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 16th Conference of the European Chapter of the Association for Computational Linguistics: Main Volume, Association for Computational Linguistics</title>
				<editor>
			<persName><forename type="first">P</forename><surname>Merlo</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">J</forename><surname>Tiedemann</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">R</forename><surname>Tsarfaty</surname></persName>
		</editor>
		<meeting>the 16th Conference of the European Chapter of the Association for Computational Linguistics: Main Volume, Association for Computational Linguistics</meeting>
		<imprint>
			<date type="published" when="2021">2021</date>
			<biblScope unit="page" from="874" to="880" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b29">
	<analytic>
		<title level="a" type="main">Large language models with controllable working memory</title>
		<author>
			<persName><forename type="first">D</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">S</forename><surname>Rawat</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Zaheer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Lukasik</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Veit</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Yu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Kumar</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Findings of the Association for Computational Linguistics: ACL 2023</title>
				<imprint>
			<date type="published" when="2023">2023</date>
			<biblScope unit="page" from="1774" to="1793" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b30">
	<analytic>
		<title level="a" type="main">Zero-shot topic classification of column headers: Leveraging llms for metadata enrichment</title>
		<author>
			<persName><forename type="first">M</forename><surname>Martorana</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Kuhn</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Stork</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Van Ossenbruggen</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Knowledge Graphs in the Age of Language Models and Neuro-Symbolic AI</title>
				<imprint>
			<publisher>IOS Press</publisher>
			<date type="published" when="2024">2024</date>
			<biblScope unit="page" from="52" to="66" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b31">
	<monogr>
		<author>
			<persName><forename type="first">N</forename><surname>Reimers</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Gurevych</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1908.10084</idno>
		<title level="m">Sentence-bert: Sentence embeddings using siamese bert-networks</title>
				<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
