<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Scalable Table-to-Knowledge Graph Matching from Metadata using LLMs</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Nathan</forename><surname>Vandemoortele</surname></persName>
							<email>nathan.vandemoortele@ugent.be</email>
							<affiliation key="aff0">
								<orgName type="institution">Ghent University -imec</orgName>
								<address>
									<addrLine>Technologiepark-Zwijnaarde 122</addrLine>
									<postCode>9052</postCode>
									<settlement>Gent</settlement>
									<country key="BE">Belgium</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Bram</forename><surname>Steenwinckel</surname></persName>
							<email>bram.steenwinckel@ugent.be</email>
							<affiliation key="aff0">
								<orgName type="institution">Ghent University -imec</orgName>
								<address>
									<addrLine>Technologiepark-Zwijnaarde 122</addrLine>
									<postCode>9052</postCode>
									<settlement>Gent</settlement>
									<country key="BE">Belgium</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Sofie</forename><surname>Van Hoecke</surname></persName>
							<email>sofie.vanhoecke@ugent.be</email>
							<affiliation key="aff0">
								<orgName type="institution">Ghent University -imec</orgName>
								<address>
									<addrLine>Technologiepark-Zwijnaarde 122</addrLine>
									<postCode>9052</postCode>
									<settlement>Gent</settlement>
									<country key="BE">Belgium</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Femke</forename><surname>Ongenae</surname></persName>
							<email>femke.ongenae@ugent.be</email>
							<affiliation key="aff0">
								<orgName type="institution">Ghent University -imec</orgName>
								<address>
									<addrLine>Technologiepark-Zwijnaarde 122</addrLine>
									<postCode>9052</postCode>
									<settlement>Gent</settlement>
									<country key="BE">Belgium</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Scalable Table-to-Knowledge Graph Matching from Metadata using LLMs</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">C138007B04597B5B59E95F2C82C33918</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T19:03+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>data linkage, knowledge graphs, large language models, retrieval augmented generation, zero-shot learning, metadata Orcid 0009-0009-0118-3519 (N. Vandemoortele)</term>
					<term>0000-0002-3488-2334 (B. Steenwinckel)</term>
					<term>0000-0002-7865-6793 (S. Van Hoecke)</term>
					<term>0000-0003-2529-5477 (F. Ongenae)</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Addressing the challenge of interoperability when integrating and interpreting large datasets from diverse sources is essential for businesses aiming to make informed, data-driven decisions. Therefore, the 2024 Semantic Web Challenge on Tabular Data to Knowledge Graph Matching (SemTab) focuses on using metadata, such as column names, to map tables to semantic concepts within standardized vocabularies or Knowledge Graphs (KGs). The challenge involves mapping two datasets, one to the DBpedia ontology and the other to a custom vocabulary. Our approach begins with applying Retrieval-Augmented Generation (RAG) for a broad search of relevant matches, ensuring scalability. We then refine these matches using a Large Language Model (LLM) with Chain-of-Thought (CoT) prompting and Self-Consistency (SC). Finally, we combine the results using Reciprocal Rank Fusion (RRF) to obtain a final ranking of the matches. This method achieves hit rates of 62% (top 1) and 82% (top 5) for the first dataset, and 84% (top 1) and 98% (top 5) for the second. The LLM's strong semantic understanding and extensive knowledge base provide significant advantages over traditional human labeling, which is often laborious and time-consuming. Furthermore, the LLM's zero-shot capability removes the need for additional task-specific training data, making this solution applicable across various domains. Despite limitations like computational costs and the need for well-defined concepts in the ontology or vocabulary, our approach remains cost-effective compared to extensive human labeling. Moreover, we leave room to trade performance for scalability if needed, pending further research.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>In today's data-driven world, many organizations and companies frequently face the challenge of integrating and interpreting large datasets from various sources to streamline their business operations and their ability to make data-driven decisions <ref type="bibr" target="#b0">[1]</ref>. For example, consider a multinational corporation that has recently acquired several smaller companies, each with a different data management system. Merging such heterogeneous data sources while maintaining the intrinsic semantic meaning of the available information is a difficult task for humans. confidentia table_name": "≈", "table_columns": ["RANG", "Museum", "Stadt", " <ref type="bibr">Facebook</ref> One possible solution to this problem involves matching tables to standardized vocabularies or Knowledge Graphs (KGs) <ref type="bibr" target="#b1">[2,</ref><ref type="bibr" target="#b2">3]</ref>. This process, known as semantic table linking or table matching, aims to align data within tables with predefined concepts. The "Semantic Web Challenge on Tabular Data to Knowledge Graph Matching (SemTab) <ref type="foot" target="#foot_0">1</ref> " has been hosted for several years at the International Semantic Web Conference (ISWC) to stimulate the creation of such automated semantic annotation systems. The goal is to automate the assignment of a DBpedia or Wikidata entity to a whole column based on table data, assign a DBpedia or Wikidata entity to different cells, infer relations between different columns where possible, or assign a KG class to an entire table. One of the objectives of this year's challenge<ref type="foot" target="#foot_1">2</ref> is to map columns to semantic concepts in a vocabulary or KG based solely on the information in the header of the columns, i.e. column and table names. The objective is illustrated in Figure <ref type="figure">1</ref>.</p><p>This automated mapping presents considerable challenges, as the true meaning of the table relies on interpreting the data within the cells.</p><p>• For instance, a column named "ID" or "Value" provides little to no information about its semantic meaning to help linking it to a concept in a KG. • Many column names are ambiguous because they can have multiple potential meanings depending on the context. For example, a column named "Date" could refer to, amongst others, a birth date, transaction date, or event date. • Tables can employ different naming conventions, languages, abbreviations, or terminologies. For instance, one dataset might use "CustID" while another uses "CustomerID" to represent the same concept. • As organizations are dynamic, new tables can be added and existing ones modified. This dynamic nature requires continuous updates and maintenance of data structures, which is resource-intensive if done manually. • As organizations accumulate new tables, the scalability of semantic mapping becomes a concern. Mapping metadata to vocabularies or KGs for potentially thousands of tables is not humanly feasible, necessitating automated and scalable solutions.</p><p>We hypothesize that Large Language Models (LLMs) offer a promising solution to the challenges of semantic mapping when relying solely on metadata. Even in the absence of actual data content (e.g., due to restricted access), LLMs can leverage their intrinsic knowledge and reasoning capabilities to tackle automated mapping problems. In this paper, we investigate the zero-shot capabilities of GPT-4o<ref type="foot" target="#foot_2">3</ref> , a state-of-the-art LLM at the time of writing, for the "Table Metadata to KG" task in the 2024 "Semantic Web Challenge on Tabular Data to Knowledge Graph Matching" challenge. Zero-shot learning refers to the ability to transfer to an unseen problem without new task-specific training data. To remain cost-effective, we adapt an advanced Retrieval Augmented Generation (RAG) <ref type="bibr" target="#b3">[4]</ref> solution that retrieves only the most relevant information within the vocabulary or KG. This ensures that the generated responses are contextually accurate while maintaining scalability. In this paper, we present the following key contributions:</p><p>• While RAG solutions exist in the context of ontology matching, we are the first to incorporate and apply (LLM-based) rerankers in this context. • Our main innovation comprises a novel completion stage, where we uniquely combine two prompting techniques, Chain-of-Thought and Self-Consistency, with Reciprocal Rerank Fusion to find the best mappings. • Our methodology efficiently scales to handle large and diverse datasets in a zero-shot manner.</p><p>The remainder of this paper is as follows. We first discuss related work in Section 2. Next, we present the table datasets and vocabularies provided by the SemTab organizers in Section 3, followed by a detailed explanation of our methodology in Section 4. In Section 5, we present the results. We conclude the paper in Section 6 and suggest additional paths for future work.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Related Work</head><p>Tasks regarding table topic annotation and column type annotation are closely related to the general problem of semantic table understanding <ref type="bibr" target="#b4">[5]</ref>. The SemTab challenge <ref type="bibr" target="#b5">[6]</ref> has always been formulated as an unsupervised table annotation problem. Systems like MTab <ref type="bibr" target="#b6">[7]</ref>, CSV2KG <ref type="bibr" target="#b7">[8]</ref>, and DAGOBAH <ref type="bibr" target="#b8">[9]</ref>, which participated in the SemTab challenge in previous years, tackle three main tasks in KG matching: creating cell entity annotations (CEA), column type annotation (CTA), and paired column property annotations (CPA). These pipelines typically begin by linking individual cell contents, such as the various column entries, to ontology entities, and then predicting the most likely column type based on these linkages. To do so, these systems need available cell values to be mapped to KG entities before column types can be identified, which is not applicable in this context as here the actual table data is not available.</p><p>Given that we can only rely on metadata information for the task addressed in this paper, the related work methodologies that are applicable align more closely with ontology alignment techniques. These can range from traditional string similarity methods <ref type="bibr" target="#b9">[10]</ref> to semantic similarity measures utilizing resources like WordNet <ref type="bibr" target="#b10">[11]</ref>. WordNet is, however, designed for general purposes and may not include specialized terms or jargon used in specific, custom vocabularies. More recent work includes vector representations for semantic similarity, using, e.g., word2vec</p><p>to get promising results in semantic textual similarity tasks <ref type="bibr" target="#b11">[12]</ref>. Nowadays, also LLM-based approaches have achieved traction for tasks like ontology alignment <ref type="bibr" target="#b12">[13]</ref>, entity matching <ref type="bibr" target="#b13">[14]</ref> and subject annotation <ref type="bibr" target="#b14">[15]</ref>. Since LLMs are trained on large data corpora, they can identify complex relations and patterns between different objects in natural language, leading to more accurate matching. LLMs demonstrate their effectiveness to improve metadata matching performance, either through additional fine-tuning on table-based tasks <ref type="bibr" target="#b15">[16]</ref> or without fine-tuning <ref type="bibr" target="#b2">[3]</ref>, and show how contextual information, such as the dataset description, can be incorporated in an LLM prompt to match a vocabulary to a table's column topics <ref type="bibr" target="#b16">[17]</ref>. This last task is similar to the one we are trying to solve in this paper. However, providing the full vocabulary as context within each prompt would not be cost-effective at scale. To this end, RAG pipelines have been proposed <ref type="bibr" target="#b17">[18,</ref><ref type="bibr" target="#b18">19]</ref>, though they have not yet been applied in this specific context. Furthermore, advanced prompt optimization strategies remain largely unexplored.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Metadata Datasets</head><p>Two datasets were provided in the challenge, which both contained a vocabulary on the one hand and table metadata (i.e., table and column names) for multiple tables without the actual cell data on the other hand:</p><p>• Dataset 1 consists of metadata from a selected set of HTML-based tables that can be found on the web, and needs to get mapped to a subset of the DBpedia ontology <ref type="foot" target="#foot_3">4</ref> . The goal is to map each table column to a corresponding ontology property. While the ontology is hierarchically structured, it is treated as a simple vocabulary of properties, each defined by labels and short descriptions, with no consideration of the relationships between them. In total, metadata for 77 web tables were provided, of which 141 columns needed to be mapped. The vocabulary consists of 2881 different properties. • Dataset 2 contains a set of tables, available in a database accessible for public use.</p><p>The metadata of these tables needs to be mapped to a custom vocabulary with clear descriptions. However, the vocabulary is not derived from an existing publicly available KG. The goal is, again, to map each table column to one vocabulary item. This dataset contains metadata for 75 tables, of which 1181 columns needed to be mapped, and has 1181 items in the vocabulary.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Methodology</head><p>Our architecture adapts an advanced RAG solution with a novel LLM-based completion stage.</p><p>RAG is a technique used to incorporate new knowledge into LLMs through in-context learning (ICL), such as a vocabulary or a knowledge graph (KG). It promises a scalable approach by retrieving only relevant information, thereby significantly driving down costs. ICL enables LLMs to learn from examples within the context window without altering their weights. Our solution's two-stage pipeline is schematically visualized in Figure <ref type="figure" target="#fig_0">2</ref> and consists of a retrieval stage and a completion stage. In this section, we explain each component of the pipeline step by step. You can find the prompt templates used in the "appendix" directory of our GitHub repository: https://github.com/predict-idlab/Meta2Concept </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Retrieval Stage Completion Stage</head><p>Top 1 Top 5 </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Selfconsistency</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1.">Preprocessing</head><p>Both tables and vocabularies can vary in size, language, and quality. For example, the descriptions of the provided vocabulary items can be rather short (see Section 3), or the provided table metadata can contains abbreviations or be multilingual (see introduction). To address these variations, a preprocessing step is applied to both the vocabulary and metadata as a first step as detailed in the following subsections.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1.1.">Enriching of vocabulary</head><p>An item within a vocabulary represents the information of a single concept or property, e.g., a DBpedia property label and its description for the first dataset. We ask GPT-4o to translate the item labels to English, fix grammar mistakes, convert proper names to their type (e.g., Harelbeke becomes → City (Harelbeke)) and write acronyms in full. Closed-source LLMs like GPT-4o enrich the embedding model in the next stage, as they are more knowledgeable and frequently updated with the latest information. We process the vocabulary in chunks of 100 items per prompt, which is overall inexpensive and enables the use of GPT-4o. In most cases, this results in property labels becoming self-explanatory. Below is an example of vocabulary enrichment using GPT-4o:</p><formula xml:id="formula_0">Input: prompt(label='ept itm') Output: 'European Poker Tour (EPT) In The Money (ITM)'</formula><p>This helps generate more accurate embeddings for the next stage. In this case, you might also find matches closer to poker, tours, or money.</p><p>Item descriptions are more fine-grained representations of the label in semantic space. However, for these descriptions to be effective as embeddings, they must be specific and not open to interpretation. For example, "season" is simply described as "season", however "season" can refer to the yearly seasons but also television seasons, or sports seasons. Attempting to assign one interpretation without knowing the ground truth could skew the embeddings vector in a specific direction, whereas it might be preferable to leave such terms undefined at this stage. This problem is specific to Dataset 1, hence, we focus on enriching and utilizing only its labels. For Dataset 2, we use and retain the original descriptions, as they are deemed to be of high quality.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1.2.">Enriching of table metadata</head><p>The metadata consists of a table name and column names, one or more of which need to be mapped. We ask GPT-4o to transform the column and table name into a sentence in the form:</p><formula xml:id="formula_1">{column} {relation} {table_name}</formula><p>No other information about the other column names in the table is included. Furthermore, we ask GPT-4o to translate words to English, fix grammar mistakes, convert names to their type and write acronyms in full, similar to what we did with the vocabulary labels. The following is an example of table metadata enrichment using GPT-4o:</p><p>Input: prompt(column='Height (m)', table_name='Mountain') Output: 'Height in meters of a mountain'   </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2.">Retrieval Stage</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2.1.">Embedding model</head><p>Dense embeddings provide a fast way to retrieve relevant documents at scale. The main idea is to perform a coarse search for the relevant items within the vocabulary or KG at a low cost, which are then refined further downstream.</p><p>The process begins with indexing the enriched vocabulary, where it is chunked by item and embedded using a dense embedding model, as shown in Figure <ref type="figure" target="#fig_1">3</ref>. We use text-embedding-3large (OpenAI) <ref type="foot" target="#foot_4">5</ref> , which converts sentences into 3072-dimensional vectors. These embeddings are stored and can be reused for comparison with each new query. Next, the enriched table metadata is embedded similarly to generate query embeddings. We then perform a similarity search, comparing the embedded query with each item in the vocabulary using cosine similarity, which can be expressed as:</p><formula xml:id="formula_2">𝑐𝑜𝑠𝑖𝑛𝑒 𝑠𝑖𝑚𝑖𝑙𝑎𝑟𝑖𝑡𝑦 = 𝑑𝑜𝑡(𝐼 𝑡𝑒𝑚 𝑣𝑒𝑐𝑡𝑜𝑟 , 𝑄𝑢𝑒𝑟𝑦 𝑣𝑒𝑐𝑡𝑜𝑟 )</formula><p>Since the vectors are normalized, there is no need to divide by their lengths.</p><p>Finally, we rank the items according to the cosine similarity and retrieve the top 150 matches to be reranked. This number was chosen based on cost considerations. Ideally, this number should be empirically validated using a validation set; however, in this challenge, we were provided with only 10 samples for Dataset 1 and none for Dataset 2. Table <ref type="table" target="#tab_3">2</ref> indicates the number is sufficient for Dataset 1 with a 100% hit rate but may be inadequate for Dataset 2, which has a 95% hit rate. This likely stems from a larger vocabulary and a higher likelihood of encountering similar items. We believe there is a probable relationship between the size of the dataset and the optimal number of retrieval candidates, which we have yet to evaluate thoroughly.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2.2.">Reranker</head><p>The top 150 candidates identified in the previous step show high semantic similarity with the given table query, but their order is highly dependent on the embedding model. Embedding models have several drawbacks: both the vocabulary item and query are embedded independently, resulting in some loss of information. They are also often trained for specific tasks and therefore require further fine-tuning to achieve the best results. This goes against our goal of making the pipeline generically reusable. Furthermore, they do not have the same knowledge or reasoning capabilities of an LLM. This is where LLM-based rerankers come in. These rerankers find logical connections between several "passages" (our vocabulary items) and a query simultaneously, albeit at a higher computational cost, which is why we only use it on the top-ranked results in order to improve scalability. We use RankGPT <ref type="bibr" target="#b19">[20]</ref>, a library that provides a prompt template to transform LLMs into reranking agents. We selected Llama-3-70B (Meta), an open-weight LLM, to ensure cost-effectiveness given the size of our datasets. We utilized RankGPT's permutation generation with a sliding window (step size of 10 and window size of 20) to limit the number of candidates needing reranking in each prompt. Moreover, this measure decreases the prompt size and improves performance by mitigating context length sensitivity. RankGPT's sliding window is applied from back to front, allowing items to be promoted in rank but not demoted outside the window size. The reranking of vocabulary items is determined by their relevance to the following new query:</p><formula xml:id="formula_3">Query:</formula></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>What is the best description of column {column} in the table '{table_name}' with columns {table_columns}?</head><p>Notice that this query now uses all the metadata of the table. Similarly, items now contain all the information of the original vocabulary, i.e. the labels with their descriptions, allowing for a more fine-grained ranking. We no longer use the information obtained during enrichment because the clean-up step in preprocessing could introduce errors that may confuse the LLM. However, for Dataset 1, we found the descriptions to be poor, which is why we included the domain and range based on the properties provided by DBpedia <ref type="foot" target="#foot_5">6</ref> . If the domain is an "owl:Thing" or the range is a datatype, the field is left empty. This results in the following items:</p><formula xml:id="formula_4">Item: Dataset 1: [&lt;domain&gt;, &lt;property&gt;, &lt;range&gt;] &lt;description&gt; Dataset 2: &lt;label&gt;, &lt;description&gt;</formula><p>After reranking the items, we retrieve the top 30 candidates for further processing. This number was chosen based on similar limitations described in the initial ranking.  </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.3.">Completion Stage</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Selfconsistency</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.3.1.">Top-k matching</head><p>To obtain the final top 1 and top 5 matches, which is the goal of the challenge, we ask GPT-4o in zero-shot setting to rank the top 5 best vocabulary items out of the top 30 candidates. We selected GPT-4o (OpenAI) as the LLM of choice as it is the most capable model based on MMLU and Chatbot Arena scores, according to lmsys.org 7 .</p><p>We use Chain-of-Thought (CoT) <ref type="bibr" target="#b20">[21,</ref><ref type="bibr" target="#b21">22]</ref> with Self-Consistency (SC) <ref type="bibr" target="#b22">[23]</ref> to retrieve multiple rankings, which are then fused using Reciprocal Rank Fusion (RRF) <ref type="bibr" target="#b23">[24]</ref>. CoT-SC is a well-known technique for improving classification task performance by sampling multiple reasoning paths (n=10), i.e. repetitively performing the same task, with a high temperature (T=0.7), i.e. encouraging exploration, and taking the consensus of the answers from an LLM. During each resampling, the vocabulary descriptions are randomly shuffled to avoid the 'lost in the middle' problem, where the LLM tends to be biased toward the beginning or end of a document. We instruct the LLM to generate structured output in free text format and extract the answers using regular expressions.</p><p>RRF is a method in information retrieval for combining the results of multiple ranked lists. The core idea behind RRF is to leverage the rankings provided by different retrieval models and fuse them into a single, more robust ranking. The formula for RRF can be expressed as follows, where 𝑑 is a document in a set of documents 𝐷 and 𝑟 is the rank in a set of rankings 𝑅:</p><formula xml:id="formula_5">𝑅𝑅𝐹 𝑠𝑐𝑜𝑟𝑒 𝑑∈𝐷 = ∑ 𝑟∈𝑅 1 𝑘 + 𝑟(𝑑)</formula><p>, where k=0.01 (a tuned parameter)</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.3.2.">Table matching</head><p>The methodology is further extended to improve the results of the second dataset. As visualized in Figure <ref type="figure" target="#fig_3">4</ref>, the completion stage includes a second prompt containing the descriptions of the columns of the entire table identified earlier. The aim is to ensure that the column descriptions are consistent with one another, which potentially reveals a pattern. For instance, it is clear from the other descriptions that the table relates to a visitor visiting a webpage, this knowledge therefore reduces ambiguity when choosing between vocabulary labels "UserID" or "VisitorID" when mapping the column "ID".</p><p>Although the same CoT-SC process can be applied here, we limit it to one inference with a low temperature (T=0.0), focusing only on improving the top 1 out of the top 5 descriptions identified in the previous step. This limitation is due to the significantly longer prompt, and we expect the gains to be marginal.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.4.">Non-contextual Methodology</head><p>In addition to the two-stage pipeline, we also present a non-contextual method. This approach tests the LLM's ability to reason and recall properties from parametric memory rather than from context. As illustrated in Figure <ref type="figure">5</ref>, we simply ask GPT-4o to rank the top 5 fine-grained properties from the DBpedia ontology and filter out those not included in the given subset. It has been shown that top LLM models like ChatGPT and GPT-4 are pretrained on DBpedia data <ref type="bibr" target="#b24">[25]</ref>. Thus, we expect the non-contextual method to perform very well on Dataset 1. However, this approach is not usable on unseen vocabularies or KGs, such as those in Dataset 2.</p><p>We prompted GPT-4o using SC (n=20) without CoT and combined the results with RRF. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Selfconsistency</head><p>Given a table 'Museum' with columns ["RANG", "Museum", "Stadt", "Facebook-Fans"].</p><p>What are the top five official DBPedia descriptions for the column 'Stadt'?</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Prompt</head><p>Figure <ref type="figure">5</ref>: Detailed overview of the non-contextual methodology: An LLM is prompted using SC to retrieve multiple rankings, which are then fused using RRF.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Results</head><p>Both datasets were evaluated using the Hit@k metric for 2 different 𝑘: 1 and 5. The Hit@k metric indicates whether the true item appears in the first 𝑘 items of the sorted rank list. This can be expressed as:</p><formula xml:id="formula_6">Hit@k = max 𝑖=1..𝑘 { 1, 𝑟 𝑖 ∈ 𝑇 0, otherwise</formula><p>in which, 𝑟 𝑖 is the item ranked at position 𝑖, and 𝑇 is the set of items that are in the test set. We take the average of this metric across the entire dataset, which is also referred to as the "hit rate".</p><p>The results of both the non-contextual method and our approach are summarized in Table <ref type="table">1</ref>. Note that we did not evaluate the non-contextual method for Dataset 2, as it is not publicly available and therefore should not have been included in the LLM's training data. Additionally, we provide the scores from the embedding model (text-embedding-3-large), during the embedding phase of our method, as a baseline for comparison.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Table 1</head><p>Comparison of the baseline (text-embedding-3-large), contextual method, and our methodology for Dataset 1 &amp; 2.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Dataset 1 Dataset 2</head><p>Hit@1 Hit@5 Hit@1 Hit@5 Baseline 33% 52% 41% 72% Non-contextual Method 75% 92% --Our Method 62% 82% 84% 98%</p><p>In Table <ref type="table" target="#tab_3">2</ref>, we provide a breakdown of performance across the different phases of the pipeline for Datasets 1 &amp; 2. The "# Items" row indicates the number of vocabulary items remaining at the end of each phase. This also determines the metrics that we can evaluate. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.">Discussion and Future work</head><p>Our two-stage pipeline performed well, achieving top 1 and top 5 hit rates of 62% and 82% respectively for the first dataset, and top 1 and top 5 scores of 84% and 98% respectively for the second dataset. LLMs significantly outperform embedding models, achieving increases of 29% and 43% for the top 1 ranking in the first and second datasets, respectively. The non-contextual method achieved the highest score on Dataset 1, confirming that the DBpedia ontology is indeed part of the original training data of the LLM. This suggests that the model formed a better understanding of the properties during pretraining, possibly because they were seen in the context of other documents, and/or is better at reasoning using parametric memory. The results of our method show a high hit rate within the top 5 candidates, indicating that it retrieves almost all relevant labels. However, the score of the top 1 candidate is lower. Analysis suggests this is partly due to poor label definitions or subjective human preferences. For instance, in Dataset 1, a human labeler might categorize a city as "location", whereas our method might specify "locationCity". Moreover, based on the limited validation dataset, it is apparent that human labelers do not consistently choose the most fine-grained property. This has shifted our problem definition from identifying the "most fine-grained" property to "matching the human labels" to maximize our score. For this reason, we did not explicitly instruct the LLM to find the most fine-grained property.</p><p>We believe that the scores we report are generally understated. There is often a lack of consensus among humans regarding the ground truth. Our observations from the limited validation dataset suggest that, in some cases, the LLM's choice may be preferable to the human one. We aim to highlight these discrepancies, although we lack access to the definitive ground truth. If the true objective is to identify the "most fine-grained" property, further improvements could be achieved, such as refining the prompt, adding hierarchical relationships within the KG, etc.</p><p>This issue highlights three key advantages over human labelers:</p><p>• Humans must be aware of all possible items in the vocabulary or KG, which is a difficult task, whereas LLMs can ingest everything exhaustively. • Our pipeline provides consistent mapping. While human labelers may have varying preferences, our approach ensures uniform assignments from an independent expert. • Table metadata is often filled with specialized jargon, LLMs can eliminate the need for separate experts in each field.</p><p>We attribute this advantage to the strong semantic understanding of LLMs, specifically their ability to extract meanings from context-i.e., how words appear together in documents. Coupled with a comprehensive knowledge base of memorized facts, which includes concepts of names, places, and acronyms, giving them an edge over traditional embedding methods and, in some respects, even human capabilities.</p><p>Another advantage is the ability of LLMs to transfer to unseen tasks in a zero-shot manner. The two-stage pipeline removes the need for additional task-specific training data. In our tests, few-shot approaches offered only marginal improvements, making zero-shot learning a more practical and resource-efficient choice. This zero-shot capability also allows the pipeline to be applied to other datasets. As shown in the second dataset, a well-defined vocabulary suffices, making our solution applicable across other domains.</p><p>The use of LLMs is not without limitations. The accuracy of our approach is highly affected by the quality of the descriptions and how well the column names represent the corresponding table data. Some current mistakes stem from poorly defined descriptions or misleading column names. If possible, providing a single row of each table as additional context to the LLM could potentially go a long way to improve clarity and also benefit human labelers. Cost remains another important factor. Table <ref type="table" target="#tab_4">3</ref> provides a breakdown of the number of input and output tokens and their associated costs for each phase within our two-stage pipeline for Dataset 2, which has the highest token count of the two datasets. The cost of an LLM is determined by the volume of tokens provided (input) and generated (output), each at different price points. Running our solution for Dataset 2 is estimated to be around $51 in total, or 4.33¢ per mapping of a column, with costs scaling linearly. This indicates that the function of a retrieval stage in this approach, as opposed to providing all items in the vocabulary or KG to the completion stage directly, is a very cost-efficient measure.</p><p>There are, however, opportunities to further reduce the cost of the provided pipeline and research the balance between cost and accuracy. For instance, during the rerank phase, one can further adjust the step and window size or use less expensive models for lower-ranked embeddings. During the completion stage, the self-consistency parameter can also be easily tweaked without making structural changes to the architecture. While these are stopgaps to the current limitations, the cost and performance of foundation models are set to continue to improve in the near future.</p><p>Looking forward, future research could further improve our pipeline at various stages. Embedding-based rankings could benefit from a hybrid approach that combines sparse retrieval models with dense models to help identify exact matches, such as names, IDs, and numbers. One could also suggest more fine-grained domain and ranges based on the usage of the property in axioms and data, or even towards the specific task or context at hand. Furthermore, we currently see an opportunity to make LLM-based rerankers more scalable through new algorithms or the aforementioned strategic use of cheaper LLMs. Additionally, their performance can potentially be improved by combining SC with RRF. As we look ahead, new solutions will need to be devised to maintain scalability if the vocabulary changes at test time.</p><p>The holy grail is still to embed knowledge into the LLM itself and retrieve the items from parametric memory. The advantages of this approach are revealed by the non-contextual method in this paper, where a single and simple prompt, together with a post-processing operation, results in very accurate table annotations. This solution is, however, still only applicable for DBpedia, as the GPT-4o model has the DBpedia properties stored in its parametric memory. Nonetheless, injecting new knowledge into LLMs through fine-tuning remains a challenging task and recent advancements such as Knowledge-based Model Editing are still in their emerging stage <ref type="bibr" target="#b25">[26]</ref>.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 2 :</head><label>2</label><figDesc>Figure 2: Overview of the two-stage methodology composed of a retrieval and completion stage after an initial preprocessing step. The hexagonal knot and lama icon refer to the OpenAI and Llama models, respectively.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figure 3</head><label>3</label><figDesc>Figure 3 illustrates the Retrieval Stage. It consists of an advanced RAG solution including an embedding step and a reranking step, both of which will be further explained in the subsequent subsections.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>[ 2 ]Figure 3 :</head><label>23</label><figDesc>Figure 3: Detailed overview of the retrieval stage: Vocabulary items are enriched with an LLM and embedded with an embedding model, then stored. Table metadata is embedded similarly to create query embeddings. A similarity search retrieves the top-k matches for the query embedding, which are reranked by RankGPT based on relevance.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_3"><head>Figure 4</head><label>4</label><figDesc>Figure 4 illustrates the Completion Stage, which aims to make a final ranking of the retrieved candidates and will be further explained in detail in the following subsections.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_4"><head>Figure 4 :</head><label>4</label><figDesc>Figure 4: Detailed overview of the completion stage: An LLM is prompted using CoT-SC to retrieve multiple rankings, which are then fused using RRF. The process includes a second prompt to ensure column descriptions are consistent across the entire table.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_5"><head>7</head><label></label><figDesc>https://huggingface.co/spaces/lmsys/chatbot-arena-leaderboard confidential location, locationCity, city,</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head></head><label></label><figDesc>-Fans"] Matching table metadata only, i.e., table and column names, to a vocabulary or ontology without any access to table data. In this example, the column "Stadt" of the "Museum" table is matched to the "locationCity" property within the given vocabulary.</figDesc><table><row><cell>table_name:</cell><cell></cell><cell>Museum</cell><cell></cell><cell></cell><cell>…</cell><cell>Vocabulary</cell></row><row><cell>table_columns: table_cells:</cell><cell>RANG</cell><cell>Name Unavailable</cell><cell>Stadt</cell><cell>Fans</cell><cell cols="2">locationName location locationCity city</cell></row><row><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell cols="2">livingPlace</cell></row><row><cell>Figure 1:</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>Table Query Enrichment</head><label>Query</label><figDesc></figDesc><table><row><cell>Embedding</cell><cell>Query Embedding</cell><cell>Top-k matches</cell><cell>Rerank</cell></row><row><cell>Model</cell><cell></cell><cell></cell><cell></cell></row><row><cell></cell><cell cols="2">Similarity</cell><cell></cell></row><row><cell></cell><cell cols="2">Search</cell><cell></cell></row><row><cell></cell><cell cols="2">{item 1}</cell><cell></cell></row><row><cell></cell><cell cols="2">{item 2}</cell><cell></cell></row><row><cell>Item Enrichment</cell><cell cols="2">{item 3} {item 4} {item 5}</cell><cell></cell></row><row><cell>Vocabulary</cell><cell cols="2">{item 6} {item 7}</cell><cell></cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_3"><head>Table 2</head><label>2</label><figDesc>Performance metrics across different stages for Dataset 1 &amp; 2.</figDesc><table><row><cell></cell><cell>Prepro-</cell><cell>Retrieval</cell><cell>Retrieval</cell><cell>Completion</cell><cell>Completion</cell></row><row><cell></cell><cell>cessing</cell><cell>Stage (em-</cell><cell>Stage</cell><cell>Stage (top-k</cell><cell>Stage (table</cell></row><row><cell></cell><cell></cell><cell>bedding)</cell><cell>(rerank)</cell><cell>matching)</cell><cell>matching)</cell></row><row><cell>Dataset 1 # Items</cell><cell>2881</cell><cell>150</cell><cell>30</cell><cell>5</cell><cell>-</cell></row><row><cell>Hit@150</cell><cell>-</cell><cell>95%</cell><cell>-</cell><cell>-</cell><cell>-</cell></row><row><cell>Hit@30</cell><cell>-</cell><cell>75%</cell><cell>91%</cell><cell>-</cell><cell>-</cell></row><row><cell>Hit@5</cell><cell>-</cell><cell>52%</cell><cell>59%</cell><cell>82%</cell><cell>-</cell></row><row><cell>Hit@1</cell><cell>-</cell><cell>33%</cell><cell>43%</cell><cell>62%</cell><cell>-</cell></row><row><cell>Dataset 2 # Items</cell><cell>1181</cell><cell>150</cell><cell>30</cell><cell>5</cell><cell>1</cell></row><row><cell>Hit@150</cell><cell>-</cell><cell>100%</cell><cell>-</cell><cell>-</cell><cell>-</cell></row><row><cell>Hit@30</cell><cell>-</cell><cell>96%</cell><cell>99%</cell><cell>-</cell><cell>-</cell></row><row><cell>Hit@5</cell><cell>-</cell><cell>72%</cell><cell>95%</cell><cell>98%</cell><cell>-</cell></row><row><cell>Hit@1</cell><cell>-</cell><cell>41%</cell><cell>77%</cell><cell>83%</cell><cell>84%</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_4"><head>Table 3</head><label>3</label><figDesc>Cost and token usage across different stages for Dataset 2</figDesc><table><row><cell></cell><cell cols="2"># Tokens (in/out) Cost per 1M</cell><cell>Total cost</cell><cell>Model</cell><cell>Inference</cell></row><row><cell></cell><cell></cell><cell>tokens</cell><cell></cell><cell></cell><cell>Platform</cell></row><row><cell></cell><cell></cell><cell>(in/out)</cell><cell></cell><cell></cell><cell></cell></row><row><cell>Preprocessing</cell><cell>16K/18K</cell><cell>$5.0/$15.0</cell><cell>$0.35</cell><cell>GPT-4o</cell><cell>OpenAI</cell></row><row><cell>Retrieval Stage</cell><cell>125K</cell><cell>$0.02</cell><cell>$0.0025</cell><cell>text-embedding-3-</cell><cell>OpenAI</cell></row><row><cell>(embedding)</cell><cell></cell><cell></cell><cell></cell><cell>large</cell><cell></cell></row><row><cell>Retrieval Stage</cell><cell>22.7M/1.4M</cell><cell>$0.59/$0.79</cell><cell>$14.50</cell><cell>Llama-3-70B-</cell><cell>DeepInfra</cell></row><row><cell>(rerank)</cell><cell></cell><cell></cell><cell></cell><cell>Instruct</cell><cell></cell></row><row><cell>Completion Stage</cell><cell>3.6M/939K</cell><cell>$5.0/$15.0</cell><cell>$32.09</cell><cell>GPT-4o</cell><cell>OpenAI</cell></row><row><cell>(top-k matching)</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell>Completion Stage</cell><cell>555K/96K</cell><cell>$5.0/$15.0</cell><cell>$4.22</cell><cell>GPT-4o</cell><cell>OpenAI</cell></row><row><cell>(table matching)</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell>ALL</cell><cell>27.0M/2.5M</cell><cell>-</cell><cell>$51.16</cell><cell>-</cell><cell>-</cell></row></table></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_0">https://www.cs.ox.ac.uk/isg/challenges/sem-tab/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_1">https://sem-tab-challenge.github.io/2024/tracks/metadata-to-kg-track.html</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_2">https://platform.openai.com/docs/models</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="4" xml:id="foot_3">http://dbpedia.org/ontology/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="5" xml:id="foot_4">https://platform.openai.com/docs/models/embeddings</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="6" xml:id="foot_5">e.g. https://mappings.dbpedia.org/index.php/OntologyProperty:ReleaseDate</note>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgments</head><p>This work was partly funded by the SolidLab Vlaanderen project (Flemish Government, EWI and RRF project VV023/10), the FWO Project FRACTION (Nr. G086822N) and the BOF Project FRACTION (Nr. BOF.24Y.2021.0048.01).</p></div>
			</div>


			<div type="availability">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Code availability</head><p>The code and additional resources used in this paper are available on GitHub at the following repository: https://github.com/predict-idlab/Meta2Concept</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Uncertainty in big data analytics: survey, opportunities, and challenges</title>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">H</forename><surname>Hariri</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><forename type="middle">M</forename><surname>Fredericks</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><forename type="middle">M</forename><surname>Bowers</surname></persName>
		</author>
		<idno type="DOI">10.1186/s40537-019-0206-3</idno>
	</analytic>
	<monogr>
		<title level="j">Journal of Big Data</title>
		<imprint>
			<biblScope unit="volume">6</biblScope>
			<biblScope unit="page">44</biblScope>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">From tabular data to knowledge graphs: A survey of semantic table interpretation tasks and methods</title>
		<author>
			<persName><forename type="first">J</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Chabot</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Troncy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V.-P</forename><surname>Huynh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Labbé</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Monnin</surname></persName>
		</author>
		<idno type="DOI">10.1016/j.websem.2022.100761</idno>
	</analytic>
	<monogr>
		<title level="j">Journal of Web Semantics</title>
		<imprint>
			<biblScope unit="volume">76</biblScope>
			<biblScope unit="page">100761</biblScope>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<monogr>
		<author>
			<persName><forename type="first">E</forename><surname>Lobo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Hassanzadeh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Pham</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Mihindukulasooriya</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Subramanian</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Samulowitz</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2309.11506</idno>
		<title level="m">Matching table metadata with business glossaries using large language models</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b3">
	<monogr>
		<author>
			<persName><forename type="first">Y</forename><surname>Gao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Xiong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Gao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Jia</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Pan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Bi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Dai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Sun</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Wang</surname></persName>
		</author>
		<idno type="DOI">10.48550/arXiv.2312.10997</idno>
		<idno type="arXiv">arXiv:2312.10997</idno>
		<title level="m">Retrieval-Augmented Generation for Large Language Models: A Survey</title>
				<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">From tables to knowledge: Recent advances in table understanding</title>
		<author>
			<persName><forename type="first">J</forename><surname>Pujara</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Szekely</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Sun</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Chen</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery &amp; Data Mining</title>
				<meeting>the 27th ACM SIGKDD Conference on Knowledge Discovery &amp; Data Mining</meeting>
		<imprint>
			<date type="published" when="2021">2021</date>
			<biblScope unit="page" from="4060" to="4061" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Resources to benchmark tabular data to knowledge graph matching systems</title>
		<author>
			<persName><forename type="first">E</forename><surname>Jiménez-Ruiz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Hassanzadeh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Efthymiou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Srinivas</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">The Semantic Web: 17th International Conference, ESWC 2020</title>
				<meeting><address><addrLine>Heraklion, Crete, Greece</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2019-05-31">2019. May 31-June 4, 2020. 2020</date>
			<biblScope unit="page" from="514" to="530" />
		</imprint>
	</monogr>
	<note>Proceedings 17</note>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Semtab 2021: Tabular data annotation with mtab tool</title>
		<author>
			<persName><forename type="first">P</forename><surname>Nguyen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Yamada</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Kertkeidkachorn</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Ichise</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Takeda</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">SemTab@ ISWC</title>
				<imprint>
			<date type="published" when="2021">2021</date>
			<biblScope unit="page" from="92" to="101" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Csv2kg: Transforming tabular data into semantic knowledge</title>
		<author>
			<persName><forename type="first">B</forename><surname>Steenwinckel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Vandewiele</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>De Turck</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Ongenae</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">SemTab, ISWC Challenge</title>
				<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Dagobah: table and graph contexts for efficient semantic annotation of tabular data</title>
		<author>
			<persName><forename type="first">V.-P</forename><surname>Huynh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Chabot</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Deuzé</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Labbé</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Monnin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Troncy</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">The 20th International Semantic Web Conference (ISWC 2021)</title>
				<imprint>
			<date type="published" when="2021">2021</date>
			<biblScope unit="volume">3103</biblScope>
			<biblScope unit="page">2</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">String similarity metrics for ontology alignment</title>
		<author>
			<persName><forename type="first">M</forename><surname>Cheatham</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Hitzler</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">The Semantic Web-ISWC 2013: 12th International Semantic Web Conference</title>
				<meeting><address><addrLine>Sydney, NSW, Australia</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2013">October 21-25, 2013. 2013</date>
			<biblScope unit="page" from="294" to="309" />
		</imprint>
	</monogr>
	<note>Proceedings, Part II 12</note>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">A survey of exploiting wordnet in ontology matching</title>
		<author>
			<persName><forename type="first">F</forename><surname>Lin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Sandkuhl</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">IFIP International Conference on Artificial Intelligence in Theory and Practice</title>
				<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2008">2008</date>
			<biblScope unit="page" from="341" to="350" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Matching ontologies with word2vec model based on cosine similarity</title>
		<author>
			<persName><forename type="first">J</forename><surname>Liao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Huang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Li</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">The International Conference on Artificial Intelligence and Computer Vision</title>
				<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2021">2021</date>
			<biblScope unit="page" from="367" to="374" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<monogr>
		<author>
			<persName><forename type="first">Y</forename><surname>He</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Dong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Horrocks</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2309.07172</idno>
		<title level="m">Exploring large language models for ontology alignment</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b13">
	<monogr>
		<author>
			<persName><forename type="first">R</forename><surname>Peeters</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Bizer</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2310.11244</idno>
		<title level="m">Entity matching using large language models</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b14">
	<monogr>
		<author>
			<persName><forename type="first">S</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Wu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Zhang</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2310.11318</idno>
		<title level="m">Utilising a large language model to annotate subject metadata: A case study in an australian national research data catalogue</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b15">
	<monogr>
		<author>
			<persName><forename type="first">T</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Yue</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Sun</surname></persName>
		</author>
		<idno>arXiv:</idno>
		<ptr target="2311.09206" />
		<title level="m">TableLlama: Towards Open Large Generalist Models for Tables</title>
				<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<monogr>
		<author>
			<persName><forename type="first">M</forename><surname>Martorana</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Kuhn</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Stork</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Van Ossenbruggen</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2403.00884</idno>
		<title level="m">Text classification of column headers with a controlled vocabulary: leveraging llms for metadata enrichment</title>
				<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b17">
	<monogr>
		<author>
			<persName><forename type="first">H</forename><forename type="middle">B</forename><surname>Giglou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Souza</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Engel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Auer</surname></persName>
		</author>
		<idno type="DOI">10.48550/arXiv.2404.10317</idno>
		<idno type="arXiv">arXiv:2404.10317</idno>
		<title level="m">LLMs4OM: Matching Ontologies with Large Language Models</title>
				<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
	<note>cs</note>
</biblStruct>

<biblStruct xml:id="b18">
	<analytic>
		<title level="a" type="main">OLaLa: Ontology Matching with Large Language Models</title>
		<author>
			<persName><forename type="first">S</forename><surname>Hertling</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Paulheim</surname></persName>
		</author>
		<idno type="DOI">10.1145/3587259.3627571</idno>
		<idno type="arXiv">arXiv:2311.03837</idno>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 12th Knowledge Capture Conference 2023</title>
				<meeting>the 12th Knowledge Capture Conference 2023</meeting>
		<imprint>
			<date type="published" when="2023">2023</date>
			<biblScope unit="page" from="131" to="139" />
		</imprint>
	</monogr>
	<note>cs</note>
</biblStruct>

<biblStruct xml:id="b19">
	<monogr>
		<title level="m" type="main">Is chatgpt good at search? investigating large language models as re-ranking agent</title>
		<author>
			<persName><forename type="first">W</forename><surname>Sun</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Yan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Ma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Ren</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Yin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Ren</surname></persName>
		</author>
		<idno>ArXiv abs/2304.09542</idno>
		<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<analytic>
		<title level="a" type="main">Chain-of-thought prompting elicits reasoning in large language models</title>
		<author>
			<persName><forename type="first">J</forename><surname>Wei</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Schuurmans</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Bosma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Ichter</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Xia</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><forename type="middle">H</forename><surname>Chi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><forename type="middle">V</forename><surname>Le</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Zhou</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 36th International Conference on Neural Information Processing Systems, NIPS &apos;22</title>
				<meeting>the 36th International Conference on Neural Information Processing Systems, NIPS &apos;22<address><addrLine>Red Hook, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>Curran Associates Inc</publisher>
			<date type="published" when="2024">2024</date>
			<biblScope unit="page" from="24824" to="24837" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b21">
	<analytic>
		<title level="a" type="main">Large Language Models are Zero-Shot Reasoners</title>
		<author>
			<persName><forename type="first">T</forename><surname>Kojima</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">S</forename><surname>Gu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Reid</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Matsuo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Iwasawa</surname></persName>
		</author>
		<ptr target="https://proceedings.neurips.cc/paper_files/paper/2022/file/8bb0d291acd4acf06ef112099c16f326-Paper-Conference.pdf" />
	</analytic>
	<monogr>
		<title level="m">Advances in Neural Information Processing Systems</title>
				<editor>
			<persName><forename type="first">S</forename><surname>Koyejo</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">S</forename><surname>Mohamed</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Agarwal</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">D</forename><surname>Belgrave</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">K</forename><surname>Cho</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Oh</surname></persName>
		</editor>
		<imprint>
			<publisher>Curran Associates, Inc</publisher>
			<date type="published" when="2022">2022</date>
			<biblScope unit="volume">35</biblScope>
			<biblScope unit="page" from="22199" to="22213" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b22">
	<monogr>
		<author>
			<persName><forename type="first">X</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Wei</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Schuurmans</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><surname>Le</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Chi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Narang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Chowdhery</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Zhou</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2203.11171</idno>
		<title level="m">Selfconsistency improves chain of thought reasoning in language models</title>
				<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b23">
	<analytic>
		<title level="a" type="main">Reciprocal rank fusion outperforms condorcet and individual rank learning methods</title>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">V</forename><surname>Cormack</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">L A</forename><surname>Clarke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Buettcher</surname></persName>
		</author>
		<idno type="DOI">10.1145/1571941.1572114</idno>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 32nd international ACM SIGIR conference on Research and development in information retrieval, SIGIR &apos;09</title>
				<meeting>the 32nd international ACM SIGIR conference on Research and development in information retrieval, SIGIR &apos;09<address><addrLine>New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>Association for Computing Machinery</publisher>
			<date type="published" when="2009">2009</date>
			<biblScope unit="page" from="758" to="759" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b24">
	<monogr>
		<title level="m" type="main">On Exploring the Reasoning Capability of Large Language Models with Knowledge Graphs</title>
		<author>
			<persName><forename type="first">P.-C</forename><surname>Lo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y.-H</forename><surname>Tsai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E.-P</forename><surname>Lim</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S.-Y</forename><surname>Hwang</surname></persName>
		</author>
		<idno>arXiv:</idno>
		<ptr target="2312.00353" />
		<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b25">
	<monogr>
		<title level="m" type="main">Knowledge Editing for Large Language Models: A Survey</title>
		<author>
			<persName><forename type="first">S</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Zhu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Zheng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Li</surname></persName>
		</author>
		<idno type="DOI">10.48550/arXiv.2310.16218</idno>
		<idno type="arXiv">arXiv:2310.16218</idno>
		<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
	<note>cs</note>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
