<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Towards LLM-augmented Creation of Semantic Models for Dataspaces</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Sayed</forename><surname>Hoseini</surname></persName>
							<email>sayed.hoseini@hs-niederrhen.de</email>
							<affiliation key="aff0">
								<orgName type="institution">Hochschule Niederrhein</orgName>
								<address>
									<settlement>Krefeld</settlement>
									<country key="DE">Germany</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Andreas</forename><surname>Burgdorf</surname></persName>
							<email>burgdorf@uni-wuppertal.de</email>
							<affiliation key="aff1">
								<orgName type="department">Institute for Technologies and Management of Digital Transformation</orgName>
								<orgName type="institution">University of Wuppertal</orgName>
								<address>
									<settlement>Wuppertal</settlement>
									<country key="DE">Germany</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Alexander</forename><surname>Paulus</surname></persName>
							<email>paulus@uni-wuppertal.de</email>
							<affiliation key="aff1">
								<orgName type="department">Institute for Technologies and Management of Digital Transformation</orgName>
								<orgName type="institution">University of Wuppertal</orgName>
								<address>
									<settlement>Wuppertal</settlement>
									<country key="DE">Germany</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Tobias</forename><surname>Meisen</surname></persName>
							<email>meisen@uni-wuppertal.de</email>
							<affiliation key="aff1">
								<orgName type="department">Institute for Technologies and Management of Digital Transformation</orgName>
								<orgName type="institution">University of Wuppertal</orgName>
								<address>
									<settlement>Wuppertal</settlement>
									<country key="DE">Germany</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Christoph</forename><surname>Quix</surname></persName>
							<email>christoph.quix@hs-niederrhen.de</email>
							<affiliation key="aff0">
								<orgName type="institution">Hochschule Niederrhein</orgName>
								<address>
									<settlement>Krefeld</settlement>
									<country key="DE">Germany</country>
								</address>
							</affiliation>
							<affiliation key="aff2">
								<orgName type="institution" key="instit1">Fraunhofer FIT</orgName>
								<orgName type="institution" key="instit2">St. Augustin</orgName>
								<address>
									<country key="DE">Germany</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">André</forename><surname>Pomp</surname></persName>
							<email>pomp@uni-wuppertal.de</email>
							<affiliation key="aff1">
								<orgName type="department">Institute for Technologies and Management of Digital Transformation</orgName>
								<orgName type="institution">University of Wuppertal</orgName>
								<address>
									<settlement>Wuppertal</settlement>
									<country key="DE">Germany</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Towards LLM-augmented Creation of Semantic Models for Dataspaces</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">CFAEE9F86E37A5C401515F456DEE40C6</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T16:36+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Dataspace</term>
					<term>Semantic Modeling</term>
					<term>LLMs</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Dataspaces aim to enable smooth and reliable data exchange between different organizations. They have gained increasing attention in Europe following the enactment of the European Data Governance Act. This legislation emphasizes trust, accessibility, and shared dataspaces, which require semantic interoperability grounded in the FAIR principles. Although semantic descriptions in the form of semantic models and ontologies are integral to dataspaces, their full potential remains underutilized. Meaningful metadata, including contextual information, enhances data usability, but manually creating semantic models can be challenging. Large Language Models (LLMs) offer a new way to utilize data in dataspaces. Their advanced natural language processing capabilities enable context-aware data processing and semantic understanding. This paper presents initial experiments on customizing and optimizing LLMs for semantic labeling and modeling tasks. The contributions of this work include research questions for future investigations, early experiments demonstrating the applicability of LLM for semantic labeling, and proposed directions to address discovered challenges.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>The European Data Governance Act <ref type="bibr" target="#b0">[1,</ref><ref type="bibr" target="#b1">2]</ref> outlines definitions and objectives aimed at bolstering trust, broadening data accessibility, and promoting shared dataspaces. Its impact extends across various data consumers and providers from academia as well as businesses. Efficient data sharing within dataspaces necessitates semantic interoperability as an essential design principle, grounded in the required adherence to FAIR principles <ref type="bibr" target="#b2">[3]</ref>. Thereby, the utilization of semantic descriptions and ontologies are already part of many dataspaces, but the potential is far from being fully utilized in actual implementations <ref type="bibr" target="#b3">[4]</ref>. An example of this is semantic interoperability and integration, a key aspect of dataspaces that requires aggregating and integrating large amounts of heterogeneous data from different sources. Managing data can be challenging, not only due to the variety of data formats such as XML, CSV, JSON, relational data, and graph data. In addition, data is often distributed across different departments within an organization, under different governance regimes, and data models. It is important to have a clear and logical structure of information, which fosters a common understanding in dataspaces, i.e., a lingua franca for data moderation <ref type="bibr" target="#b4">[5]</ref> based on the Linked Data principles.</p><p>Meaningful metadata is crucial for enhancing data usability, particularly for users with limited domain knowledge or those unfamiliar with a dataset. Annotating raw data from heterogeneous data sources with semantically rich models enhances data interpretability and usability <ref type="bibr" target="#b5">[6,</ref><ref type="bibr" target="#b6">7]</ref>. This type of semantic data expands beyond typical extractable metadata, such as schema, data types, sizes, and formats, to include contextual information that is not inherent to the specific data source. The field of Semantic Data Management (SDM) <ref type="bibr" target="#b7">[8]</ref> aims to represent the metadata about heterogeneous data sources in the form of ontologies or knowledge graphs (KG) serialized in a language of the Semantic Web. Hence, the goal is to establish an additional layer between the data and the knowledge layer <ref type="bibr" target="#b8">[9]</ref>. This is highly relevant for dataspaces because they integrate data from various systems and platforms, which requires data to be interoperable and seamlessly exchangeable between systems. In order to implement SDM in practice, conceptualizations in the form of KGs and/or ontologies <ref type="bibr" target="#b9">[10]</ref>, and a mapping between concepts and data items are required. Semantic models provide these mappings from single datasets to a common data model to represent data consistently across different applications in a way that is understandable and interpretable by both humans and machines <ref type="bibr" target="#b10">[11]</ref>.</p><p>With companies increasingly acknowledging the importance of data for their business operations, semantic descriptions are often integrated into data management and governance strategies <ref type="bibr" target="#b11">[12,</ref><ref type="bibr" target="#b12">13]</ref>, where an ontology or KG serves as a conceptual representation of an organization's data assets. A data source that is semantically well-annotated can be identified and interpreted by leveraging conceptual representations of the data and by comprehending the provided context information stored in the model. However, a huge initial overhead, coming from the time-consuming manual process of creating meaningful semantic descriptions for data sources, hampers the widespread adoption of SDM in practise <ref type="bibr" target="#b7">[8]</ref>. Creating semantic models entails deciphering the existing data source, consulting appropriate conceptualizations, and establishing connections between data attributes and concepts provided by the conceptualization.</p><p>Automating this task can be challenging and complex. Futia et al. <ref type="bibr" target="#b13">[14]</ref> present a method based on graph neural networks covering the process of the process of semantic modeling. However, the model can only optimize semantic models for which historic training data exists. Xu et al. <ref type="bibr" target="#b14">[15]</ref> train a cross-modal network to learn semantic features between data sources and semantic models. They admit that the method has shortcomings in dynamically augmenting the semantic models to cover concepts that are not part of the original training data. Moreover, challenges remain with detecting and correcting potentially incorrect attribute types if a source attribute has more than one attribute type, and distinguishing similar attributes with the same entities and semantic types.</p><p>Following the rise of Large Language Models (LLMs), one can expect to see a major impact on the landscape of data utilization and exchange within dataspaces. LLMs, such as OpenAI's GPT-3.5 and GPT-4.0, have demonstrated remarkable capabilities in understanding, generating, and processing vast amounts of textual data <ref type="bibr" target="#b15">[16,</ref><ref type="bibr" target="#b16">17]</ref>. Their abilities in natural language processing  enable advanced semantic understanding and context-aware data processing within dataspaces. A promising field of LLM application is the integration of heterogeneous data sources stored in a dataspace.</p><formula xml:id="formula_0">R-50 R-100 R-150 R-200 R-250 R-300 R-350 R-400 R-450 R-500 R-550 R-600 R-650 R-700 -60 -120 -</formula><p>In this article, we highlight some initial experiments in this direction to examine the question of how such general-purpose AI systems can be customized and optimized for data integration tasks in the sense of SDM. In particular, we make the following contributions:</p><p>• A set of research questions to be answered in future research endeavors • Early experiments to illustrate the applicability of LLMs to the tasks of semantic labeling • Potential future research directions to address the identified challenges with the applicability of LLMs to the tasks of semantic labeling and modeling.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Semantic Data Management</head><p>Figure <ref type="figure" target="#fig_0">1</ref> illustrates the basic idea of semantic models. The raw datasets in the dataspace are represented at the bottom; they can be in different formats and structures, such as tabular data or hierarchical JSON data, but have partially overlapping content. The semantic model is a projection from a shared conceptualization onto the different datasets. It utilizes relevant entities and relationships of the conceptualization, in this case, the schema.org ontology (prefix schema:), to formalize the context information of the dataset. An essential part of each semantic model are the mappings, indicated as dotted lines, which link attributes in the datasets to classes in the semantic model using properties of these classes. These elementary mappings are referred The semantic model creation process as formalized by <ref type="bibr" target="#b20">[21]</ref>. Specific terms used for available automation are stated in purple.</p><formula xml:id="formula_1">R-50 R-100 R-150 R-200 R-250 R-300 R-350 R-400 R-450 R-500 R-550 R-600 R-650 R-</formula><p>to as semantic labels. The semantic model captures the precise meaning of the dataset, explicitly encoding the semantic types and relationships among its attributes within the graph.</p><p>Following the definition of semantic labeling by Pham et al. <ref type="bibr" target="#b17">[18]</ref>, a semantic label is an annotation of a dataset attribute by a tuple consisting of a class (subject) and a property (predicate). In this work, a semantic label is represented as a triple: (subject, predicate, schema attribute). For example, the semantic label of the table's column 'Title' is constructed through the subject 'schema:Movie' and the predicate 'schema:title' modeling the relationship between them. This connects the table's content to the attribute 'titel' in the JSON object, indicating an entry point for data integration between the two (heterogeneous) datasets. Moreover, the semantic model doesn't merely rely on a static conceptualization; it can also introduce novel classes and properties <ref type="bibr" target="#b18">[19]</ref>. The necessity for evolution becomes apparent when users contribute datasets containing concepts and relationships not yet included in the conceptualization. The semantic label for the JSON key 'verkaeufe' is represented as the triple: (schema:ScreeningEvent, :ticketsSold, 'verkaeufe'). Here, the predicate is a novel property for that specific domain, which is not (yet) present in this form in the general-purpose schema.org ontology. This new knowledge can be systematically integrated, thus perpetually advancing the conceptualization layer <ref type="bibr" target="#b19">[20]</ref>. Semantic models complement extractable metadata (such as data types, sizes, formats, etc.) to convey context information that may not be inherent to the dataset at hand, for instance, a starting date of a 'schema:ScreeningEvent' as shown in Figure <ref type="figure" target="#fig_0">1</ref>.</p><p>The underlying process of semantic model creation has been formalized by <ref type="bibr" target="#b20">[21]</ref> and is visualized in Figure <ref type="figure" target="#fig_1">2</ref>, starting with the identification of the schema of the dataset, followed by a semantic labeling phase, in which basic concepts are assigned to the identified attributes. Automated semantic labeling, referred to as Semantic Type Detection <ref type="bibr" target="#b21">[22]</ref>, is the process of identifying these labels using algorithms and machine learning models. Subsequent to the semantic labeling, the semantic modeling phase builds the remaining semantic model by formalizing the context information. During semantic modeling, semantic relation inference <ref type="bibr" target="#b13">[14]</ref> refers to the process of automated identification of relationships and additional concepts, resulting in a generated semantic model. All automation is followed by the semantic refinement phase, where the modeler is involved in the modeling process to correct any errors present before the semantic model is finalized and stored for documentation purposes. In practice, semantic relation inference depends heavily on accurate semantic labels <ref type="bibr" target="#b22">[23,</ref><ref type="bibr" target="#b23">24]</ref>, which underscores the importance of semantic type detection in fully automated systems to induce as few errors as possible for the modeler to correct.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1.">Research Questions</head><p>With an introduction to the field of SDM at hand, we move on to formulate the research questions that have motivated this work. The goal is to leverage LLMs to improve the automation of semantic model creation for large quantities of heterogeneous data sources that share a common domain in a dataspace.</p><p>• RQ1: How to utilize LLMs to perform semantic type detection with a fixed set of labels coming from a pre-selected conceptualization (such as WikiData, or schema.org)? • RQ2: How to utilize LLMs to perform semantic type detection against an arbitrary domain ontology, i.e., with no labeled dataset or zero-shot classification? • RQ3: How can LLMs be utilized to identify and formalize the context of a given dataset, creating a full semantic model?</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Related Work</head><p>First, in the scope of this contribution, we consider only works after the launch of ChatGPT in November 2022. While one can make a strong case that (large) language models have existed before this date, we decided to draw a line due to the massive performance increase which was quite suddenly accessible to the public. Most of the works found in this limited range are pre-prints that have not yet been peer-reviewed and published in scientific journals. To the best of our knowledge, so far, except for the below-mentioned approaches, there seems to be no further LLM-based efforts on the integration of the semantics of several heterogeneous data sources modeled directly in a language of the Semantic Web in order to generate semantic models in the sense of Figure <ref type="figure" target="#fig_0">1</ref>. Korini et al. <ref type="bibr" target="#b24">[25]</ref> are among the first to report the application of LLMs for the Column Type Annotation (CTA) task. CTA is a schema-level annotation task that represents a simplified form of our interpretation of semantic labeling (no predicate) as it aims to map the underlying table schema to a conceptualization. They view CTA as a multi-class classification problem and evaluate different prompt designs. One important artifact that is highlighted is that ChatGPT tends to ignore the instruction to use terms from the label space, and instead answers using different terms. This is a known drawback of contemporary LLMs known as the Hallucination problem <ref type="bibr" target="#b25">[26]</ref>. Their proposed solutions for this challenge involve first determining a set of classes of the entities described in the table and depending on this set asking ChatGPT to annotate columns using only the relevant subset of the overall vocabulary. The evaluation of the approach reports competitive performance when evaluated against the more traditional models which are mostly directly fine-tuned for the CTA task and require significant amounts of task-specific training data <ref type="bibr" target="#b26">[27]</ref>.</p><p>A dataspace stores heterogeneous data of any type of which the majority may be in some form relational, but an important fraction may be more complex, e.g. nested and even unstructured (video, text, audio, ...). We found several works <ref type="bibr" target="#b27">[28,</ref><ref type="bibr" target="#b28">29,</ref><ref type="bibr" target="#b29">30</ref>] that aim at customizing LLMs via fine-tuning for tables in particular. The goal is to solve the challenge of table understanding which is closely related to understanding the semantics of a data source as it includes the CTA task for example. Usmani et al. <ref type="bibr" target="#b12">[13]</ref> highlight the importance of multi-modal knowledge graphs for dataspaces, present a review of the current state, and propose an ontology towards further development. Furthermore, since important use cases for datasets can be attributed to numerical data, it is important to have solid numerical reasoning skills <ref type="bibr" target="#b30">[31]</ref>. Several works suggest that modern LLMs excel in simple problem settings, but they fall short of human expert performance in problems requiring numerical reasoning over long contexts. As the complexity of challenging mathematical problems increases, LLMs currently exhibit suboptimal performance <ref type="bibr" target="#b16">[17,</ref><ref type="bibr" target="#b31">32]</ref>.</p><p>There exist several works that investigate the use of LLMs for Knowledge Graph Engineering <ref type="bibr" target="#b32">[33,</ref><ref type="bibr" target="#b33">34,</ref><ref type="bibr" target="#b34">35]</ref>. Here the goal is to utilize the LLM for common tasks related to KGs. For example, Meyer et al. <ref type="bibr" target="#b35">[36]</ref> investigate SPARQL query generation as well as knowledge extraction from fact sheets and KG exploration among others. They present several prompts that essentially test how to pose questions in natural language executed against serialized KGs. The experiments show that LLMs can return syntactically correct SPARQL queries and even entire serialized RDF models with the desired form based on a task formulated in natural language. Further, LLMs can find relationships in KG and answer basic questions correctly, e.g., "Are there any connections between US and UK?". They report major performance differences between GPT-3.5 and GPT-4 in favor of the latter. One particular prompt aims at extracting knowledge from tables and converting it into a serialized KG, which is very close to the idea of semantic modeling. The experiment illustrates several problems with the output of contemporary LLMs:</p><p>• A tendency to prioritize the usage of schema.org vocabulary. While this works well for well-known entities and properties, the LLMs invent reasonable, but non-existent classes and properties (in the schema.org namespace) for concepts and relations that are too specific. • Non-deterministic output: For multiple runs of the same prompt to the LLM, the output varies. For instance, while in three out of four runs a printer manufacturer was represented as a separate typed entity, in one run it was only expressed as a string literal. • Invention of non-existent properties, prefixes, and classes: If the LLM cannot identify a fully matching class for a concept or a relation, URIs for those elements are invented for the raw RDF output. While this would be possible in its own namespace, the classes and properties are placed in existing namespaces, such as schema.org, resulting in the generated URIs not being resolvable. • Non-functional queries: SPARQL queries generated by ChatGPT -3 did not return the expected results when executed against a knowledge graph, albeit being syntactically correct. All queries needed slight modifications to work, such as correcting the referencing of non-existent classes.</p><p>To conclude, the results obtained by Meyer et al. show that the problems commonly observed with LLMs also limit their ability to conduct tasks in the semantic domain. It is therefore not possible to use LLMs out-of-the-box for semantic relation inference. Since semantic type detection is simpler than semantic relation inference and also relies heavily on obtaining context, this area of automation is investigated more closely.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Semantic Type Detection with LLMs</head><p>To investigate the suitability of LLMs for semantic type detection, we conducted four exemplary experiments using ChatGPT 4.0. Therefore, we manually selected three datasets from the VC-SLAM corpus <ref type="bibr" target="#b36">[37]</ref> which contains datasets created by human modelers in combination with their data description and semantic model as evaluation datasets. Dataset 1 (VC-SLAM 0001) has seven labels that are close to natural language, Dataset 2 (VC-SLAM 0018) has 21 labels that are mostly human readable, and Dataset 3 (VC-SLAM 0068) consists of 24 labels, some of which are abbreviations. The experiments are briefly described in the following and the results are given in Table <ref type="table">1</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1.">Experiments</head><p>Experiment 1 -Mapping to VC-SLAM: This experiment explores ChatGPT 's ability to map dataset labels to the corresponding concepts within the VC-SLAM ontology, provided solely in Turtle (TTL) format, without any additional contextual information. This setup aims to assess the base capability of ChatGPT to utilize the ontology's structure and content for semantic type detection. The task involves presenting dataset labels to ChatGPT and instructing it to identify the most fitting ontology concept for each label.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Prompt Experiment 1</head><p>• You are a tool for semantic type detection. I will provide you an owl ontology that consists of all the concepts you know. This ontology is called VC-SLAM. Later I will additionally provide the labels of three data sets. For each label you return the fitting concept from the ontology. • The first data set consists of the following labels: type, longitude, address, latitude, tvm_identifier, pay_by_credit_card, pay_by_cash Please return the results in the following form: label,concept Experiment 2 -VC-SLAM with Documentation: In this experiment, the methodology is similar to the Mapping to VC-SLAM experiment, but it includes comprehensive documentation of the VC-SLAM ontology and datasets. This tests the hypothesis that additional contextual information enhances the accuracy of semantic type detection.</p><p>Experiment 3 -schema.org Ontology: This experiment shifts the focus to a generalpurpose ontology to compare ChatGPT 's adaptability and performance with a different ontology structure. This experiment provides insights into the model's versatility and the challenges of applying a broad ontology like schema.org to a specific dataset, highlighting differences in specificity and applicability.</p><p>Experiment 4 -Simplified VC-SLAM: The final experiment aims to investigate the impact of ontology complexity on semantic type detection accuracy. By reducing the VC-SLAM ontology to only include concept names and their descriptions without further relations, this experiment seeks to determine whether a simplified ontology framework would enhance ChatGPT 's mapping accuracy due to decreased complexity and ambiguity. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Table 1</head><p>The table demonstrates the results of the four experiments that we run on three datasets. The datasets refer to the following original datasets from VC-SLAM <ref type="bibr" target="#b36">[37]</ref>: 1: 0001, 2: 0018, 3: 0068</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2.">Results</head><p>The outcomes of the experiments are measured across the three different datasets. Experiment 1 -Mapping to VC-SLAM reveals a varying performance with a 57.1% accuracy rate for the first dataset (4 out of 7 labels correctly mapped), 33.3% for the second (7 out of 21), and a notably lower 12.5% for the third (3 out of 24). These results indicate that while ChatGPT can achieve some level of correct mapping based on the ontology structure alone. Experiment 2 -VC-SLAM with Documentation showed improved performance with a 71.4% accuracy rate for the first dataset (5 out of 7 labels correctly mapped), 61.9% for the second (13 out of 21), and 20.8% for the third (5 out of 24). These results highlight the significant impact of extra contextual information in enhancing ChatGPT 's semantic type detection capabilities, leading to more accurate mappings.</p><p>Experiment 3 -schema.org Ontology further demonstrated ChatGPT 's adaptability with impressive accuracies: 100% for the first dataset (7 out of 7 labels correctly mapped), 76.2% for the second (16 out of 21), and 45.8% for the third (11 out of 24). Reasons for this may be that ChatGPT is better at handling ontologies that were already part of the training data, or that the descriptions in schema.org are more meaningful than those of the VC-SLAM ontology.</p><p>Experiment 4 -Simplified VC-SLAM yielded mixed results: 57.1% accuracy for the first dataset (4 out of 7 labels correctly mapped), 42.9% for the second (9 out of 21), and 50% for the third (12 out of 24). These outcomes suggest that simplification of the ontology does not necessarily lead to improved performance across all datasets, reflecting the complex balance between ontology complexity and the effectiveness of semantic type detection with ChatGPT.</p><p>During these experiments, several key findings emerged. First, the availability of additional contexts, such as ontology documentation, significantly improves ChatGPT 's ability to accurately map dataset labels to ontology concepts, underscoring the importance of rich contextual information for semantic type detection tasks. Second, the experiments revealed ChatGPT 's adaptability to different ontologies, with performance variations highlighting the model's capability to handle both specialized and general-purpose ontologies. Lastly, the simplification of the ontology structure was shown to potentially enhance semantic type detection accuracy, suggesting that the complexity of an ontology can affect the efficiency and effectiveness of label mapping. These findings contribute valuable insights into the potential of leveraging LLMs for semantic type detection, indicating promising pathways for automating and refining the data categorization process. The experiments underscore the significance of ontology design and contextual information in optimizing the performance of semantic type detection tasks using AI models like ChatGPT.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Semantic Model Creation with LLMs</head><p>Adding LLMs into the process of semantic model creation provides automation algorithms with the ability to profit from the advantages that these pre-trained models provide. Taking the findings from both related work and our experiments into account, it can be deduced that results obtained from LLMs need to be verified and checked before being applied in a semantic model creation scenario within dataspaces. In the following, two approaches on how to utilize LLM-generated output in automated semantic model creation are conceptualized. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.1.">Unifying KGs with LLMs for Semantic Modeling</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Table 2</head><p>The pros and cons of LLMs vs KGs as described by <ref type="bibr" target="#b37">[38]</ref> Although the technologies for linked data and the Semantic Web have become more mature in recent years, the amount of data considered in Semantic Web applications is far less than in Big Data applications <ref type="bibr" target="#b38">[39]</ref>. Thus, scalability to large, heterogeneous data sets is a major challenge for applying Semantic Web technologies in dataspaces for which LLMs can be a great help. However, even though LLMs can effectively possess rich knowledge learned from massive amounts of training data and benefit downstream tasks at the fine-tuning stage, as previously described, they still have significant limitations due to the lack of external knowledge <ref type="bibr" target="#b16">[17]</ref>. In contrast, KGs are structured knowledge models that explicitly store rich factual knowledge. However, KGs are difficult to construct and evolve by nature, making it challenging to generate new facts and represent unseen knowledge <ref type="bibr" target="#b39">[40]</ref>. Therefore, it is reasonable to view KGs and LLMs as two complementary technologies whose integration has the potential to produce synergy, capitalizing on the strengths of each while mitigating their respective weaknesses.</p><p>For a detailed discussion on research towards the unification of language models and KGs we refer to the survey by Pan et al. <ref type="bibr" target="#b37">[38]</ref>. They contrast the pros and cons of (large) language models vs KGs. Table <ref type="table">2</ref> confirms the previous findings about the drawbacks of LLMs, namely hallucination, a lack of domain-specific knowledge, indecisiveness, and a lack of interpretability. Conversely, the automated construction of KGs is equally challenging, and current approaches to KGs are inadequate in handling the incomplete and dynamically changing nature of realworld KGs. Additionally, many of the current techniques for KGs are tailored to particular tasks and, therefore not easy to generalize to broader applications. This suggests that KG and LLMs indeed complement and may synergize with each other. Pan et al. further predict three main directions for future research toward this goal: KG-enhanced LLMs, which incorporate KGs during the pre-training and inference phases of LLMs to enhance understanding of the knowledge learned by LLMs. Here we direct the interested reader to the survey by Hu et al. <ref type="bibr" target="#b40">[41]</ref>. Then there are LLM-augmented KGs, mentioned in a similar form by Meyer et al. <ref type="bibr" target="#b35">[36]</ref> (see Section 3). Ultimately, these directions may be integrated to produce Synergized LLMs + KGs, in which LLMs and KGs play equal roles and work in a mutually beneficial way to facilitate reasoning driven by both data and knowledge. This fusion may possibly address some of the contemporary challenges discussed in Section 3 and represent one answer to the third research question (RQ3) on how LLMs can enhance the semantic model creation process. Figure <ref type="figure" target="#fig_2">3</ref> illustrates the integration of the two directions and its stated goals. As the most basic semantic unit, entities play a crucial role, and incorporating their knowledge into LLMs helps to improve semantic understanding. In addition, there are also a large number of relational triples in the knowledge graph, which can provide sufficient structured information to further improve the semantic understanding. Since conventional LLMs trained on plain text data are not designed to understand (graph-)structured data such as knowledge graphs, they might not fully grasp or understand the information conveyed by the KG structure. This assumption is confirmed by our experiments (see Section 4), since reducing the representation of ontologies to plain text significantly improves the performance. This indicates that ChatGPT does not handle the graph structure well. Synergized LLMs + KGs promise to be able to understand the underlying graph structure which could improve the performance of KG technology e.g. in discovering unseen facts and exploration for example.</p><p>Multimodal KGs are becoming increasingly important for dataspaces as they integrate different modalities, including text, image, audio, and video data, into a single graph, allowing for a comprehensive representation of complex data <ref type="bibr" target="#b12">[13]</ref>. As previously described, it is important for semantic modeling to have solid numerical reasoning skills. Similar to KGs, regular datasets like CSVs or JSONs can be viewed as (semi-)structured data that represent a further modality. Therefore, effectively leveraging representations from multiple modalities, in particular tables and spreadsheets, would be a significant milestone towards the unification of KGs and LLMs.</p><p>Finally to remedy the problems with hallucinations and updates of the internal knowledge of LLMs as real-world situations change, the incorporation of knowledge from KGs represents a logical solution. For hallucination, the KGs can be leveraged as an external source to validate or fact-check the output of an LLM. Editing the knowledge of an LLM live without re-training is an attractive idea. However, current methods have severe problems, and further research is required <ref type="bibr" target="#b41">[42,</ref><ref type="bibr" target="#b42">43]</ref>. A potential solution to this problem is presented in the next section.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.2.">LLM-Supported Interactive Semantic Model Design</head><p>Since semantic model creation is usually performed inside a semantic modeling platform, integrating LLMs into the semantic model creation following an interactive pattern requires the surrounding platforms to offer additional functions. None of the existing semantic modeling platforms, such as SAND <ref type="bibr" target="#b43">[44]</ref>, MantisTable <ref type="bibr" target="#b44">[45]</ref> or PLASMA <ref type="bibr" target="#b45">[46]</ref>, offer the ability to communicate with a generative AI to refine semantic models. Future platforms to support manual semantic model creation will likely integrate the interaction with an LLM as a central component of their design. For example, a semantic model creation could be fitted into a session with an interactive LLM. Any type of generative LLM can be used, however, using knowledge-enhanced LLMs (see Section 5.1) helps to reduce the effects of unwanted phenomena, such as hallucinations.</p><p>Users input their desired changes using natural language, and the LLM alters the model accordingly. This process requires two central features to be realized: First, users must be able to formulate their changes using a prompt-like interface. Any identified shortcomings can be expressed in text form, even using natural language, to interact with the system and improve the usability of the semantic modeling platform for users with little or no previous knowledge of semantic technologies. Additionally, the platform should provide a process for piping and filtering LLM output to minimize the impact of known drawbacks, such as hallucinations.</p><p>Figure <ref type="figure" target="#fig_3">4</ref> visualizes this process in which the modeler and the LLM serve as the interacting participants of a communication. All interactions between both parties are conducted through various services, such as the modeling platform and the LLM's API. These services apply modifications and transform the contained data to match the other side's data model. For example, when a semantic model generation is requested using an LLM, the current semantic model is provided to the LLM, preferably using a pre-configured, session-based GPT specialized in semantic model creation. The request is appended to the interactive session, resulting in an updated model being generated by the LLM. The LLM's extensive knowledge and advanced capability to process natural speech input allows it to modify the semantic model based on the modeler's intentions, proposing a formalized solution to shortcomings such as syntactical errors. The LLM-generated output undergoes post-processing to ensure presentability to the modeler, particularly when generating large semantic models. The changes made by the LLM in the last iteration are highlighted in separate steps in the generated model, making it easy to identify the changes made based on the last input when displaying the results in the modeling platform. In case the LLM generates corresponding textual output, it is parsed and attached to the updated model using a special set of RDF properties. This enables the modeler to verify the reasoning behind the modifications made to specific elements. Once the post-processing is complete, the proposed semantic model is transferred back to the user and displayed.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.">Conclusion</head><p>This article explores the applicability of modern LLMs for semantic data management in dataspaces, in particular to the tasks of semantic model creation. Our objective is to address the provided research questions to offer directions of future research for preparing LLMs for the complex task of creating semantic models for vast amounts of heterogeneous data sources in dataspaces.</p><p>Regarding RQ1 and RQ2, the experiments in Section 4 demonstrate the feasibility of utilizing LLMs for semantic type detection with a fixed or limited set of labels derived from legacy knowledge graphs. LLMs show promise in achieving significant accuracy in semantic type detection tasks, especially when additional contextual information or documentation is provided alongside the ontology. in particular, Experiment 3, which used the schema.org ontology, showcases the high adaptability and potential of LLMs to accurately map dataset labels to ontology concepts with an accuracy reaching up to 100% for certain datasets. This indicates that LLMs can serve as a powerful tool for semantic type detection. Experiment 4's approach, using a simplified version of the VC-SLAM ontology, offers insight into how LLMs might tackle semantic type detection tasks when the ontology is minimized to basic concept names and descriptions, achieving up to 57.1% accuracy in some cases. The findings suggest that LLMs, including ChatGPT, can effectively engage in semantic type detection tasks even when presented with new, unfamiliar, or arbitrary domain ontologies, by leveraging their inherent understanding of language and context.</p><p>Regarding RQ3, exploiting the vast knowledge and reasoning capabilities of LLMs to automate semantic modeling is an attractive idea. However, significant research is still necessary to integrate KGs with LLMs to produce synergy between these two complementary technologies (see Section 5.1). LLMs do not navigate on graphs or handle numerical data sets well. They may suffer from hallucinations and cannot acquire domain-specific knowledge <ref type="bibr" target="#b46">[47]</ref> easily.</p><p>The LLM-supported interactive semantic model design (see Section 5) establishes a unique way of generating semantic models, providing another possible answer to RQ3 on how LLMs can enhance the semantic model creation process. However, it requires several additions to today's semantic modeling platforms. In theory, the creation of a semantic model can be a fully immersive experience, where modifications can even be made through voice commands. These modifications are then converted to prompts and interpreted by the natural language processing capabilities of LLMs. The resulting changes are automatically visualized, effectively utilizing the LLM as a semantic modeling system. While the presented results and concepts represent a first approach to the topic, the stated research questions remain open to inspire future research in this area.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: Semantic models creating a unified mapping from different data sources (datasets) to an ontology.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figure 2 :</head><label>2</label><figDesc>Figure 2:The semantic model creation process as formalized by<ref type="bibr" target="#b20">[21]</ref>. Specific terms used for available automation are stated in purple.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>Figure 3 :</head><label>3</label><figDesc>Figure 3:The unification of KGs and LLMs. Adopted from<ref type="bibr" target="#b37">[38]</ref>.</figDesc><graphic coords="10,150.55,84.18,291.64,126.96" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_3"><head>Figure 4 :</head><label>4</label><figDesc>Figure 4: Communication-based interactive semantic model generation process</figDesc></figure>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgements</head><p>This work has been sponsored by the German Federal Ministry of Education and Research in the funding program "Forschung an Fachhochschulen" (grant no. 13FH557KX0) and funding program "Datenkompetenzzentren für die Wissenschaft" (grant no. 16DKZ2056B). Declaration of generative AI and AI-assisted technologies in the writing process: During the preparation of this work, the author(s) used OpenAI's generative AI (ChatGPT v3.5 &amp; v4), DeepL and Grammarly to improve the writing, make suggestions, and for rephrasing. After using this tool/service, the author(s) reviewed and edited the content as needed and take(s) full responsibility for the content of the publication.</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">White paper on the data governance act</title>
		<author>
			<persName><forename type="first">J</forename><surname>Baloup</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Bayamlıoğlu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Benmayor</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Ducuing</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Dutkiewicz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Lalova-Spinks</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Miadzvetskaya</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Peeters</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">CiTiP Working Paper</title>
		<imprint>
			<biblScope unit="page">2021</biblScope>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<monogr>
		<author>
			<persName><forename type="first">L</forename><surname>Nagel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">J</forename><surname>Hierro</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Perea</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Lycklama</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Mertens</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A.-S</forename><surname>Taillandier</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Marques</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Gelhaar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Marguglio</surname></persName>
		</author>
		<author>
			<persName><forename type="first">U</forename><surname>Ahle</surname></persName>
		</author>
		<title level="m">Design Principles for Data Spaces: Position Paper</title>
				<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
		<respStmt>
			<orgName>E. ON Energy Research Center</orgName>
		</respStmt>
	</monogr>
	<note type="report_type">Technical Report</note>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">The fair guiding principles for scientific data management and stewardship</title>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">D</forename><surname>Wilkinson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Dumontier</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><forename type="middle">J</forename><surname>Aalbersberg</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Appleton</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Axton</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Baak</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Blomberg</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J.-W</forename><surname>Boiten</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><forename type="middle">B</forename><surname>Da Silva Santos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><forename type="middle">E</forename><surname>Bourne</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Scientific data</title>
		<imprint>
			<biblScope unit="volume">3</biblScope>
			<date type="published" when="2016">2016</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Semantics in dataspaces: Origin and future directions</title>
		<author>
			<persName><forename type="first">J</forename><surname>Theissen-Lipp</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Kocher</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Lange</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Decker</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Paulus</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Pomp</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Curry</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Companion Proceedings of the ACM Web Conference 2023, WWW &apos;23 Companion</title>
				<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<monogr>
		<title level="m" type="main">Semantic Integration and Interoperability</title>
		<author>
			<persName><forename type="first">S</forename><surname>Auer</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2022">2022</date>
			<publisher>Springer International Publishing</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">The web and linked data as a solid foundation for dataspaces</title>
		<author>
			<persName><forename type="first">S</forename><surname>Meckler</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Dorsch</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Henselmann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Harth</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Companion Proceedings of the ACM Web Conference</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Semantic web and knowledge graphs for industry 4.0</title>
		<author>
			<persName><forename type="first">M</forename><surname>Yahya</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">G</forename><surname>Breslin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">I</forename><surname>Ali</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Applied Sciences</title>
		<imprint>
			<biblScope unit="volume">11</biblScope>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<monogr>
		<author>
			<persName><forename type="first">S</forename><surname>Hoseini</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Theissen-Lipp</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Quix</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2310.15373</idno>
		<title level="m">Semantic data management in data lakes</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Applying semantics to reduce the time to analytics within complex heterogeneous infrastructures</title>
		<author>
			<persName><forename type="first">A</forename><surname>Pomp</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Paulus</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Kirmse</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Kraus</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Meisen</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Technologies</title>
		<imprint>
			<biblScope unit="volume">6</biblScope>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Knowledge graphs</title>
		<author>
			<persName><forename type="first">A</forename><surname>Hogan</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">ACM Comput. Surv</title>
		<imprint>
			<biblScope unit="volume">54</biblScope>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Enabling data spaces: existing developments and challenges</title>
		<author>
			<persName><forename type="first">G</forename><surname>Solmaz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Cirillo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Fürst</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Jacobs</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Bauer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Kovacs</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">R</forename><surname>Santana</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Sánchez</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 1st International Workshop on Data Economy, DE &apos;22</title>
				<meeting>the 1st International Workshop on Data Economy, DE &apos;22</meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Using semantic technologies to manage a data lake: Data catalog, provenance and access control</title>
		<author>
			<persName><forename type="first">H</forename><surname>Dibowski</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Schmid</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Svetashova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Henson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Tran</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. Scalable Semantic Web Knowledge Base Systems Workshop</title>
				<meeting>Scalable Semantic Web Knowledge Base Systems Workshop</meeting>
		<imprint>
			<publisher>CEUR WS</publisher>
			<date type="published" when="2020">2020</date>
			<biblScope unit="volume">2757</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Towards multimodal knowledge graphs for data spaces</title>
		<author>
			<persName><forename type="first">A</forename><surname>Usmani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">J</forename><surname>Khan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">G</forename><surname>Breslin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Curry</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Companion Proceedings of the ACM Web Conference 2023</title>
				<imprint/>
	</monogr>
	<note>WWW &apos;23</note>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">Semi: A semantic modeling machine to build knowledge graphs with graph neural networks</title>
		<author>
			<persName><forename type="first">G</forename><surname>Futia</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Vetrò</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">C</forename><surname>De Martin</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">SoftwareX</title>
		<imprint>
			<biblScope unit="volume">12</biblScope>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">Automatic semantic modeling of structured data sources with cross-modal retrieval</title>
		<author>
			<persName><forename type="first">R</forename><surname>Xu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Mayer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Chu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H.-Y</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Feng</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Pattern Recognition Letters</title>
		<imprint>
			<biblScope unit="volume">177</biblScope>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<monogr>
		<title level="m" type="main">Evaluating the logical reasoning ability of chatgpt and gpt-4</title>
		<author>
			<persName><forename type="first">H</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Ning</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Teng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><surname>Zhou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Zhang</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2304.03439</idno>
		<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">A survey on evaluation of large language models</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Chang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Wang</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">ACM Trans. Intell. Syst. Technol</title>
		<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
	<note>Just Accepted</note>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">Semantic Labeling: A Domain-Independent Approach</title>
		<author>
			<persName><forename type="first">M</forename><surname>Pham</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Alse</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">A</forename><surname>Knoblock</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Szekely</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">The Semantic Web -ISWC 2016</title>
				<imprint>
			<publisher>Springer International Publishing</publisher>
			<date type="published" when="2016">2016</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<analytic>
		<title level="a" type="main">Bottom-up Knowledge Graph-based Data Management</title>
		<author>
			<persName><forename type="first">A</forename><surname>Pomp</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Berichte aus dem Maschinenbau</title>
				<meeting><address><addrLine>Shaker</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<analytic>
		<title level="a" type="main">You are missing a concept! enhancing ontology-based data access with evolving ontologies</title>
		<author>
			<persName><forename type="first">A</forename><surname>Pomp</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Lipp</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Meisen</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. ICSC, IEEE</title>
				<meeting>ICSC, IEEE</meeting>
		<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<analytic>
		<title level="a" type="main">Recent advances and future challenges of semantic modeling</title>
		<author>
			<persName><forename type="first">A</forename><surname>Paulus</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Burgdorf</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Pomp</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Meisen</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. 15th IEEE ICSC, IEEE</title>
				<meeting>15th IEEE ICSC, IEEE</meeting>
		<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b21">
	<analytic>
		<title level="a" type="main">Sherlock: A deep learning approach to semantic data type detection</title>
		<author>
			<persName><forename type="first">M</forename><surname>Hulsebos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Hu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Bakker</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 25th ACM SIGKDD</title>
				<meeting>the 25th ACM SIGKDD</meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b22">
	<analytic>
		<title level="a" type="main">Learning Semantic Models of Data Sources Using Probabilistic Graphical Models</title>
		<author>
			<persName><forename type="first">B</forename><surname>Vu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Knoblock</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Pujara</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">The World Wide Web Conference, WWW &apos;19</title>
				<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b23">
	<analytic>
		<title level="a" type="main">Leveraging Linked Data to Infer Semantic Relations within Structured Sources</title>
		<author>
			<persName><forename type="first">M</forename><surname>Taheriyan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">A</forename><surname>Knoblock</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Szekely</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">L</forename><surname>Ambite</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Chen</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 6th International Workshop on Consuming Linked Data</title>
				<meeting>the 6th International Workshop on Consuming Linked Data<address><addrLine>COLD</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2015">2015. 2015</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b24">
	<monogr>
		<author>
			<persName><forename type="first">K</forename><surname>Korini</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Bizer</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2306.00745</idno>
		<title level="m">Column type annotation using chatgpt</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b25">
	<analytic>
		<title level="a" type="main">Survey of hallucination in natural language generation</title>
		<author>
			<persName><forename type="first">Z</forename><surname>Ji</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Lee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Frieske</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Yu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Su</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Xu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Ishii</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><forename type="middle">J</forename><surname>Bang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Madotto</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Fung</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">ACM Comput. Surv</title>
		<imprint>
			<biblScope unit="volume">55</biblScope>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b26">
	<analytic>
		<title level="a" type="main">Annotating columns with pre-trained language models</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Suhara</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Demiralp</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W.-C</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><surname>Tan</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2022 International Conference on Management of Data, SIGMOD &apos;22</title>
				<meeting>the 2022 International Conference on Management of Data, SIGMOD &apos;22</meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b27">
	<monogr>
		<author>
			<persName><forename type="first">P</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>He</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Yashar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Cui</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Ge</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">R</forename><surname>Fainman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Chaudhuri</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2310.09263</idno>
		<title level="m">Table-gpt: Table-tuned gpt for diverse table tasks</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b28">
	<analytic>
		<title level="a" type="main">Tabllm: Few-shot classification of tabular data with large language models</title>
		<author>
			<persName><forename type="first">S</forename><surname>Hegselmann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Buendia</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Lang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Agrawal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Jiang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Sontag</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">International Conference on Artificial Intelligence and Statistics</title>
				<meeting><address><addrLine>PMLR</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b29">
	<monogr>
		<author>
			<persName><forename type="first">T</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Yue</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Sun</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2311.09206</idno>
		<title level="m">Tablellama: Towards open large generalist models for tables</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b30">
	<analytic>
		<title level="a" type="main">Tab2kg: Semantic table interpretation with lightweight semantic profiles</title>
		<author>
			<persName><forename type="first">S</forename><surname>Gottschalk</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Demidova</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Semantic Web</title>
		<imprint>
			<biblScope unit="volume">13</biblScope>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b31">
	<monogr>
		<author>
			<persName><forename type="first">Y</forename><surname>Zhao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Long</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Liu</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2311.09805</idno>
		<title level="m">Docmath-eval: Evaluating numerical reasoning capabilities of llms in understanding long documents with tabular data</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b32">
	<monogr>
		<author>
			<persName><forename type="first">B</forename><forename type="middle">P</forename><surname>Allen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Stork</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Groth</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2310.00637</idno>
		<title level="m">Knowledge engineering using large language models</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b33">
	<monogr>
		<author>
			<persName><forename type="first">L.-P</forename><surname>Meyer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Frey</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Junghanns</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Brei</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Bulert</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Gründer-Fahrer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Martin</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2308.16622</idno>
		<title level="m">Developing a scalable benchmark for assessing large language models in knowledge graph engineering</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b34">
	<monogr>
		<author>
			<persName><forename type="first">J</forename><surname>Frey</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L.-P</forename><surname>Meyer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Arndt</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Brei</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Bulert</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2309.17122</idno>
		<title level="m">Benchmarking the abilities of large language models for rdf knowledge graph creation and comprehension: How well do llms speak turtle?</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b35">
	<monogr>
		<author>
			<persName><forename type="first">L.-P</forename><surname>Meyer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Stadler</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Frey</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2307.06917</idno>
		<title level="m">Llm-assisted knowledge graph engineering: Experiments with chatgpt</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b36">
	<analytic>
		<title level="a" type="main">VC-SLAM -A Handcrafted Data Corpus for the Construction of Semantic Models</title>
		<author>
			<persName><forename type="first">A</forename><surname>Burgdorf</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Paulus</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Pomp</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Meisen</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Data</title>
		<imprint>
			<biblScope unit="volume">7</biblScope>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b37">
	<analytic>
		<title level="a" type="main">Unifying large language models and knowledge graphs: A roadmap</title>
		<author>
			<persName><forename type="first">S</forename><surname>Pan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Luo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Wang</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Transactions on Knowledge and Data Engineering</title>
		<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b38">
	<analytic>
		<title level="a" type="main">An analysis of links in wikidata</title>
		<author>
			<persName><forename type="first">A</forename><surname>Haller</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Polleres</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Dobriy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Ferranti</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">J</forename><surname>Rodríguez Méndez</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">European Semantic Web Conference</title>
				<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b39">
	<analytic>
		<title level="a" type="main">Scaling up knowledge graph creation to large and heterogeneous data sources</title>
		<author>
			<persName><forename type="first">E</forename><surname>Iglesias</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Jozashoori</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M.-E</forename><surname>Vidal</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Web Semantics</title>
		<imprint>
			<biblScope unit="volume">75</biblScope>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b40">
	<analytic>
		<title level="a" type="main">A survey of knowledge enhanced pre-trained language models</title>
		<author>
			<persName><forename type="first">L</forename><surname>Hu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Zhao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Hou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Nie</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Li</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Transactions on Knowledge and Data Engineering</title>
		<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b41">
	<monogr>
		<author>
			<persName><forename type="first">Y</forename><surname>Yao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Tian</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Cheng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Deng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Zhang</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2305.13172</idno>
		<title level="m">Editing large language models: Problems, methods, and opportunities</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b42">
	<monogr>
		<author>
			<persName><forename type="first">R</forename><surname>Cohen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Biran</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Yoran</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Globerson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Geva</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2307.12976</idno>
		<title level="m">Evaluating the ripple effects of knowledge editing in language models</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b43">
	<analytic>
		<title level="a" type="main">SAND : A Tool for Creating Semantic Descriptions of Tabular Sources</title>
		<author>
			<persName><forename type="first">B</forename><surname>Vu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">A</forename><surname>Knoblock</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">The semantic web</title>
				<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2022">2022</date>
			<biblScope unit="volume">13384</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b44">
	<analytic>
		<title level="a" type="main">Mantistable v: A novel and efficient approach to semantic table interpretation</title>
		<author>
			<persName><forename type="first">R</forename><surname>Avogadro</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Cremaschi</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">SemTab@ ISWC</title>
				<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b45">
	<analytic>
		<title level="a" type="main">PLASMA: Platform for Auxiliary Semantic Modeling Approaches</title>
		<author>
			<persName><forename type="first">Alexander</forename><surname>Paulus</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Andreas</forename><surname>Burgdorf</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Lars</forename><surname>Puleikis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Tristan</forename><surname>Langer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">André</forename><surname>Pomp</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Tobias</forename><surname>Meisen</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">International Conference on Enterprise Information Systems</title>
				<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b46">
	<monogr>
		<title level="m" type="main">Large Language Models Struggle to Learn Long-Tail Knowledge</title>
		<author>
			<persName><forename type="first">N</forename><surname>Kandpal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Deng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Roberts</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Wallace</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Raffel</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
