<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <issn pub-type="ppub">1613-0073</issn>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>KONDA: An LLM-based Tool for Semantic Annotation and Knowledge Graph Creation Using Ontologies for Research Data</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Soo-Yon Kim</string-name>
          <email>kim@dbis.rwth-aachen.de</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Martin Görz</string-name>
          <email>martin.goerz@rwth-aachen.de</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Sandra Geisler</string-name>
          <email>geisler@dbis.rwth-aachen.de</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Workshop</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>5th International Workshop on Scientific Knowledge: Representation, Discovery, and Assessment</institution>
          ,
          <addr-line>Nov 2025, Nara</addr-line>
          ,
          <country country="JP">Japan</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Data Stream Management and Analysis, RWTH Aachen University</institution>
          ,
          <addr-line>Ahornstr. 55, Aachen, 52074</addr-line>
          ,
          <country country="DE">Germany</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2004</year>
      </pub-date>
      <fpage>3</fpage>
      <lpage>18</lpage>
      <abstract>
        <p>The increasing demand for FAIR (Findable, Accessible, Interoperable, and Reusable) research data management (RDM) practices has underscored the need for tools that support semantic annotation and structuring of datasets. However, integrating ontologies and knowledge graphs into research workflows often presents a technical, cognitive, and time-consuming challenge for researchers. We investigate how large language models (LLMs) can be leveraged to lower this barrier by guiding researchers through the ontology-aligned annotation and transformation of heterogeneous research data into knowledge graphs. To this end, we design and implement a modular, domain-agnostic tool that is able to process and annotate various types of research data by combining methods such as vector-based semantic matching, human-in-the-loop validation, and integration with terminology lookup services and graph databases. The tool provides a guided end-to-end workflow, from data and context ingestion to entity and relation extraction, ontology-based annotation, and graph assembly, designed to support non-expert users through an intuitive, lightweight web interface. Results from a usability survey show that the tool is considered practical and useful. We argue that LLMs, when embedded in an interactive and user-centered system, can significantly enhance the accessibility and efectiveness of semantic research data management. Future work may investigate tool performance in real-world settings.</p>
      </abstract>
      <kwd-group>
        <kwd>semantic annotation</kwd>
        <kwd>knowledge graphs</kwd>
        <kwd>ontologies</kwd>
        <kwd>large language models</kwd>
        <kwd>research data management</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>Ontologies for Research</p>
    </sec>
    <sec id="sec-2">
      <title>1. Introduction</title>
      <p>https://sandra-geisler.de/ (S. Geisler)</p>
      <p>CEUR</p>
      <p>ceur-ws.org
concepts to use and how to apply them often requires an initial coordination efort among collaborators,
particularly in interdisciplinary projects. This adds another layer of complexity to ontology-based work,
even before any technical implementation begins.</p>
      <p>Although a wide range of ontologies and semantic web frameworks and tools exist1,2,3, their
integration into everyday research workflows remains limited. The tools frequently require data in specific
formats or assume a high level of data quality, which is often unrealistic in heterogeneous research
contexts. Further, many tools are designed for specific domains and therefore ofer limited applicability
outside their original scope. Lastly, solutions are often technically cumbersome to operate and ofer
limited support for users without prior experience in semantic technologies. As a result, much of
the research data produced today remains poorly described, which limits its potential for integration,
discovery, and reuse across domains.</p>
      <p>Addressing this gap calls for a tool that provides low-barrier access to semantic technologies, supports
diverse data types and domains, and enables the gradual annotation and transformation of data without
demanding extensive upfront modeling. Large language models (LLMs) ofer promising capabilities in
this regard, as they can be operated with natural language, are able to process various types of data,
and can conduct semantic matching.</p>
      <p>Building on these considerations, we developed KONDA (Knowledge graph creation and ONtology
enrichment for research DAta), a tool that builds on LLM functionalities to facilitate semantic annotation
and knowledge graph creation in research contexts. It supports users in transforming research data
into semantically enriched graph representations through a guided, end-to-end workflow. The system
integrates modular key functionalities into a single web interface, including data upload, terminology
lookup, ontology embedding, entity and relation extraction, semantic similarity matching, graph
visualization, and export options, with continuous human-in-the-loop validation.</p>
      <p>
        The contributions of the tool presented in this paper are as follows. (
        <xref ref-type="bibr" rid="ref1">1</xref>
        ) The tool is agnostic with
respect to both domain and data format. (
        <xref ref-type="bibr" rid="ref2">2</xref>
        ) It enables the reuse of existing ontologies. (3) Semantic
annotation is semi-automated through a combination of LLM-based recommendations and
human-inthe-loop validation. (4) A knowledge graph construction component supports the visual exploration of
the semantically enriched data. (5) All steps are unified under a single web interface. (6) The modular
architecture supports self-hosting of the tool and the deployment of various LLMs.
      </p>
      <p>To assess the tool’s usability and practical value, we conducted a feasibility demonstration and a
user study focusing on the accessibility of the workflow, the quality of the generated suggestions,
and the overall fit with real-world research practices. The results indicate that KONDA functions
as intended, and is perceived as useful and approachable, even by users without prior experience in
semantic technologies. The findings support the potential of LLM-based systems to lower the entry
threshold for semantic data management and to foster broader adoption of FAIR-aligned practices.</p>
      <p>The remainder of the paper is structured as follows. Section 2 reviews related work in ontology-based
annotation and knowledge graph creation, and the application of language models in semantic research
data management. Section 3 develops a solution concept. Section 4 describes the system architecture and
implementation of KONDA. Section 5 presents the evaluation design and results. Section 6 discusses key
insights and limitations. Section 7 summarizes the paper and gives implications for future development.</p>
    </sec>
    <sec id="sec-3">
      <title>2. Related Work</title>
      <p>Various tools and approaches have been developed that address semantic technologies for research data
management using LLMs. These works cover areas such as ontology-based annotation, knowledge
graph generation, and holistic semantic research data management systems. In the following, we review
related work and position it in relation to the goals and design of KONDA.</p>
      <sec id="sec-3-1">
        <title>1https://protege.stanford.edu/ (accessed on 25.07.2025) 2https://bioportal.bioontology.org/annotator (accessed on 25.07.2025) 3https://inception-project.github.io/ (accessed on 25.07.2025)</title>
        <p>Ontology-based annotation. Recent works have explored how LLMs can support the semantic
annotation tasks of entity and relation recognition and their linking to ontology concepts. Tools such
as BioLinkerAI [10] and frameworks such as NSSC [11] focus on the medical domain, where entities
identified in clinical or biomedical texts are linked to terms of the Unified Medical Language
System4. NeOn-GPT [12] combines a systematic ontology engineering methodology with prompt-based
interactions to translate domain descriptions into formal representations. OntoKGen [13] provides an
interactive workflow for identifying domain-relevant concepts and relations in technical documents
and helps users formalize these into custom ontologies. These approaches apply diferent methods:
Some combine rule-based methods for entity and relationship extraction with LLM-based methods for
selecting and disambiguating candidate concepts based on contextual information. Other systems show
how LLMs can support the construction of semantic resources from text. However, there are some
limitations. Many tools rely on the existence of predefined rules and established ontologies, requiring a
substantial amount of a-priori symbolic knowledge and making the solutions not suitable for use across
diverse research domains. In the works surrounding ontology integration, there is a focus primarily on
generating new ontologies, and little support is ofered for integrating existing ontologies into data
annotation workflows.</p>
        <p>Knowledge graph creation. Further works explore how LLMs can support the construction of
knowledge graphs. Yang et al. [14] have proposed a method for having LLMs expand ontologies based on
competency questions and subsequently construct knowledge graphs. In OntoKGen [13], ontologies can
be custom-built and then be used to generate knowledge graphs. Vizcarra et al. [15] have developed a
system for constructing a knowledge graph which describes user-product interactions. While these
approaches present efective solutions for knowledge graph construction, they underscore the importance
of ontology integration in ensuring that the resulting graphs are semantically meaningful. However,
their reliance on domain-specific prior work for the integration of ontologies limits the scalability of
the suggested pipelines across diferent domains.</p>
        <p>End-to-end semantic data management systems. Few works address the holistic (semi-)automated
support of semantic research data management. The Leibniz Data Manager [16] integrates automatic
components for the annotation of data with concepts e.g. from Wikidata5, its integration into a
knowledge graph, and querying functions. Further systems such as AGENTiGraph [17] and the
materials science platform NOMAD [18], while ofering functionalities for semantic data management,
focus the automated support rather to downstream tasks such as analysis. Hence, there is a general lack
of systems that provide end-to-end, (semi-)automated pipelines for semantic research data management
workflows.</p>
        <p>While these existing systems cover a range of tasks related to semantic annotation and graph
construction, important aspects remain underdeveloped. First, few approaches support the semantic
transformation of heterogeneous research data in a way that is applicable across domains and data
formats. Second, existing tools rarely facilitate the reuse of established ontologies during annotation,
often focusing on the creation of new ontologies instead. Third, available solutions are often fragmented
across diferent tools and interfaces, lacking a unified system for end-to-end ontology-based annotation
and knowledge graph creation, with a specific focus on low-barrier usability for time- and
resourceconstrained researchers. Together, these gaps motivate the development of a system that combines
these components into a single, accessible workflow.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>3. Methodology</title>
      <p>To address these gaps, we propose KONDA, a tool designed to support researchers in structuring their
data using ontologies and generating machine-readable knowledge graphs.The conceptual foundation
of KONDA centers around a modular pipeline architecture, in which LLMs are embedded into a
transparent, human-in-the-loop workflow. The core design goal is to preserve semantic rigor while</p>
      <sec id="sec-4-1">
        <title>4https://www.nlm.nih.gov/research/umls/index.html (accessed on 23.07.2025) 5www.wikidata.org/ (accessed on 25.07.2025)</title>
        <p>Context Files</p>
        <p>Readme
Entity</p>
        <p>Tomato</p>
        <p>Assigned Broader Category</p>
        <sec id="sec-4-1-1">
          <title>Edible Plant ▼ ✔ X</title>
          <p>4 Relation Extraction
Subject</p>
          <p>Tomato</p>
          <p>Verb
grown in ▼</p>
          <p>Object</p>
          <p>Greenhouse
✔ X</p>
          <p>Tomato</p>
          <p>Plant
has concept</p>
          <p>...
has related match</p>
          <p>plant_
dataset.zip
has concept</p>
          <p>has concept Greenhouse
reducing cognitive and technical overhead, enabling users to guide, validate, and refine the semantic
transformation of their data.</p>
          <p>
            At the highest level, the tool guides the researcher through a sequence of seven conceptual stages,
depicted in Figure 1: (
            <xref ref-type="bibr" rid="ref1">1</xref>
            ) data ingestion, (
            <xref ref-type="bibr" rid="ref2">2</xref>
            ) ontology selection, (3) entity and (4) relation extraction,
(5) ontology-based annotation, (6) structured knowledge graph assembly, and (7) export. Each stage
is conceptually modular, with clearly defined intermediate results that are presented to the user in
natural language for inspection and revision, the mechanism of which is depicted in Figure 6 in Section
A of the appendix. This design ensures not only transparency and autonomy for researchers, but also
adaptability to diferent research domains and varying degrees of semantic web familiarity.
          </p>
          <p>
            The process begins with the ingestion of heterogeneous research materials. As illustrated in stage
(
            <xref ref-type="bibr" rid="ref1">1</xref>
            ) of Figure 1, users upload both a research dataset and a collection of supplementary documentation
(e.g., README files, related publications, notes, or protocols) that contextualize the dataset. Because
raw research data often lacks suficient semantic context, the tool uses few-shot prompting with an
LLM to generate summaries of the accompanying documentation [19]. This stage creates a compact,
coherent representation of the research context, which is reused across subsequent tasks to ground
entity recognition, relation modeling, and ontology alignment. This design decision emerged from our
early stage experiments, where the presence or absence of context dramatically afected the relevance
and specificity of extracted entities.
          </p>
          <p>
            Once the semantic context has been established, users proceed to the ontology selection stage (
            <xref ref-type="bibr" rid="ref2">2</xref>
            ).
To support this stage, the tool ofers a keyword-based ontology search interface integrated with a
terminology lookup service6 to query relevant domain ontologies. Researchers may select ontologies
from this search or upload custom ontologies as needed. This stage ensures that subsequent extraction
and alignment tasks are grounded in vocabularies that reflect the intended domain.
          </p>
          <p>In stage (3), KONDA prompts the LLM to identify a set of relevant entities that capture the key
concepts in the dataset, guided by both the extracted context and the chosen ontologies. To structure
these entities more efectively, they are grouped into broader semantic categories that provide scafolding
for interpretation and alignment. For example, in a community gardening dataset, the tool may extract
“Chamomile” as an entity and assign it the category “Herbal Plant”. These categories improve ontology
mapping in stage (5). The extracted entities and categories are presented to the user through an</p>
        </sec>
      </sec>
      <sec id="sec-4-2">
        <title>6https://terminology.tib.eu/ts/ontologies (accessed on 25.07.2025)</title>
        <p>interactive interface. Each extraction can be reviewed, edited, or extended, ensuring human-in-the-loop
validation. Users may adjust the number of extracted items to control complexity as well as granularity.
The inputs remain in natural language at this stage, therefore lowering the entry barrier while guiding
alignment in downstream stages to ensure both usability and semantic consistency.</p>
        <p>Next, the tool identifies relationships between the validated entities in stage (4). Prompted again
with contextual information and the current list of entities and categories, the LLM proposes subject,
predicate, object triples that represent meaningful semantic links. Relationship labels are initially
suggested in plain language, such as “is grown in” or “is used for”, and are subject to user revision. Only
entities validated in the previous stage are used as subject and object nodes, ensuring referential integrity
across the graph. This iterative model avoids ambiguous relationships and allows the researcher to use
their use case-specific knowledge to validate results.</p>
        <p>While entity and relation extraction form the conceptual backbone of the knowledge graph,
representing them as free-form text alone does not guarantee interoperability. To address this, the tool includes
a dedicated ontology-based annotation stage. Here, each free-form entity, category, and relationship is
mapped to a formal ontology term using a two-step strategy, shown in stage (5) in Figure 1. First, the
tool performs a vector-based similarity search: embeddings are computed for both user-defined terms
and ontology labels (including descriptions) using an embedding model. Second, these candidates are
matched through LLM-assisted reasoning: the model evaluates conceptual fit between the user term
and the ontology candidate, taking into account the dataset’s context summaries. Final suggestions are
presented to the user, who can approve, override, or reject each mapping. If no suitable match is found,
users may fall back to their original term, preserving semantic intent while maintaining a clear record
of unmatched items.</p>
        <p>A key design consideration is to support a flexible representation of user-defined terms and their
ontology alignments in the final graph. During annotation, users can choose whether to replace their
original label with the matched ontology term, which is useful when an exact or highly appropriate match
is found, or to retain the original label and link it to the ontology term via a looser semantic relation (e.g.,
skos:relatedMatch). This approach maintains both formal interoperability and the expressive nuance
of domain-specific terminology, accommodating cases where exact matches do not exist or where
preserving original phrasing is important for interpretation.</p>
        <p>The final stages are the knowledge graph assembly stage (6), in which all prior components, entities,
categories, relations, and ontology annotations are combined into a structured, machine-readable
graph, and the export stage (7). The graph is centered around a dataset node that links to contextual
documents, extracted entities (coined concepts), and their broader categories. Edges between entities
reflect validated relationships, and each node optionally links to an ontology term via explicit match
relations. The schema is depicted in Figure 7 in Section A of the appendix and is designed to be
compatible with SKOS7 and other common vocabularies, ensuring downstream compatibility with
query engines, metadata registries, and discovery platforms. All outputs are exportable and conform to
standard formats, supporting integration into broader scientific knowledge graph infrastructures.</p>
        <p>In summary, KONDA supports ontology-based annotation and knowledge graph generation by
embedding LLMs into an interactive, user-centered pipeline. It transforms heterogeneous, informal
research data into semantically structured, ontology-aligned graphs while keeping the researcher in
control at every step. The next section details the implementation of this conceptual workflow, including
tool architecture, backend components, and the orchestration of LLM tasks and user interaction.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>4. Implementation</title>
      <p>Building upon the conceptual workflow outlined in the previous chapter, the tool was implemented as
a modular web-based application designed to be a user-friendly and guided tool for researchers. The
implementation emphasizes accessibility, responsiveness, and adaptability to diferent domains and
dataset formats. These priorities are reflected in both the tool’s architecture and its choice of technologies.</p>
      <p>Context Dataset
Paper
Uploads Data
Researcher</p>
      <p>htmx
User Interface</p>
      <p>Neo4j
Graph DB</p>
      <p>VM / Cluster</p>
      <p>Server Rendered UI</p>
      <p>Files, Configs,</p>
      <p>Feedback</p>
      <p>Go Backend
store KG
and Onto</p>
      <p>provides
semantic similarity</p>
      <p>API
provides text embeddings</p>
      <p>LLM Server</p>
      <p>OpenAI /</p>
      <p>Ollama
(Self-) Hosted
Structured Generated</p>
      <p>Responses
LLM
Prompts</p>
      <p>This chapter details how the conceptual components, i.e., file preprocessing, contextual enrichment,
semantic structuring, ontology alignment, and knowledge graph construction, were translated into a
fully operational tool, while maintaining human-in-the-loop validation.</p>
      <p>The application follows a client-server architecture in which responsibilities are clearly separated
between a dynamic, form-driven web frontend and a processing-intensive backend. This separation
allows for asynchronous background processing while keeping the user experience lightweight and
responsive. Researchers interact with the tool through a stepwise interface composed of dynamically
updated forms, while the backend handles file extraction, interaction with LLMs, vector similarity
searches, and graph database operations, as shown in Figure 2.</p>
      <p>The frontend was developed using a minimalist stack that avoids complex frameworks, in favor
of technologies that support a sequential interaction model. A combination of HTMX8 and Go’s9
native templating system is used to render dynamic forms without requiring a single-page application
framework. HTMX enables partial page updates triggered by user input or server responses, keeping
the interface reactive without heavy client-side logic. TailwindCSS10 and AlpineJS11 are used to provide
styling and minor interactivity, such as toggling inputs or dynamically updating sections. This approach
focuses on providing a guided user interaction through sequentiality while allowing for dynamic
feedback and validation at each stage of the pipeline.</p>
      <p>The backend is implemented in Go and is responsible for all computational tasks. The server manages
session-specific workspaces, processes uploaded files, orchestrates background tasks, generates and
interprets LLM prompts, and constructs the final knowledge graph in a graph database. Each researcher
session is isolated, with its own temporary storage and database instance, ensuring data privacy and
system stability.</p>
      <p>Once a user initiates a session by uploading files, the backend begins processing them asynchronously.
A file type detection mechanism is used to determine the appropriate file extractor for each input.
Extractors are format-specific and capable of operating in either full or sampled mode depending on
ifle size. For example, large CSV files are processed by extracting headers and a sample of rows to
preserve structure without exceeding LLM input limits, while PDF or Markdown files are converted
into plain text for summarization. These extracted segments are then fed into structured LLM prompts</p>
      <sec id="sec-5-1">
        <title>8https://htmx.org/ (accessed on 23.07.2025)</title>
        <p>9https://go.dev/ (accessed on 23.07.2025)
10https://tailwindcss.com/ (accessed on 23.07.2025)
11https://alpinejs.dev/ (accessed on 23.07.2025)
designed to generate dataset and context summaries. The resulting summaries form the semantic basis
for downstream processing and are reused across prompts to reduce cost and redundancy.</p>
        <p>To manage these asynchronous operations, the tool introduces a task orchestration layer in which
each transformation step (e.g., summarization, extraction, or alignment) is defined as a discrete task
with explicit dependencies. If a user modifies an input or updates a configuration, dependent tasks are
automatically invalidated and marked for recomputation. This approach ensures that all outputs remain
consistent with the current state of the data while supporting a responsive and fluid user experience.</p>
        <p>LLM integration is abstracted behind an API layer supporting OpenAI and Ollama style API calls12.
Prompts are constructed using Go templates that encapsulate few-shot examples, context, and output
format instructions. For each major transformation task, i.e., entity recognition, relation extraction,
and ontology annotation, the tool constructs structured few-shot prompts in a consistent format. The
output is always constrained to JSON, enabling reliable parsing and integration into the backend. While
the tool was primarily tested with Azure-hosted models13 for production stability, the API interface
supports any OpenAI-compatible provider, including self-hosted alternatives with Ollama. This allows
researchers to select between commercial and private deployment scenarios based on their requirements
for performance, cost, or data sensitivity.</p>
        <p>A key implementation detail lies in the handling of ontology alignment. All user-uploaded ontologies,
as well as those selected through the in-app ontology search interface, are parsed and stored in a
dedicated Neo4j14 graph database. Neo4j was selected for its ability to eficiently store and traverse
graph-structured data, its flexible node and edge modeling, and its capability to easily generate and
store text embeddings. Each ontology class, attribute, or property, and, if available, its description, is
embedded using OpenAI’s text-embedding-3-large15 model. These embeddings are stored in a vector
index within the graph database and used to retrieve semantically similar candidates when annotating
extracted entities and relations. A two-level structure is used: a session-specific vector store for selected
domain-specific ontologies, and a shared foundational store that covers domain-agnostic vocabularies.
This layered approach ensures broad term coverage while prioritizing relevant domain-specific matches.</p>
        <p>The matching process combines semantical and reasoning-based techniques. First, cosine similarity
is used to retrieve candidate matches from the vector stores. The matches are ranked by their similarity
score. Then, the best ranked candidates are passed to an LLM along with the dataset and context
summaries. The LLM selects the most appropriate ontology term for each extracted element, based on
semantic fit. This dual-step matching process balances scalability with contextual understanding and
supports the alignment of informal labels to formal vocabularies across disciplines.</p>
        <p>The final knowledge graph is constructed using a Neo4j instance dedicated to each user session. This
graph captures the structure and semantics of the dataset, enriched by context and ontology alignment.
Nodes are created for the dataset itself, context files, extracted entities, categories, and matched ontology
terms. Relationships such as “has concept”, “has broader”, and predicate triples between entities are
encoded as edges, using either matched ontology URIs or generated URIs based on the user’s configured
namespace. Where applicable, SKOS vocabulary is used to represent hierarchical relationships and
ontology matches, enabling future inference or reasoning.</p>
        <p>Neo4j also provides the infrastructure for in-app visualization and export. The final graph is rendered
in an interactive viewer, allowing researchers to explore and verify the result. Export options include
RDF/XML16, Turtle17, and JSON-LD18 formats, ensuring compatibility with semantic web standards.</p>
        <p>To manage computational resources, the tool implements a cleanup routine that removes
sessionspecific data, including files, summaries, vector indexes, and Neo4j instances, after a period of inactivity.
This ensures scalability across users while maintaining a consistent and responsive experience.
12https://ollama.com/blog/openai-compatibility (accessed on 23.07.2025)
13https://azure.microsoft.com/en-us/products/ai-services/openai-service (accessed on 23.07.2025)
14https://neo4j.com/ (accessed on 23.07.2025)
15https://platform.openai.com/docs/models/text-embedding-3-large (accessed on 26.09.2025)
16https://www.w3.org/TR/rdf-syntax-grammar/ (accessed on 23.07.2025)
17https://www.w3.org/TR/turtle/ (accessed on 23.07.2025)
18https://json-ld.org/ (accessed on 23.07.2025)</p>
        <p>A collage of the diferent screens of the interface is depicted in Figure 3.</p>
        <p>In summary, the implementation efectively realizes the conceptual framework outlined previously
by combining minimalistic, sequential user interaction with powerful backend automation. Through a
modular architecture, robust background processing, provider-flexible LLM integration, and
ontologyaware graph construction, the tool transforms the abstract process of semantically enriching data into a
guided and transparent application for researchers. The next chapter evaluates this implementation
with a feasibility and a user study to assess its applicability and usability.</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>5. Evaluation</title>
      <p>To assess the feasibility, efectiveness, and usability of the proposed system, we conducted two
complementary evaluations. First, we examined whether the tool can successfully transform realistic research
data into semantically enriched knowledge graphs using a structured example scenario. Second, we
carried out a user study to evaluate both the usability and the acceptance of the tool among researchers
from diverse disciplinary backgrounds.
5.1. Functional Demonstration of the Transformation Pipeline</p>
      <p>Use
Tea</p>
      <p>Salad</p>
      <p>To assess whether the tool delivers on its core objective of transforming research data into semantically
structured, ontology-aligned knowledge graphs, we conducted a detailed example run using a synthetic
dataset. The aim of this demonstration is to evaluate its efectiveness when applied to a realistic,
interpretable scenario.</p>
      <p>The dataset originates from a fictional community gardening initiative, containing records of edible
plants, such as “Mint” and “Tomato”, along with their typical uses, growing locations, and the individuals
who recommended them. An excerpt is given in Table 1, and the full dataset is provided in Table 2 in
Section B of the appendix. This CSV file is accompanied by a short README that provides background
on the dataset, clarifies the meaning of each column, and gives narrative examples to contextualize the
data. The README file is given in Section B.2 of the appendix.</p>
      <p>Once uploaded, the tool processed both files and generated semantic interpretations without further
user input. For instance, plant names were correctly identified as content-bearing entities, and grouped
under the broader label “Edible Plant”. Similarly, values like “Greenhouse”, and “Urban Garden” were
grouped as “Growing Locations”, while “Tea” and “Salad” were categorized as “Use Cases”. These
extractions proved coherent and appropriate when reviewed in the context of the dataset and accompanying
description.</p>
      <p>Crucially, the tool also inferred meaningful relations between these entities. For example, it proposed
triples such as “Tomato – is grown in – Greenhouse” and “Mint – used in – Tea”, correctly capturing
relationships that are implicit in the tabular structure and narrative context. These relations were
both syntactically well-formed and semantically valid, indicating that the pipeline accurately captured
relevant domain kno.w.. ledge. ... ...</p>
      <p>To evaluate semantic enrichment, we selected two ontologies for alignment: the Common Greenhouse
Ontology (CGO)19 and the BioAssay Ontology (BAO)20. The tool matched “Greenhouse” directly to
“CGO:Greenhouse”, and successfully aligned “Tomato” with “CGO:Tomato Plant”. Additionally, the
19https://gitlab.com/ddings/common-greenhouse-ontology/-/blob/main/Ontologies/CGO-v2.4-current-stable-version.ttl?ref_
type=heads (accessed on 23.07.2025)
20https://bioportal.bioontology.org/ontologies/BAO (accessed on 23.07.2025)
predicate “grown in” was matched to “BAO:BAO_0002924”, demonstrating that the tool not only finds
appropriate classes, but also meaningful property mappings. Importantly, when concepts such as “Mint”
lacked a suitable match in the selected ontologies, the tool retained them using a custom identifier,
ensuring no information was discarded in the process.</p>
      <p>The resulting knowledge graph consisted of 27 nodes and 53 edges. An excerpt is shown in Figure 4.
Each extracted element was either matched to an ontology term or clearly retained with its original label.
Relations between entities were labeled with readable, accurate predicates, and semantic groupings
such as categories were preserved to enhance interpretability. The graph was rendered interactively and
could be exported in standard semantic web formats, as depicted fully in Section B.3 of the appendix.</p>
      <p>This example run confirms that the tool can produce coherent, semantically enriched knowledge
graphs from real-world inspired data with minimal intervention, setting the stage for a deeper evaluation
of quality and usability in the subsequent sections.
5.2. Usability and Acceptance</p>
      <p>SUS Score
by Ontology Familiarity</p>
      <p>SUS Score
100 by Knowledge Graph Familiarity 100</p>
      <p>SUS Score
by Research Area
100
80
60
e
r
o
c
S
S
U
S 40
20
0</p>
      <p>Mean
SUS Score</p>
      <p>To assess whether the tool meets its research objective of providing a user-friendly experience
suitable for researchers of varying backgrounds, a structured usability study was conducted. The
evaluation targeted researchers from diverse disciplines and difering levels of expertise with semantic
technologies, and focused on determining whether the researchers could successfully complete the
semantic enrichment pipeline and found the tool intuitive, understandable, and efective in practice.</p>
      <p>A total of 10 participants from various domains, including biology, computer science, materials
science, and economics, took part in the study. All participants were afiliated with RWTH Aachen
University and had prior exposure to working with research data. The study was conducted in three
phases: a questionnaire on the participant’s background, a guided example task using a sample dataset,
and a structured post-task survey. The example task required participants to upload the aforementioned
plant dataset and its context, run the pipeline through entity extraction, ontology-based annotation,
and graph export, and inspect the results.</p>
      <p>To quantitatively assess usability, the System Usability Scale (SUS) was used alongside additional
Likert-scale questions that addressed more specific aspects of usability, functionality, and perceived
impact [20]. The SUS yielded a mean score of 76, surpassing the commonly accepted usability benchmark
of 68 [21]. This result indicates strong overall usability across participants. Furthermore, the average
rating for overall usability was 8.8 out of 10 (SD: 0.42), with 9 being the most commonly selected value,
suggesting high satisfaction with the interface and workflow.</p>
      <p>A central result for evaluating the accessibility of the tool is shown in Figure 5, which splits SUS
scores by participants’ familiarity with ontologies, knowledge graphs, and research domain. Most
notably, participants who self-reported being “Not at all familiar” with ontologies or knowledge graphs
still gave high usability ratings, comparable to those with moderate or high familiarity. This suggests
that we reached the objective of lowering the barrier to entry for non-experts. The guided interface,
structured prompts, and stepwise validation mechanisms enabled users with no prior exposure to
semantic technologies to complete the full pipeline efectively.</p>
      <p>The figure also highlights diferences across disciplinary backgrounds. While researchers from
biology and other non-informatics fields reported the highest SUS scores, users from computer science
also rated the tool positively. These results suggest that the tool’s usability is robust across varying
research domain backgrounds.</p>
      <p>Likert-scale responses further reinforced these results. Participants rated the intuitiveness of the
interface at 4.6/5 (SD: 0.52), the clarity of instructions at 4.5/5 (SD: 0.71), and reported they would feel
comfortable to use the tool again in future projects (mean: 4.5/5, SD: 0.53). Most users agreed that they
were able to complete the task without issues and understood the purpose of each pipeline step.</p>
      <p>Free-text feedback echoed these positive trends. Participants highlighted the ease of reviewing
extracted entities and the clarity of the interface. Suggested improvements focused on secondary
aspects rather than core functionality, such as improving inline help, refining the ontology selection
experience, and more clearly separating entities from relations during annotation.</p>
      <p>In summary, the survey results strongly support the conclusion that the tool enables non-expert users
to create semantically enriched knowledge graphs with minimal onboarding through the transparent,
modular, and well-guided interface. High SUS scores across diverse familiarity levels and research
backgrounds, combined with positive qualitative feedback, support that the tool is both usable and
accessible.</p>
    </sec>
    <sec id="sec-7">
      <title>6. Discussion</title>
      <p>The evaluation results indicate that KONDA significantly contributes to low-barrier ontology-based
semantic annotation and knowledge graph construction. The tool provides an accessible interface
that allows researchers from diferent domains and with varying levels of technical expertise to create
semantically enriched representations of scientific data. High usability scores across disciplines and
user backgrounds support the system’s applicability for use in various research contexts.</p>
      <p>A particular strength of KONDA is its capacity to process heterogeneous input formats and support
semantic structuring without requiring prior experience with semantic web technologies. The
integration of key annotation and transformation steps, such as terminology lookup, ontology alignment, and
graph construction, into a single interface enables a guided workflow that helps translate informal and
semantically unstructured data into formal representations. The reuse of existing ontologies contributes
to interoperability even in the absence of prior agreement on shared terminologies.</p>
      <p>Nonetheless, certain limitations remain. The quality of semantic matching between identified entities
and ontology concepts depends on the capabilities of the underlying language model. While this
approach is efective in many cases, it can produce imprecise or overly general matches, particularly
when relevant concepts are underrepresented in the selected ontology. In addition, although the tool
is designed to support data from diferent domains, its efectiveness depends on the availability and
coverage of suitable ontologies. Domains with limited ontology support may require additional efort
to achieve meaningful annotations.</p>
      <p>The tool’s performance with larger datasets and more complex graph structures also remains a
challenge. While the manual validation step is important, it introduces additional efort for users which
grows with the size of the dataset. While the current implementation supports minimal interactive use,
future developments should consider optimization methods for scalability.</p>
      <p>Lastly, the current system ofers few functionalities for supporting the exploration of the resulting
knowledge graphs. Advanced post-processing or querying mechanisms on the knowledge graph remain
therefore underexplored.</p>
    </sec>
    <sec id="sec-8">
      <title>7. Conclusion</title>
      <p>This paper introduced KONDA, a tool for semantic annotation and knowledge graph construction for
research data. It leverages the integration of LLMs to provide an accessible, user-friendly workflow. The
tool supports researchers in processing heterogeneous data, reusing existing ontologies, and producing
structured semantic representations without requiring prior expertise in semantic technologies. An
initial user study confirmed the usability and applicability of the system across a range of diferent user
contexts.</p>
      <p>Future work may focus on evaluating the tool in real-world settings. Improvement of domain-specific
performance may be explored by integrating recommendation mechanisms for ontologies. In addition,
the implementation of persistent ontology embeddings may reduce redundancy, improve eficiency, and
enhance matching quality in repeated use scenarios. The integration of further functionalities for the
support of FAIR-aligned practices may be considered, such as enabling FAIR Digital Objects as export
formats. Finally, expanding the tool’s capabilities to support tasks such as semantic search, filtering, or
integration with reasoning engines may increase its value for further research workflows.</p>
    </sec>
    <sec id="sec-9">
      <title>Acknowledgments</title>
      <p>Funded by the Deutsche Forschungsgemeinschaft (DFG, German Research Foundation) under Germany’s
Excellence Strategy – EXC-2023 Internet of Production – 390621612.</p>
    </sec>
    <sec id="sec-10">
      <title>Declaration on Generative AI</title>
      <p>During the preparation of this work, the authors used GPT-4o in order to: Paraphrase and reword.
After using this tool/service, the authors reviewed and edited the content as needed and take full
responsibility for the publication’s content.
[3] H. Rahman, M. I. Hussain, A comprehensive survey on semantic interoperability for internet of
things: State–of–the–art and research challenges, Transactions on Emerging Telecommunications
Technologies 31 (2020). doi:10.1002/ett.3902.
[4] B. H. de Mello, S. J. Rigo, C. A. Da Costa, R. Da Rosa Righi, B. Donida, M. R. Bez, L. C. Schunke,
Semantic interoperability in health records standards: a systematic literature review, Health and
technology 12 (2022) 255–272. doi:10.1007/s12553-022-00639-w.
[5] L. Westhofen, C. Neurohr, M. Butz, M. Scholtes, M. Schuldes, Using ontologies for the
formalization and recognition of criticality for automated driving, IEEE Open Journal of Intelligent
Transportation Systems 3 (2022) 519–538. doi:10.1109/OJITS.2022.3187247.
[6] S. Appukuttan, L. L. Bologna, F. Schürmann, M. Migliore, A. P. Davison, Ebrains live papers
interactive resource sheets for computational studies in neuroscience, Neuroinformatics 21 (2023)
101–113. doi:10.1007/s12021-022-09598-z.
[7] T. Benson, G. Grieve, Why interoperability is hard, in: T. Benson, G. Grieve (Eds.), Principles
of Health Interoperability, Health Information Technology Standards, Springer International
Publishing, Cham, 2021, pp. 21–40. doi:10.1007/978-3-030-56883-2{\textunderscore}2.
[8] A. O. Alkhamisi, M. Saleh, Ontology opportunities and challenges: Discussions from semantic
data integration perspectives, in: 2020 6th Conference on Data Science and Machine Learning
Applications (CDMA), IEEE, 2020, pp. 134–140. doi:10.1109/CDMA47397.2020.00029.
[9] S. Reichmann, T. Klebel, I. Hasani-Mavriqi, T. Ross-Hellauer, Between administration and research:
Understanding data management practices in an institutional context, Journal of the Association
for Information Science and Technology 72 (2021) 1415–1431. doi:10.1002/asi.24492.
[10] A. Sakor, K. Singh, M.-E. Vidal, Biolinkerai: Capturing knowledge using llms to enhance biomedical
entity linking, in: M. Barhamgi, H. Wang, X. Wang (Eds.), Web Information Systems Engineering
– WISE 2024, volume 15439 of Lecture Notes in Computer Science, Springer Nature Singapore,
Singapore, 2025, pp. 262–272. doi:10.1007/978-981-96-0573-6{\textunderscore}19.
[11] Á. García-Barragán, A. Sakor, M.-E. Vidal, E. Menasalvas, J. C. S. Gonzalez, M. Provencio, V. Robles,
Nssc: a neuro-symbolic ai system for enhancing accuracy of named entity recognition and linking
from oncologic clinical notes, Medical &amp; biological engineering &amp; computing 63 (2025) 749–772.
doi:10.1007/s11517-024-03227-4.
[12] N. Fathallah, A. Das, S. de Giorgis, A. Poltronieri, P. Haase, L. Kovriguina, Neon-gpt: A large
language model-powered pipeline for ontology learning, in: A. Meroño Peñuela, O. Corcho,
P. Groth, E. Simperl, V. Tamma, A. G. Nuzzolese, M. Poveda-Villalón, M. Sabou, V. Presutti,
I. Celino, A. Revenko, J. Raad, B. Sartini, P. Lisena (Eds.), The Semantic Web: ESWC 2024 Satellite
Events, volume 15344 of Lecture Notes in Computer Science, Springer Nature Switzerland, Cham,
2025, pp. 36–50. doi:10.1007/978-3-031-78952-6{\textunderscore}4.
[13] M. S. Abolhasani, R. Pan, Ontokgen: A genuine ontology and knowledge graph generator using
large language model, in: 2025 Annual Reliability and Maintainability Symposium (RAMS), IEEE,
2025, pp. 1–6. doi:10.1109/RAMS48127.2025.10935139.
[14] H. Yang, L. Xiao, R. Zhu, Z. Liu, J. Chen, An llm supported approach to ontology and knowledge
graph construction, in: 2024 IEEE International Conference on Bioinformatics and Biomedicine
(BIBM), IEEE, 2024, pp. 5240–5246. doi:10.1109/BIBM62325.2024.10822222.
[15] J. Vizcarra, S. Haruta, M. Kurokawa, Representing the interaction between users and products
via llm-assisted knowledge graph construction, in: 2024 IEEE 18th International Conference on
Semantic Computing (ICSC), IEEE, 2024, pp. 231–232. doi:10.1109/ICSC59802.2024.00043.
[16] A. Sakor, M. Brunet, E. Iglesias, A. Rivas, P. D. Rohde, A. Kraft, M.-E. Vidal, Integrating knowledge
graphs and neuro-symbolic ai: Ldm enables fair and federated research data management, in:
W. Nejdl, S. Auer, O. Karras, M. Cha, M.-F. Moens, M. Najork (Eds.), Proceedings of the Eighteenth
ACM International Conference on Web Search and Data Mining, ACM, New York, NY, USA, 2025,
pp. 1044–1047. doi:10.1145/3701551.3704125.
[17] X. Zhao, M. Blum, R. Yang, B. Yang, L. M. Carpintero, M. Pina-Navarro, T. Wang, X. Li, H. Li, Y. Fu,
R. Wang, J. Zhang, I. Li, Agentigraph: An interactive knowledge graph platform for llm-based
chatbots utilizing private data, 2024. URL: https://arxiv.org/abs/2410.11531. arXiv:2410.11531.</p>
    </sec>
    <sec id="sec-11">
      <title>A. Conceptual Design Details</title>
      <p>Question Decomposition Prompting
1. Named Entity Recognition
2. Relation Extraction
3. Ontology Based Annotation</p>
      <p>Task Statement
t
p
m
Po Modeling Knowledge
r
t
o
-hS Output Format
w
e
F</p>
      <p>Input
Response</p>
      <p>Task Statement
Modeling Knowledge</p>
      <p>Output Format</p>
      <p>Input</p>
      <p>Response
Human-in-the-Loop Validation</p>
      <p>Task Statement
Modeling Knowledge</p>
      <p>Output Format</p>
      <p>Input
Response</p>
    </sec>
    <sec id="sec-12">
      <title>B. Full Synthetic Example</title>
      <p>B.1. CSV File
Running example dataset listing plants, their utilized parts, associated use, recommender names, and locations
B.2. README File: Community Garden Edible Plants Log
This dataset is based on notes collected from a local community gardening initiative, where members
share their favorite edible plants, how they use them, and where they grow them.</p>
      <p>Each entry includes:
• The name of the plant,
• the part of the plant that is commonly used (e.g., leaf, fruit, root, or flower),
• what it is typically used for (e.g., culinary, herbal, or fresh consumption),
• the name of the person who recommended the plant,
• and the growing location (e.g., greenhouse, balcony, or countryside).</p>
      <sec id="sec-12-1">
        <title>For example:</title>
        <p>• Emily Rose grows mint and spinach in an urban garden and uses them for tea and salad.
• Sofia Bloom gardens on the balcony and recommends basil and lemon balm for cooking and tea.
• Noah Fields contributes countryside plants like carrots and beets, both used in hearty meals and
fresh dishes.</p>
        <p>The dataset reflects diverse preferences, gardening styles, and cultural uses for everyday edible plants.
B.3. Knowledge Graph Export
ex:tea skos:broader ex:use-case .
ex:beetroot &lt;http://www.bioassayontology.org/bao#BAO_0002924&gt; ex:greenhouse;</p>
        <p>skos:broader ex:edible-plant .</p>
      </sec>
    </sec>
    <sec id="sec-13">
      <title>C. Online Resources</title>
      <p>The original implementation of the tool is available as a Coscine resource from the corresponding author
upon reasonable request. The interface of the tool is available from the corresponding author to users
with institutional credentials upon reasonable request. External access to the interface is restricted due
to security requirements.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>M. D.</given-names>
            <surname>Wilkinson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Dumontier</surname>
          </string-name>
          ,
          <string-name>
            <given-names>I. J. J.</given-names>
            <surname>Aalbersberg</surname>
          </string-name>
          , G. Appleton,
          <string-name>
            <given-names>M.</given-names>
            <surname>Axton</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Baak</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Blomberg</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.-W.</given-names>
            <surname>Boiten</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L. B.</given-names>
            <surname>Da Silva Santos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. E.</given-names>
            <surname>Bourne</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Bouwman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. J.</given-names>
            <surname>Brookes</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Clark</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Crosas</surname>
          </string-name>
          ,
          <string-name>
            <given-names>I.</given-names>
            <surname>Dillo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Dumon</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Edmunds</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C. T.</given-names>
            <surname>Evelo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Finkers</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Gonzalez-Beltran</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. J. G.</given-names>
            <surname>Gray</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Groth</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Goble</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. S.</given-names>
            <surname>Grethe</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Heringa</surname>
          </string-name>
          , P. A. C. '
          <string-name>
            <surname>t Hoen</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Hooft</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Kuhn</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Kok</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Kok</surname>
            ,
            <given-names>S. J.</given-names>
          </string-name>
          <string-name>
            <surname>Lusher</surname>
            ,
            <given-names>M. E.</given-names>
          </string-name>
          <string-name>
            <surname>Martone</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Mons</surname>
            ,
            <given-names>A. L.</given-names>
          </string-name>
          <string-name>
            <surname>Packer</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Persson</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Rocca-Serra</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Roos</surname>
            , R. van Schaik,
            <given-names>S.-A.</given-names>
          </string-name>
          <string-name>
            <surname>Sansone</surname>
            , E. Schultes,
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Sengstag</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Slater</surname>
            , G. Strawn,
            <given-names>M. A.</given-names>
          </string-name>
          <string-name>
            <surname>Swertz</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Thompson</surname>
            ,
            <given-names>J. van der</given-names>
          </string-name>
          <string-name>
            <surname>Lei</surname>
            , E. van Mulligen,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Velterop</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Waagmeester</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Wittenburg</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          <string-name>
            <surname>Wolstencroft</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Zhao</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Mons</surname>
          </string-name>
          ,
          <article-title>The fair guiding principles for scientific data management and stewardship</article-title>
          ,
          <source>Scientific data 3</source>
          (
          <year>2016</year>
          )
          <article-title>160018</article-title>
          . doi:
          <volume>10</volume>
          .1038/sdata.
          <year>2016</year>
          .
          <volume>18</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>D.</given-names>
            <surname>Hodapp</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Hanelt</surname>
          </string-name>
          ,
          <article-title>Interoperability in the era of digital innovation: An information systems research agenda</article-title>
          ,
          <source>Journal of Information Technology</source>
          <volume>37</volume>
          (
          <year>2022</year>
          )
          <fpage>407</fpage>
          -
          <lpage>427</lpage>
          . doi:
          <volume>10</volume>
          .1177/ 02683962211064304.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>L.</given-names>
            <surname>Sbailò</surname>
          </string-name>
          , Á. Fekete,
          <string-name>
            <given-names>L. M.</given-names>
            <surname>Ghiringhelli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Schefler</surname>
          </string-name>
          ,
          <article-title>The nomad artificial-intelligence toolkit: turning materials-science data into knowledge and understanding</article-title>
          ,
          <source>npj Computational Materials</source>
          <volume>8</volume>
          (
          <year>2022</year>
          ). doi:
          <volume>10</volume>
          .1038/s41524- 022- 00935- z.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>M.</given-names>
            <surname>Agrawal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Hegselmann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Lang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Kim</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Sontag</surname>
          </string-name>
          ,
          <article-title>Large language models are few-shot clinical information extractors</article-title>
          ,
          <year>2022</year>
          . URL: https://arxiv.org/abs/2205.12689. arXiv:
          <volume>2205</volume>
          .
          <fpage>12689</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [20]
          <string-name>
            <surname>John</surname>
            <given-names>Brooke,</given-names>
          </string-name>
          <article-title>Sus: A 'quick and dirty' usability scale</article-title>
          , in: P. W.
          <string-name>
            <surname>Jordan</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Thomas</surname>
            ,
            <given-names>I. L.</given-names>
          </string-name>
          <string-name>
            <surname>McClelland</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          Weerdmeester (Eds.), Usability Evaluation In Industry, CRC Press,
          <year>1996</year>
          , pp.
          <fpage>207</fpage>
          -
          <lpage>212</lpage>
          . doi:
          <volume>10</volume>
          . 1201/
          <fpage>9781498710411</fpage>
          -
          <lpage>35</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>J. R.</given-names>
            <surname>Lewis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Sauro</surname>
          </string-name>
          ,
          <article-title>Item benchmarks for the system usability scale</article-title>
          .,
          <source>Journal of Usability studies 13</source>
          (
          <year>2018</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          <article-title>ex:growing-location skos:relatedMatch obo</article-title>
          :AEON_
          <fpage>0000155</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          <article-title>ex:urban-garden skos:relatedMatch obo:AEON_0000155; skos:broader ex:growing-location .</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          <article-title>ex:carrot skos:broader ex:edible-plant;</article-title>
          &lt;http://www.bioassayontology.org/bao#BAO_0002924&gt; ex:countryside .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>