<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Large Language Models as Assistants for Ontology Engineering</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Mohammad Javad Saeedizade</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Linköping University</institution>
          ,
          <addr-line>Linköping</addr-line>
          ,
          <country country="SE">Sweden</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2025</year>
      </pub-date>
      <fpage>2</fpage>
      <lpage>6</lpage>
      <abstract>
        <p>Ontology engineering is often a complex, time-consuming and costly process that relies heavily on expert engineers. Even experienced ontology engineers introduce errors, such as incompleteness in terms of requirements, and fall into common ontology pitfalls, underscoring the challenge of producing high-quality ontologies. This PhD proposal aims to address these issues by creating an LLM-based assistant for both ontology development and ontology evaluation. The envisioned assistant will ofer suggestions during conceptual modelling, patternbased suggestions for class and property definitions, and real-time validation checks to identify modelling errors. By embedding these capabilities into a unified tool, the research seeks to reduce dependence on expert intervention, enabling mid-level and novice ontology engineers and organisations to develop reliable ontologies more independently, while simultaneously accelerating the workflow of expert ontologists. The outcome of this work will be a software tool that supports and streamlines the ontology engineering lifecycle-facilitating creation, error detection, and quality assessment-thereby making ontology creation faster, less error-prone, and more accessible to non-experts.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Large Language Models</kwd>
        <kwd>Ontology Development</kwd>
        <kwd>Ontology Evaluation</kwd>
        <kwd>Ontology Engineering</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>
        Ontology engineering is a challenging task that relies heavily on domain experts for both the creation
and the evaluation of ontologies. Even when developed by ontology experts, ontologies frequently
exhibit errors—ranging from pitfalls flagged by OOPS! (OntOlogy Pitfall Scanner!) [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], inaccuracy in the
ontology, logical inconsistencies, completeness, and inadequate modelling of the intended requirements—
that undermine their reliability and reuse. A key step in ensuring that an ontology meets its intended
requirements is competency question (CQ) verification [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ], in which an ontology engineer represents
each CQ as a SPARQL query and evaluates it against the ontology. However, formulating and executing
these queries imposes an additional burden on developers and further raises the barrier to entry for
novice ontology engineers.
      </p>
      <p>
        Meanwhile, many real-world tasks have been (semi-)automated through large language models
(LLMs), such as GitHub Copilot [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. The rapid pace of LLM development continually yields models that
outperform their predecessors on a variety of knowledge-centric tasks. Although a handful of studies
have explored the use of LLMs for discrete aspects of ontology engineering, no comprehensive tool
currently exists to guide users through the end-to-end process of ontology modelling or evaluation.
In this work, we propose tackling two tasks in ontology creation: (i) ontology development and (ii)
ontology evaluation.
      </p>
      <p>
        Based on eXtreme Design (XD), ontology creation is usually an incremental process of developing
and evaluating [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. Ontologists typically begin by framing a set of CQs and using ontology narratives
as contextual background. They then represent the CQs and narrative fragments using OWL (the Web
Ontology Language) [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] and propose a model. In the evaluation phase, they apply both structural
and functional methods. Structural aspects are mostly measured with tools such as OntoMetric [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ],
consistency checking by running reasoners, OOPS! [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], FOOPS! [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ], etc., which report structural
statistics, logical issues, and common mistakes (pitfalls). On the other hand, functional methods
concern functionalities of the ontology, such as the intended modelling. These methods are mostly
time-consuming and manual. One such method is CQ verification, which requires generating SPARQL
queries for each CQ, executing them against the current ontology, interpreting the results to determine
whether the ontology satisfies the intended requirement, and repairing any detected shortcomings in
the next iteration of ontology revision.
      </p>
      <p>In this work, we propose the development of an ontology engineering assistant that ofers
contextaware suggestions during conceptual modelling and semi-automates CQ verification by providing
suggestions as shown in Figure 1. By embedding LLMs guidance and validation into a single pipeline,
the tool is designed to accelerate ontology creation, reduce dependence on expert intervention, and
make ontology engineering more accessible to novice ontology engineers.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Importance</title>
      <p>The complexity of ontology engineering, from creation to maintenance, coupled with the necessity
for expert knowledge engineers, often makes it challenging for organisations to use semantic web
technologies and holds back novice ontology engineers, resulting in increased costs for organisations
and creating an obstacle when it comes to adapting semantic web technologies.</p>
      <p>By introducing an ontology engineering assistant that incorporates LLMs to create suggestions
for ontologists during the development and evaluation phase, this work stands to benefit multiple
stakeholders. Expert ontologists can streamline repetitive processes and reduce the risk of common
errors, while novices can leverage guidance to accelerate their learning and contribute more efectively
to ontology projects. Organisations gain the ability to develop cleaner, more maintainable ontologies
with reduced expert involvement, thereby lowering overhead and fostering a wider adoption of semantic
technologies. Ultimately, this research makes ontology engineering more accessible to companies by
making this task simpler, faster and less costly.</p>
    </sec>
    <sec id="sec-3">
      <title>3. Related work</title>
      <p>
        LLMs for ontology development/generation. Recent studies have used LLMs to draft OWL
ontologies from requirements. For example, Lippolis et al. [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] introduced Ontogenia, prompting LLMs to
generate ontologies from user stories and CQs. In [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] and later in [
        <xref ref-type="bibr" rid="ref10 ref11">10, 11</xref>
        ], we likewise leverage LLMs
to formalise requirements (CQs and user narratives), resulting in ontologies comparable to or better
than those created by novice ontology engineers. Fathallah et al. [
        <xref ref-type="bibr" rid="ref12 ref13">12, 13</xref>
        ] developed NeOn-GPT and
LLMs4Life pipelines for automated ontology modelling without user-based evaluation of the pipeline.
In the work of Alharbi et al.[
        <xref ref-type="bibr" rid="ref14">14</xref>
        ], who developed DIAMOND-KG, we observed that participants’
performance is significantly influenced by the accuracy of the LLM’s predictions. Similarly, in our previous
study[
        <xref ref-type="bibr" rid="ref15">15</xref>
        ], we demonstrated that more accurate knowledge graphs lead to improved performance in
applications leveraging KGs. Mateiu &amp; Groza [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ] fine-tuned GPT-3 to translate natural language
sentences into OWL axioms. They integrated their tool into Protégé, although they provided no proper
evaluation of the tool. However, this tool is the closest thing to what this proposal is proposing.
      </p>
      <p>
        LLMs for ontology evaluation. While the work mentioned in the previous paragraph has
incorporated evaluation for their generated ontologies, evaluation has rarely been the central concern. For
example, in Lippolis et al. [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] we proposed a set of criteria specifically designed for automatically
generated ontologies; however, our framework still depended largely on manual judgments, revealing a
broader need for more automated, scalable evaluation methods. Tsaneva et al. [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ] used GPT-4’s chat
interface (ChatGPT) to compare automatically inferred axioms against human expert assessments, yet
this work remained purely structural and was demonstrated on only a single toy ontology, limiting its
broader applicability. Similarly, Benson et al. [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ] examined GPT-4’s capacity to both produce and
critique class definitions within the Basic Formal Ontology. While they showed that a human-in-the-loop
refinement process can enhance productivity in ontology tasks, their experiments were limited to a
small set of illustrative classes and did not tackle functional evaluation aspects.
      </p>
      <p>
        In our recent study [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ], we investigated the use of LLMs to support ontology engineers in evaluating
whether an ontology adequately addresses a given CQ. Our findings indicate that the performance
of ontology engineers is strongly influenced by the accuracy of the suggestions provided by the
LLM. Specifically, when the LLM ofers a correct suggestion, the evaluators’ performance improves
substantially. Conversely, incorrect suggestions from the LLM lead to a marked decline in performance.
It is important to note that, at this stage, we have not yet developed a tool suitable for deployment in
industrial settings.
      </p>
      <p>Thus, despite promising results, the use of LLMs for ontology evaluation, especially concerning
functional adequacy, remains underdeveloped. In this context, Garijo et al. [20] leave the categorisation
of existing resources on LLM use for ontology evaluation blank, highlighting this gap and pointing to
directions where further research, with appropriate setup configurations, could yield more conclusive
suggestions.</p>
    </sec>
    <sec id="sec-4">
      <title>4. Research questions and hypotheses</title>
      <p>This proposal to create an ontology creation assistant has two parts: (i) ontology development assistance
(ii) ontology evaluation assistance. The main research questions related to the first component of the
tool—ontology development assistance—are related to the capability of LLMs in creating ontologies by
themselves or assisting ontology engineers in the development phase. They are as follows:
• RQ1.1 To what extent can LLMs be used to support the generation of ontologies that meet a
predefined set of requirements? Which LLMs and what prompting techniques are more efective?
• RQ1.2 What evaluation criteria are suitable for evaluating LLM-generated ontologies?
• RQ1.3 What are the strengths and weaknesses of ontologies generated using LLMs?
• RQ1.4 To what extent can LLMs assist ontology engineers in ontology development, and what
are the benefits and drawbacks of a hybrid approach combining LLM suggestions with expert
validation compared to traditional human-only methods?</p>
      <p>Similarly to the development phase, the following research questions examine the capabilities of
LLMs in supporting ontology engineers during the evaluation phase:
• RQ2.1 To what extent can LLMs evaluate ontologies using CQ verification?
• RQ2.2 To what extent can LLMs assist ontology engineers in evaluating ontologies through
CQ verification, and what are the benefits and drawbacks of a hybrid approach combining LLM
suggestions with expert validation compared to traditional human-only methods?
This work hypothesises that LLMs can efectively assist ontology engineers in developing and
evaluating ontologies.</p>
    </sec>
    <sec id="sec-5">
      <title>5. Preliminary results</title>
      <p>There have been several experiments done to answer the research questions to some extent with
reproducible results [21], and some work in the future work section will try to give more complete
answers. The results are divided into two sections: (i) ontology development and (ii) ontology evaluation.</p>
      <sec id="sec-5-1">
        <title>5.1. Ontology development</title>
        <p>The research questions RQ1.1–RQ1.3 have been answered partially based on our previous work.</p>
        <p>
          In our first work, Saeedizade and Blomqvist [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ], we explored RQ1.1 through an automatic ontology
generation pipeline shown in Figure 2. First, we filtered LLMs and prompting techniques on some simple
ontology generation tasks shown in Figure 2a. Then, in the main experiment, Figure 2b, we evaluated
the generated ontologies for the remaining LLMs and prompting techniques from the initial experiment.
This work was submitted in December 2023, and at that time, GPT-4 and the sub-task decomposed
prompting technique could generate ontologies similar to those of novice ontology engineers concerning
the only presented criteria (based on CQ verification).
        </p>
        <p>
          (a) Initial Experiment: Finding the best LLMs and prompting techniques [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ]
        </p>
        <p>(b) Main experiment, evaluating the generated ontologies manually.</p>
        <p>
          To extend this work and provide a better answer for RQ1.1 and also answer RQ1.2 and RQ1.3, we
performed (To be presented in ESWC 2025) [
          <xref ref-type="bibr" rid="ref11">11</xref>
          ]. This work is shown in Figure 3, tackles several
limitations of the previous work [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ] by providing a holistic evaluation of the generated ontologies to
address RQ1.2 and RQ1.3. Also, it shows that the generated ontologies are diferent from human-created
ontologies; therefore, they need a diferent evaluation. We showed that some LLMs generated a lot of
unnecessary classes and properties. Then we showed that counting unnecessary components, OOPS!,
and CQ verification gives comparable evaluation results to expert ontology evaluation results. By
December 2024, o1-preview with the introduced evaluation criteria outperformed novice ontology
engineers and showed promising results to using the generated ontologies as a starting point for
ontology development.
        </p>
        <p>
          There were some risks related to the generalizability of LLMs on domain-specific tasks, which could
result in LLMs performing well on one domain but poorly on another. We performed [
          <xref ref-type="bibr" rid="ref10">10</xref>
          ] (To be
presented at ELMKE workshop at ESWC 2025) to investigate this risk, and we showed LLMs perform
similarly on six domains used in the work. We also showed that some ontology development tasks
that were considered complex, LLMs performed them with the same performance (concerning CQ
verification) as simple ones.
        </p>
      </sec>
      <sec id="sec-5-2">
        <title>5.2. Ontology Evaluation</title>
        <p>
          We developed a prototype assistant for ontology evaluation, as described in our recent work [
          <xref ref-type="bibr" rid="ref19">19</xref>
          ]
and illustrated in Figure 4. The primary objective of this assistant is to facilitate both automated and
semi-automated verification of CQs.
        </p>
        <p>In the upper part of Figure 4, a CQ, its context, and an ontology are provided to an LLM, which
is then prompted to assess whether the ontology correctly models the given CQ. The LLM responds
with a binary answer (yes or no), which is subsequently compared against a gold standard to evaluate
performance. Using the o1-preview model, this automated approach achieved a macro-F1 score of
0.68.</p>
        <p>The lower part of the figure illustrates the semi-automated evaluation process. In this setup, we
presented the output generated by the LLM to ontology engineers, who were then asked to determine
whether the ontology adequately models the CQ, using the LLM’s suggestion as guidance with the
ontology opened in Protégé. We deliberately mixed both correct and incorrect LLM suggestions
across CQs that were either correctly or incorrectly modelled in the ontology. The results reveal
that participants’ performance was significantly afected by the accuracy of the LLM’s predictions.
Specifically, correct LLM suggestions improved human performance by 13%, while incorrect suggestions
caused a decline of 26%. However, because the LLM provided more correct suggestions than incorrect
ones, the opposing efects largely cancelled each other out, resulting in no net improvement in human
performance.</p>
        <p>In future work, we plan to extend this tool by incorporating visualisations and additional interactive
features, guided by the feedback received during our experimental evaluation.</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>6. Evaluation</title>
      <p>To evaluate the hypothesis in our chosen setup, the usefulness of LLMs in assisting ontology engineers
in developing and evaluating ontologies should be assessed by means of a user-based study. We
conducted some experiments to partially answer the mentioned research questions through a manual
evaluation of LLMs’ outputs by ontology engineers. Furthermore, we created a prototype to measure
users’ performance only in ontology evaluation with CQ verification. Overall, the main evaluation of
tools for ontology engineering with LLMs as assistants should be done by measuring users’ performance
in ontology engineering, with and without LLMs’ suggestions, and comparing their performance in
each task.</p>
      <p>After creating a tool for ontology engineering, we should hire/invite ontology engineers to develop
and evaluate ontologies following our setup. To evaluate the users’ performance using the tool, we will
measure how accurately and eficiently ontology engineers perform the task of ontology engineering
with the tool when (i) LLMs’ suggestions are available to a user, and (ii) without LLMs’ suggestions.
The performance can be measured by measuring time to show how long it took users to complete
the task, and assessing how accurate users were in each setting regarding ontology development and
evaluation metrics in line with the selected CQ, such as measuring the proportion of correctly modelled
CQs, quality of the final ontology concerning OOPS! warnings and expert evaluation.</p>
    </sec>
    <sec id="sec-7">
      <title>7. Reflection and future work</title>
      <p>In this paper, I propose creating an ontology engineering assistant tool. This tool helps ontologists to
develop and evaluate ontologies by leveraging LLMs. Specifically, the approach relies on suggestions
from LLMs to address relevant CQ throughout the ontology development process. For the ontology
development part of the tool, the next step involves designing the appropriate setup and creating a
tool, followed by defining a task that systematically measures its efectiveness via human-in-the-loop
evaluation. Furthermore, during the ontology evaluation phase, we will capitalise on the user feedback
we received to further enhance the tool by incorporating additional features and refined suggestions.</p>
    </sec>
    <sec id="sec-8">
      <title>Declaration on Generative AI</title>
      <p>The author declares the use of generative AI tools in preparing this manuscript. Specifically, ChatGPT
were used for sentence polishing, paraphrasing, and grammar fixing; all AI-generated text was reviewed,
edited, and approved by the author, who takes full responsibility for the final content.</p>
    </sec>
    <sec id="sec-9">
      <title>Acknowledgments</title>
      <p>I would like to express my gratitude to my supervisors, Eva Blomqvist and Robin Keskisärkkä, for their
guidance, encouragement, and expertise throughout this work. Their insights and support have been
instrumental in shaping both the direction and the quality of this work.
[20] D. Garijo, M. Poveda-Villalón, E. Amador-Domínguez, Z. Wang, R. García-Castro, O. Corcho, Llms
for ontology engineering: A landscape of tasks and benchmarking challenges (2022).
[21] M. J. Saeedizade, R. Alharbi, H. B. Giglou, A. S. Lippolis, E. Blomqvist, V. Tamma, F. Grasso,
T. R. Payne, J. D’Souza, S. Auer, et al., A framework for assessing llm consistency in knowledge
engineering, 2025.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>M.</given-names>
            <surname>Poveda-Villalón</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Gómez-Pérez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. C.</given-names>
            <surname>Suárez-Figueroa</surname>
          </string-name>
          ,
          <article-title>OOPS! (OntOlogy Pitfall Scanner!): An On-line Tool for Ontology Evaluation</article-title>
          ,
          <source>International Journal on Semantic Web and Information Systems (IJSWIS) 10</source>
          (
          <year>2014</year>
          )
          <fpage>7</fpage>
          -
          <lpage>34</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>E.</given-names>
            <surname>Blomqvist</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. Seil</given-names>
            <surname>Sepour</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Presutti</surname>
          </string-name>
          ,
          <article-title>Ontology testing-methodology and tool, in: Knowledge Engineering and Knowledge Management: 18th International Conference</article-title>
          , EKAW 2012,
          <string-name>
            <given-names>Galway</given-names>
            <surname>City</surname>
          </string-name>
          , Ireland, October 8-
          <issue>12</issue>
          ,
          <year>2012</year>
          . Proceedings 18, Springer,
          <year>2012</year>
          , pp.
          <fpage>216</fpage>
          -
          <lpage>226</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>S.</given-names>
            <surname>Peng</surname>
          </string-name>
          , E. Kalliamvakou,
          <string-name>
            <given-names>P.</given-names>
            <surname>Cihon</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Demirer</surname>
          </string-name>
          ,
          <article-title>The impact of ai on developer productivity: Evidence from github copilot</article-title>
          ,
          <source>arXiv preprint arXiv:2302.06590</source>
          (
          <year>2023</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>V.</given-names>
            <surname>Presutti</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Daga</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Gangemi</surname>
          </string-name>
          , E. Blomqvist,
          <article-title>extreme design with content ontology design patterns</article-title>
          ,
          <source>in: Proc. Workshop on Ontology Patterns, CEUR-WS</source>
          ,
          <year>2009</year>
          , pp.
          <fpage>83</fpage>
          -
          <lpage>97</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>D. L.</given-names>
            <surname>McGuinness</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Van Harmelen</surname>
          </string-name>
          , et al.,
          <article-title>Owl web ontology language overview</article-title>
          ,
          <source>W3C recommendation 10</source>
          (
          <year>2004</year>
          )
          <year>2004</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>B.</given-names>
            <surname>Lantow</surname>
          </string-name>
          , Ontometrics:
          <article-title>Putting metrics into use for ontology evaluation</article-title>
          .,
          <source>in: KEOD</source>
          ,
          <year>2016</year>
          , pp.
          <fpage>186</fpage>
          -
          <lpage>191</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>D.</given-names>
            <surname>Garijo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Corcho</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Poveda-Villalón</surname>
          </string-name>
          ,
          <article-title>Foops!: An ontology pitfall scanner for the fair principles</article-title>
          ., in: ISWC (Posters/Demos/Industry),
          <year>2021</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>A. S.</given-names>
            <surname>Lippolis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Ceriani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Zuppiroli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. G.</given-names>
            <surname>Nuzzolese</surname>
          </string-name>
          , Ontogenia:
          <article-title>Ontology generation with metacognitive prompting in large language models</article-title>
          ,
          <source>in: European Semantic Web Conference</source>
          , Springer,
          <year>2024</year>
          , pp.
          <fpage>259</fpage>
          -
          <lpage>265</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>M. J.</given-names>
            <surname>Saeedizade</surname>
          </string-name>
          , E. Blomqvist,
          <article-title>Navigating ontology development with large language models</article-title>
          , in: A.
          <string-name>
            <surname>Meroño Peñuela</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Dimou</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Troncy</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          <string-name>
            <surname>Hartig</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Acosta</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Alam</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          <string-name>
            <surname>Paulheim</surname>
          </string-name>
          , P. Lisena (Eds.),
          <source>The Semantic Web</source>
          , Springer Nature Switzerland, Cham,
          <year>2024</year>
          , pp.
          <fpage>143</fpage>
          -
          <lpage>161</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>A. S.</given-names>
            <surname>Lippolis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. J.</given-names>
            <surname>Saeedizade</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Keskisarkka</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Gangemi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Blomqvist</surname>
          </string-name>
          ,
          <string-name>
            <surname>A. G.</surname>
          </string-name>
          <article-title>Nuzzolese, Assessing the capability of large language models for domain-specific ontology generation</article-title>
          ,
          <source>arXiv preprint arXiv:2504.17402</source>
          (
          <year>2025</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>A. S.</given-names>
            <surname>Lippolis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. J.</given-names>
            <surname>Saeedizade</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Keskisärkkä</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Zuppiroli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Ceriani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Gangemi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Blomqvist</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. G.</given-names>
            <surname>Nuzzolese</surname>
          </string-name>
          ,
          <article-title>Ontology generation using large language models</article-title>
          ,
          <source>arXiv preprint arXiv:2503.05388</source>
          (
          <year>2025</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>N.</given-names>
            <surname>Fathallah</surname>
          </string-name>
          ,
          <string-name>
            <surname>A. Das</surname>
            ,
            <given-names>S. D.</given-names>
          </string-name>
          <string-name>
            <surname>Giorgis</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Poltronieri</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Haase</surname>
          </string-name>
          , L. Kovriguina,
          <article-title>Neon-gpt: a large language model-powered pipeline for ontology learning</article-title>
          ,
          <source>in: European Semantic Web Conference</source>
          , Springer,
          <year>2024</year>
          , pp.
          <fpage>36</fpage>
          -
          <lpage>50</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>N.</given-names>
            <surname>Fathallah</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Staab</surname>
          </string-name>
          ,
          <string-name>
            <surname>A. Algergawy,</surname>
          </string-name>
          <article-title>Llms4life: Large language models for ontology learning in life sciences</article-title>
          ,
          <year>2024</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>R.</given-names>
            <surname>Alharbi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>U.</given-names>
            <surname>Ahmed</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Dobriy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Łajewska</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Menotti</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. J.</given-names>
            <surname>Saeedizade</surname>
          </string-name>
          ,
          <string-name>
            <surname>M.</surname>
          </string-name>
          <article-title>Dumontier, Exploring the role of generative ai in constructing knowledge graphs for drug indications with medical context, in: 15th International Semantic Web Applications and Tools for Healthcare and Life Sciences (SWAT4HCLS</article-title>
          <year>2024</year>
          )
          <article-title>(to appear</article-title>
          ),
          <year>2024</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <surname>M. J. Saeedizade</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          <string-name>
            <surname>Torabian</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Minaei-Bidgoli</surname>
          </string-name>
          ,
          <article-title>Kgrefiner: Knowledge graph refinement for improving accuracy of translational link prediction methods</article-title>
          ,
          <source>arXiv preprint arXiv:2106.14233</source>
          (
          <year>2021</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>P.</given-names>
            <surname>Mateiu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Groza</surname>
          </string-name>
          ,
          <article-title>Ontology engineering with large language models</article-title>
          ,
          <source>in: 2023 25th International Symposium on Symbolic and Numeric Algorithms for Scientific Computing (SYNASC)</source>
          , IEEE,
          <year>2023</year>
          , pp.
          <fpage>226</fpage>
          -
          <lpage>229</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>S.</given-names>
            <surname>Tsaneva</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Vasic</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Sabou</surname>
          </string-name>
          ,
          <article-title>Llm-driven ontology evaluation: Verifying ontology restrictions with chatgpt</article-title>
          ,
          <source>The Semantic Web: ESWC Satellite Events</source>
          <year>2024</year>
          (
          <year>2024</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>C.</given-names>
            <surname>Benson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Sculley</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Liebers</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Beverley</surname>
          </string-name>
          ,
          <article-title>My ontologist: Evaluating bfo-based ai for definition support</article-title>
          ,
          <source>arXiv preprint arXiv:2407.17657</source>
          (
          <year>2024</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>A. S.</given-names>
            <surname>Lippolis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. J.</given-names>
            <surname>Saeedizade</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Keskisärkkä</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Gangemi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Blomqvist</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. G.</given-names>
            <surname>Nuzzolese</surname>
          </string-name>
          ,
          <source>Large language models assisting ontology evaluation</source>
          ,
          <year>2025</year>
          . URL: https://arxiv.org/abs/2507.14552. doi:
          <volume>10</volume>
          .48550/arXiv.2507.14552. arXiv:
          <volume>2507</volume>
          .14552, submitted July 19,
          <year>2025</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>