<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>Enschede, The
Netherlands.
$ fneuhaus@ovgu.de (F. Neuhaus); janna.hastings@uzh.ch (J. Hastings)</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>The illusory goal of automating ontology development - With or without large language models (extended abstract)</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Fabian Neuhaus</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Janna Hastings</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Faculty of Medicine, Institute for Implementation Science in Health Care, University of Zurich</institution>
          ,
          <country country="CH">Switzerland</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Institute of Intelligent Cooperating Systems, Otto von Guericke University Magdeburg</institution>
          ,
          <country country="DE">Germany</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>School of Medicine, University of St. Gallen</institution>
          ,
          <country country="CH">Switzerland</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2024</year>
      </pub-date>
      <volume>000</volume>
      <fpage>0</fpage>
      <lpage>0002</lpage>
      <abstract>
        <p>This abstract is based on two papers, namely 'Ontology Development Is Consensus Creation, Not (merely) Representation', Applied ontology 17 (2022) 495-513.1 and 'Ontologies in the Era of Large Language Models - a Perspective', Applied Ontology 18 (2023) 399-407 2. A naïve view of ontology development is the following: to develop an ontology, all that one needs to do is to gather the relevant knowledge about a given domain (through studying textual resources or by talking to an expert in that domain) and then formalise it in some appropriate logical language, such as the Web Ontology Language (OWL) or variants of first-order logic (FOL). From this conception, it is often thought to follow that the whole process of ontology development should be amenable to automation. After all, the formalisation of the content of natural language text in a formal language is akin to a translation task from one natural language to another. Because widely available automated natural language translation tools translate between natural languages, it seems as though the automation of ontology development should be possible. Since large language models (LLM) became available, - according to the same naïve view - even the task of gathering the relevant knowledge is superfluous. Large language models learn the relevant knowledge when they are trained on text corpora. Thus, they can be used to automate the knowledge acquisition phase. Hence, ontology development does not require the painfully slow process of talking to domain experts but instead can be replaced by some clever prompt engineering. This naïve view of ontology development is based on a misunderstanding of what ontologies are, what they contain, and how they are built. In particular, it misconstrues the task of the ontologist as a mere translator, who is tasked with specifying a shared conceptualisation by formalising it in a logical language. It also misconstrues the capabilities of LLMs. In Studer et al. (1998): Knowledge engineering: Principles and methods, ontologies are defined as formal, explicit specifications of some shared conceptualisation , where '[...] an ontology captures consensual knowledge, that is, it is not private to some individual, but accepted by a group'. There are several reasons why an ontology ought to represent a consensus by experts in the domain: (a) if the ontology represents only one point of view on a controversial subject, it reduces the likelihood of adoption by the relevant community, (b) a divergent interpretation of the vocabulary of the ontology by the users will lead to interoperability issues, and (c) if conflicting views are represented unresolved in the same ontology, the ontology will likely be incoherent and, thus, hard to understand for users. The naïve view of ontology development is based on the misconception that such a consensus exists before the ontology engineering process starts. This may be the case in rare situations (e.g. if one is tasked to develop an ontology that represents the content of a standard document), but in the vast majority of ontology development projects, there is no pre-existing consensus. Because in any non-trivial domain, there are divergent views on the domain, diferent opinions, and competing conceptualisations. This situation will be reflected in the literature (or other texts) that are written by experts on the domain, although, often the diferences of opinion will only be implicit. Thus, a major task during ontology development is to create a consensus, which is afterward represented in the ontology. 'Consensus' in this sense does not imply that all issues are settled and that arguments have ceased. Rather, it denotes a working agreement about the types of entities to which the discourses in a given domain are referring and their most important relationships. This consensus is reflected by the meaning postulates in the ontology that determines both the logical and the intended semantics of its vocabulary. Consensus creation is a social process, which cannot be replaced by LLMs. Because if the available literature does not contain a consensus on a given domain, LLMs are not able to learn one. The reason is that LLMs are not designed to resolve divergent points of view and ambiguities, rather they learn to navigate them during their training by adjusting a mapping from texts to probability distributions of tokens appropriately. For this reason, even minor variants of the same prompt or, depending on the configuration, even the same prompt may lead to the generation of contradicting statements. Thus, simple ontological questions like “Is an X a kind of Y?” may be answered by the same LLM both positively and negatively, if both possible views are supported by the literature. Given how LLMs work, we should not expect the output of an LLM to reflect a logically consistent and ontologically sound view of any given domain. Let alone some views that reflect a consensus that does not exist.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>Consensus creation for ontology development is a complex social process, which involves two distinct
steps: reaching agreement between domain experts that are actively engaged in ontology development
(microlevel consensus), and the wider community that the ontology is intended to serve (macro-level consensus).
Consensus creation requires an open, participatory, and transparent ontology development process that engages
the stakeholders of the larger community. Since, by definition, a participatory process cannot be fully automated,
it is not possible to fully automate ontology engineering – with or without LLMs.</p>
    </sec>
  </body>
  <back>
    <ref-list />
  </back>
</article>