<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>R. Alharbi);</journal-title>
      </journal-title-group>
      <issn pub-type="ppub">1613-0073</issn>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Methods for Competency Question Elicitation from Ontology Requirements</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Reham Alharbi</string-name>
          <email>R.Alharbi@liverpool.ac.uk</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Jacopo de Berardinis</string-name>
          <email>Jacopo.De-Berardinis@liverpool.ac.uk</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Terry R. Payne</string-name>
          <email>T.R.Payne@liverpool.ac.uk</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Valentina Tamma</string-name>
          <email>V.Tamma@liverpool.ac.uk</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Workshop</string-name>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>University of Liverpool</institution>
          ,
          <addr-line>Liverpool</addr-line>
          ,
          <country country="UK">UK</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2025</year>
      </pub-date>
      <volume>000</volume>
      <fpage>0</fpage>
      <lpage>0001</lpage>
      <abstract>
        <p>Competency Questions (CQs) are used guide ontology development, yet formulating them in such a way as to align them to the stakeholder needs remains challenging. This paper presents a comparative analysis of three CQ elicitation methods: manual authoring by ontology engineers; template-based instantiation; and automated generation using diferent LLMs (GPT-4.1, Gemini 2.5). Each CQ is evaluated across dimensions of suitability, readability, and complexity. To facilitate this evaluation we introduce AskCQ, a dataset of 204 CQs derived from a shared user story in the cultural heritage domain. Our results show that manually authored CQs are consistently more acceptable, readable, and concise. LLM-generated CQs are more complex and diverse but require refinement. These findings highlight the importance of human expertise and suggest potential hybrid approaches.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
    </sec>
    <sec id="sec-2">
      <title>Methodology</title>
      <p>To investigate the impact of diferent Competency Question (CQ) elicitation strategies on the
characteristics of the resulting questions, we conducted a comparative analysis across three representative</p>
      <p>CEUR</p>
      <p>ceur-ws.org
approaches: fully manual authoring; template-based instantiation; and automatic generation via Large
Language Models (LLMs). Each approach was applied independently to the same ontology requirement
source to ensure a fair and controlled comparison.</p>
      <p>
        1. Manual (Human-Authored): Two ontology engineers (HA-1 and HA-2), each with over five
years of professional experience in ontology design and requirement engineering, independently
read and interpreted the same user story. Based solely on their expert understanding of the
personas, goals, and informational needs described therein, each formulated a set of CQs without
constraints on format or style. This condition serves as the expert-driven baseline and reflects
common manual practice in ontology development.
2. Template-Based (Pattern Instantiation): An ontology engineer with similar domain
experience instantiated a curated set of 19 CQ patterns derived from Ren et al. [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. These patterns use
archetypal structures such as “Which [CE1] [OPE] [CE2]?” and “Is the [CE1] [CE2]?” and were
manually populated with entities and relations extracted from the user story. The instantiation
process required the identification of suitable fillers from the story content, and their mapping
to the syntactic slots defined by the patterns. This semi-automated method ofers structured
linguistic support but limited flexibility.
3. LLM-Based (Generative AI): Two state-of-the-art LLMs — GPT-4.1 and Gemini 2.5 Pro — were
prompted to generate CQs directly from a markdown-formatted version of the user story. Prompts
were intentionally minimal and neutral: no explicit instructions were given regarding CQ format,
number, or examples, to avoid priming or biasing the output. This open-ended configuration
was intended to test each model’s intrinsic ability to extract ontology-relevant requirements and
phrase them as competency questions.
      </p>
      <sec id="sec-2-1">
        <title>2.1. AskCQ Dataset Construction</title>
        <p>
          All three approaches were applied to the same textual requirement source: a detailed user story developed
for a cultural heritage ontology use case. The story, adapted from the methodology proposed by de
Berardinis et. al. [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ] is centred on two personas, a music archivist and a curator, and describes their
activities and data needs relating to a museum’s music memorabilia collection, including acquisition,
loan, metadata management, and display. The output comprises five CQ sets: HA-1 (44 CQs), HA-2 (54
CQs), Pattern (38 CQs), GPT-4.1 (26 CQs), and Gemini 2.5 Pro (42 CQs), totalling 204 distinct questions.1
        </p>
      </sec>
      <sec id="sec-2-2">
        <title>2.2. Evaluation Dimensions and Feature Extraction</title>
        <p>
          To assess the quality and characteristics of the generated CQs, we adopted a multi-dimensional,
mixedmethods evaluation framework encompassing both qualitative expert judgment and quantitative feature
analysis: CQ suitability, structural and semantic properties, and inter-method agreement.
1. Suitability (Expert Evaluation): Each CQ was independently reviewed by three ontology experts,
who rated its acceptability for guiding ontology engineering in the context of the user story. Scores
ranged from -3 (unanimous rejection) to +3 (unanimous acceptance). The experts were not provided
with explicit criteria to preserve their interpretive autonomy, analogous to the elicitation setup. A
Fleiss’ Kappa of  = 0.35 indicated fair inter-expert agreement.
2. Readability: Each CQ was assessed to gauge its ease of understanding. We assess readability in a
similar way to Ciroku, et al. [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ], where a suite of established readability indices designed to capture
diferent aspects of textual dificulty were initially computed for each CQ using the textstat Python
1The resulting AskCQ dataset is publicly released under a CC-BY license, and all CQs were anonymized and randomly shufled
prior to evaluation to minimize bias regarding their origin.
library.2 In this paper we report only the Flesch-Kincaid Grade Level (FKGL) and the Dale-Chall
Readability Score (DCR) as representative readability features.
        </p>
        <p>
          • Flesch-Kincaid Grade Level (FKGL) — Estimates the education level (U.S. grade) required for
comprehension [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ].
• Dale-Chall Readability Index (DCR) — Penalizes complex vocabulary based on a restricted
list of familiar words [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ].
3. Relevance: The alignment of each CQ, together with the user story, was assessed by Gemini 2.5
Pro and rated on a 4-point scale Likert scale using the following criteria: (4) directly stated in the story,
(3) inferable and necessary, (2) tangentially relevant, (1) of-topic. The evaluation prompt was carefully
designed and spot-validated on a selected sample of CQs.
4. Complexity: The following four complementary metrics were defined to quantify diferent facets
of CQ complexity:
• c0 (Length): The total number of characters, as a coarse indicator of verbosity and potential
elaboration.
• c1 (Requirement Complexity): The number of distinct concepts, properties, relations, and
iflters identified in the CQ by Gemini 2.5 Pro.
• c2 (Linguistic Complexity): A count of syntactic and lexical features (noun phrases, verbs,
prepositions, modifiers, etc.) extracted via spaCy.
• c3 (Syntactic Complexity): The structural depth and richness of the dependency parse tree,
including depth, node count, and key dependency types. These metrics were selected from the
linguistic complexity heuristics from the Universal Dependency set [
          <xref ref-type="bibr" rid="ref10">10</xref>
          ].
        </p>
        <p>
          Overall, these four dimensions are expected to provide complementary perspectives on CQ complexity.
A CQ might be semantically complex (e.g., requiring navigation of intricate partonomy or causality
relations) yet linguistically simple (e.g., “What caused this event?”), scoring high on requirement metrics
but low on linguistic/syntactic ones. Conversely, a CQ might lack ontological complexity but is phrased
using complex sentence structures, thereby scoring high on syntactic metrics but low on semantic ones.
5. Semantic Overlap: To analyse the semantic characteristics of CQ sets generated by diferent
approaches, we conducted a study on their embeddings. This utilised Sentence-BERT embeddings from
the all-MiniLM-L6-v2 model [
          <xref ref-type="bibr" rid="ref11">11</xref>
          ], which generates vectors e ∈ ℝ384 capturing the semantic meaning
of each CQ (this method follows that adopted by several other studies [
          <xref ref-type="bibr" rid="ref3 ref5">3, 5</xref>
          ]). Furthermore, to identify
semantically equivalent CQs, a pre-defined similarity threshold (  = 0.75 ) was determined empirically.
This study quantifies the semantic overlap between pairs of CQ sets (e.g.,  ↔  ). For each pair,
we denote   = | | and   = | | as the number of CQs in each set, respectively, and measure:
• Centroid cosine similarity. The cosine similarity between the centroids of Set A and Set B
provides a measure of the overall alignment of their central semantic representation. A score
that is closer to 1 will indicate that the two sets are, on average, focused on similar concepts.
• Coverage analysis. We measured how well one set covers the semantic content of another.
        </p>
        <p>This was performed in both directions, i.e. for the coverage of Set A by Set B (Set A ← Set B) we
determine:
– Mean Maximum Similarity (MMS). For each CQ embedding e, in Set A, its maximum
cosine similarity to any CQ embedding in Set B,    → = max cos(e, , e, ), was identified.
The mean of these    → scores (and the standard deviation) indicates how well each CQ in
Set A is semantically represented by its closest counterpart in Set B. A higher mean suggests
stronger semantic parallels ofered by Set B.
2Scores were computed using the textstat Python library and interpreted comparatively, given the short, interrogative
nature of CQs.</p>
        <p>The same metrics were computed for the coverage of Set B by Set A.
• Bidirectional coverage: This symmetric metric quantifies the overall mutual semantic overlap.</p>
        <p>cov +  →co v , where  →
It was calculated as  → cov is the number of CQs in Set A covered by Set B (i.e.,
   → ≥  ), and   →co v is t+h e number of CQs in Set B covered by Set A. Hence, a higher percentage
indicates greater shared conceptual space between the two sets.</p>
        <p>Together, these dimensions provide a comprehensive, multidimensional view of the suitability,
expressiveness, and diversity of CQs produced by diferent elicitation methods, grounded in both expert
assessment and computational analysis.</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Overview of Evaluation Outcomes</title>
      <p>The results for the expert evaluation clearly favoured manually authored CQs. The manual method
(HA-1 and HA-2) achieved a mean suitability score of 2.65, with 94.5% of questions accepted by a
majority of annotators. This indicates that domain experts are highly efective at producing suitable
CQs. The LLM-based methods (GPT-4.1 and Gemini) achieved an average score of 1.24 with 76.0%
acceptance, suggesting moderate reliability. Pattern-based CQs scored lowest, with a mean suitability
of 0.11 and only 50% acceptance. For readability, human-authored CQs had the lowest FKGL and DCR
scores, indicating clearer phrasing. GPT-4.1 generated the most complex and least readable CQs (FKGL
11.64). LLM-generated CQs were also significantly longer (c0), richer in ontological references (c1), and
more syntactically complex (c3) than manual or pattern-based ones. Figure 1 consolidates our findings,
showing distinct feature profiles from the CQs generated by each elicitation approach. 3</p>
      <p>The pairwise comparison results (Table 1) show that the cosine similarities between the centroids of
most pairs are relatively high (typically ranging from 0.61 to 0.85). This suggests that, at a high level,
all sets tend to address the same core thematic area defined by the user story. The lowest centroid
similarities were observed in comparisons involving Gemini (e.g., 0.61 with HA-2), indicating its central
theme might be slightly more distinct than the other sets.</p>
      <p>CQ Feature Profile</p>
      <p>Linguistic (c2)
Syntactic (c3)</p>
      <p>Requirement (c1)
0.25
0.50</p>
      <p>Despite these relatively high centroid similarities, the specific semantic coverage between sets is low,
denoting high degrees of novelty, i.e. a high number of CQs not previously generated. The percentage
of CQs in one set covered by another (i.e., having a CQ in the other set with similarity ≥ 0.75) is
consistently below 21%, and often below 10%. Between the two human annotators (  − 1 ↔   − 2 ),
who shared a high centroid similarity (0.82), HA-2 covered 20.5% of HA-1’s CQs, and HA-1 covered 11.1%
of HA-2’s CQs (HA-2 has 10 more CQs than HA-1), yielding a bidirectional coverage proportion of 15.3%.</p>
    </sec>
    <sec id="sec-4">
      <title>4. Discussion and Conclusion</title>
      <p>Our findings suggest that CQs manually crafted by ontology engineers tend to demonstrate the
highest suitability for OE, due to achieving better readability, lower complexity, and uniquely capturing
inferential requirements (implicit functional requirements) essential for robust ontology design. While
LLMs can produce relevant and thematically coherent outputs, the resulting CQs exhibited higher
complexity, lower readability, and their semantic coverage, though broad, exhibits limited overlap with
human-generated CQs and amongst each other. These results suggest that while LLMs can provide
reasonable CQs, these are not comparable to expert authored ones, and that human expertise still
remains critical in Ontology Engineering. Crucially, our insights on CQ characteristics and limitations
of current automated approaches can be leveraged to directly inform and improve their elicitation
methods, aiming to better align their outputs with the desiderata of ontology engineers.</p>
    </sec>
    <sec id="sec-5">
      <title>Declaration on Generative AI</title>
      <p>
        Generative AI was only used in the experiments described in the paper, and no Gen AI tool was used to
compose or edit the text.
3A full discussion of the results and analysis for this evaluation is available in [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ].
      </p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>G. K. Q.</given-names>
            <surname>Monfardini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. S.</given-names>
            <surname>Salamon</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. P.</given-names>
            <surname>Barcellos</surname>
          </string-name>
          ,
          <article-title>Use of competency questions in ontology engineering: A survey</article-title>
          ,
          <source>in: Proc. of the Conceptual Modeling: 42nd International Conference</source>
          , ER, Springer-Verlag,
          <year>2023</year>
          , p.
          <fpage>45</fpage>
          -
          <lpage>64</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>C. M.</given-names>
            <surname>Keet</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z. C.</given-names>
            <surname>Khan</surname>
          </string-name>
          ,
          <article-title>Discerning and characterising types of competency questions for ontologies</article-title>
          ,
          <year>2024</year>
          . URL: https://arxiv.org/abs/2412.13688. arXiv:
          <volume>2412</volume>
          .
          <fpage>13688</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>R.</given-names>
            <surname>Alharbi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Tamma</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Grasso</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T. R.</given-names>
            <surname>Payne</surname>
          </string-name>
          ,
          <article-title>A review and comparison of competency question engineering approaches</article-title>
          ,
          <source>in: Proc. 24th International Conference on Knowledge Engineering and Knowledge Management</source>
          ,
          <string-name>
            <surname>EKAW</surname>
          </string-name>
          , Springer Nature,
          <year>2024</year>
          , pp.
          <fpage>271</fpage>
          -
          <lpage>290</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Ren</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Parvizi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Mellish</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. Z.</given-names>
            <surname>Pan</surname>
          </string-name>
          ,
          <string-name>
            <surname>K. van Deemter</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Stevens</surname>
          </string-name>
          ,
          <article-title>Towards competency questiondriven ontology authoring</article-title>
          ,
          <source>in: Proc. of the 11th Extended Semantic Web Conference</source>
          , ESWC, Springer International Publishing,
          <year>2014</year>
          , pp.
          <fpage>752</fpage>
          -
          <lpage>767</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>F.</given-names>
            <surname>Ciroku</surname>
          </string-name>
          , J. de Berardinis,
          <string-name>
            <given-names>J.</given-names>
            <surname>Kim</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Meroño-Peñuela</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Presutti</surname>
          </string-name>
          , E. Simperl, Revont:
          <article-title>Reverse engineering of competency questions from knowledge graphs via language models</article-title>
          ,
          <source>Journal of Web Semantics</source>
          <volume>82</volume>
          (
          <year>2024</year>
          )
          <fpage>100822</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>B.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V. A.</given-names>
            <surname>Carriero</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Schreiberhuber</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Tsaneva</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L. S.</given-names>
            <surname>González</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Kim</surname>
          </string-name>
          , J. de Berardinis,
          <article-title>Ontochat: A framework for conversational ontology engineering using language models</article-title>
          ,
          <source>in: Proc. of the 21st Extended Semantic Web conference, ESWC</source>
          , Springer Nature Switzerland,
          <year>2025</year>
          , pp.
          <fpage>102</fpage>
          -
          <lpage>121</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <surname>J. de Berardinis</surname>
            ,
            <given-names>V. A.</given-names>
          </string-name>
          <string-name>
            <surname>Carriero</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          <string-name>
            <surname>Jain</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          <string-name>
            <surname>Lazzari</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Meroño-Peñuela</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Poltronieri</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          <string-name>
            <surname>Presutti</surname>
          </string-name>
          ,
          <article-title>The polifonia ontology network: Building a semantic backbone for musical heritage</article-title>
          ,
          <source>in: Proc. of the 22nd International Semantic Web Conference</source>
          , ISWC, Springer,
          <year>2023</year>
          , pp.
          <fpage>302</fpage>
          -
          <lpage>322</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>J. P.</given-names>
            <surname>Kincaid</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. P.</given-names>
            <surname>Fishburne</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. L.</given-names>
            <surname>Rogers</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B. S.</given-names>
            <surname>Chissom</surname>
          </string-name>
          ,
          <source>Derivation of New Readability Formulas (Automated Readability Index, Fog Count and Flesch Reading Ease Formula) for Navy Enlisted Personnel</source>
          ,
          <source>Technical Report Research Branch Report 8-75</source>
          , Naval Air Station Memphis, Research Branch,
          <string-name>
            <surname>Millington</surname>
            <given-names>TN</given-names>
          </string-name>
          ,
          <year>1975</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>E.</given-names>
            <surname>Dale</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. S.</given-names>
            <surname>Chall</surname>
          </string-name>
          , A Formula for Predicting Readability: Instructions, Ohio State University Bureau of Educational Research,
          <year>1948</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <surname>M.-C. de Marnefe</surname>
            , T. Dozat,
            <given-names>N.</given-names>
          </string-name>
          <string-name>
            <surname>Silveira</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          <string-name>
            <surname>Haverinen</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Ginter</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Nivre</surname>
            ,
            <given-names>C. D.</given-names>
          </string-name>
          <string-name>
            <surname>Manning</surname>
          </string-name>
          ,
          <article-title>Universal Stanford dependencies: A cross-linguistic typology</article-title>
          ,
          <source>in: Proc. of the Ninth International Conference on Language Resources and Evaluation (LREC'14)</source>
          ,
          <source>European Language Resources Association (ELRA)</source>
          ,
          <year>2014</year>
          , pp.
          <fpage>4585</fpage>
          -
          <lpage>4592</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>N.</given-names>
            <surname>Reimers</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Gurevych</surname>
          </string-name>
          ,
          <article-title>Sentence-bert: Sentence embeddings using siamese bert-networks</article-title>
          ,
          <source>in: Proc. of the 2019 Conference on Empirical Methods in Natural Language Processing, Association for Computational Linguistics</source>
          ,
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>R.</given-names>
            <surname>Alharbi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Tamma</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T. R.</given-names>
            <surname>Payne</surname>
          </string-name>
          , J. de Berardinis,
          <article-title>A comparative study of competency question elicitation methods from ontology requirements</article-title>
          ,
          <year>2025</year>
          . URL: https://arxiv.org/abs/2507.02989. arXiv:
          <volume>2507</volume>
          .
          <fpage>02989</fpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>