=Paper= {{Paper |id=Vol-2954/paper-30 |storemode=property |title=Metadata-based Term Selection for Modularization and Uniform Interpolation of OWL Ontologies |pdfUrl=https://ceur-ws.org/Vol-2954/paper-30.pdf |volume=Vol-2954 |authors=Xinhao Zhu,Xuan Wu,Ruiqing Zhao,Yu Dong,Yizheng Zhao |dblpUrl=https://dblp.org/rec/conf/dlog/ZhuWZDZ21 }} ==Metadata-based Term Selection for Modularization and Uniform Interpolation of OWL Ontologies== https://ceur-ws.org/Vol-2954/paper-30.pdf
        Metadata-based Term Selection for
     Modularization and Uniform Interpolation of
                  OWL Ontologies

      Xinhao Zhu2 , Xuan Wu1,2 , Ruiqing Zhao2 , Yu Dong2 , Yizheng Zhao1,2
1
    National Key Laboratory for Novel Software Technology, Nanjing University, China
              2
                School of Artificial Intelligence, Nanjing University, China



        Abstract. This paper explores the problem of selecting good terms as
        seed signature for abstraction of OWL ontologies. Existing methods gen-
        erate seed signatures based on geographic connections, which is far from
        sufficient to produce a satisfactory abstract. This restricts the reusability
        of OWL ontologies from the aspect of knowledge management. In this
        paper, we propose a signature extension approach to generate seed signa-
        tures for modularization and uniform interpolation of OWL ontologies,
        both of which are ontology abstraction techniques. The approach estab-
        lishes the semantic relevance of terms by taking into account as much
        as possible metadata information of an OWL ontology, and computes a
        numerical value to measure the relevance of terms using their embedding
        transformed based on a so-called OWL2Vec* framework. An empirical
        evaluation of the approach shows that the proposed method significantly
        outperforms other term selection baselines in making accurate selections.
        Besides, a case study on ontology abstraction tasks shows that modular-
        ization tools can make more complete and precise abstractions using the
        signature extended by our method.



1     Introduction
Because of the heterogeneous nature of web resources, ontologies developed for
the semantic web are typically large, sometimes monolithic, and knowledge mod-
elled therein is rich and covers multiple topics. This may however restrict the
reusability and interoperability of ontologies in real-world application scenarios,
since large ontologies can be difficult to manage, unwieldy to manipulate, and
moreover costly to reason about.
    Consider an ontology reuse use case where an ontologist wants to import a
football ontology into a growing sports knowledge base. Currently the only well-
established ontology concerning football is the BBC Sports Ontology3 , which,
however, publishes data about all types of competitive physical activities, per-
taining not only to the topic of football. Importing the whole ontology into the
  Copyright © 2021 for this paper by its authors. Use permitted under Creative
  Commons License Attribution 4.0 International (CC BY 4.0).
3
  https://www.bbc.co.uk/ontologies/sport
knowledge base is not difficult from an engineering perspective, but as one can
expect, many web services upon the knowledge base such as search, querying,
retrieval, which typically involve extensive reasoning, may become problematic,
as too much irrelevant information has been, automatically yet unnecessarily,
introduced. Such information makes no contribution to the formalization of the
information about football but increases the computational cost.
    A straightforward way to tackle these challenges of reusability and interop-
erability is to extract a fragment of an ontology that can behave in the same
way as the original ontology in a specific context, but is significantly smaller. In
the above case, this means to extract from the BBC Sports Ontology a fragment
that contains sufficiently many logical statements to summarize all knowledge
about football. Ideally, this fragment should be as small as possible.
    Two logic-based approaches have been developed for computing fragments of
ontologies. One is based on modularization [5,9,13,8,3,15], which seeks to identity
from an ontology a subset (module) that preserves several reasoning tasks for a
sub-vocabulary of the ontology, namely a seed signature.4 The other is uniform
interpolation [18,16,6,17], which computes a more compact representation of a
module of an ontology which preserves the underlying logical definitions of the
terms in the seed signature.
    As one could expect, the quality of extracted fragments depends largely on
the seed signature fed to modularization and uniform interpolation procedures.
We may say that a fragment is complete if it covers all essential information
about the topic of interest, and a fragment is precise if it is complete and in
addition, it does not include too much irrelevant information about the topic
of interest. More specifically, if we selected as seed signature too few terms to
summarize all materials of the topic, we would lose important information that
a user may be interested in, and if we selected as seed signature too many terms
with some of them not strongly relevant to the topic, we would include too
much additional information. Importing more information can also change the
definitions of the terms in the original ontology, and destroy the coherence and
consistency of the original ontology [9].
    Nevertheless, very little attention has been paid to the problem of term selec-
tion for ontology extraction. Chen et al. [3] have proposed a signature extension
algorithm to generate seed signatures for ontology modularization. The idea is
to (1) fix a primitive seed signature Σ, often containing several domain expert-
suggested terms, and (2) extend Σ with new terms collected from the axioms
which contain the current Σ-terms. This step is iterated until no new terms can
be added to Σ. One may understand this as: if two people p1 and p2 live together
in a house h1 on an island, then they are relevant and team up as Σ = {p1 , p2 },
and if there exists a road connecting h1 with another house h2 , then the people
living in h2 are collected into Σ. Iteratively, the same strategy applies to the
entire island, and in the end, Σ will probably have collected all habitants on
the island. However, a person who lives on another island will never be collected

4
    A signature of an ontology is the set of all concept and role names in the ontology.
by Σ since there is no road connecting two islands; islands are geographically
isolated.
    Evidently, following this signature extension strategy one must obtain a larger
seed signature with which, a more informative fragment will be produced, but
we may argue that the seed signature obtained in this way, i.e., using a signature
extension algorithm based merely on geographic connections, could hardly yield
a complete fragment. Our argument is that: the relevance between a term and
the expanding seed signature should be evaluated based on a consideration of all
metadata of the participating terms in the context of the host ontology, rather
than based merely on their geographic connections.




                   Fig. 1: A snippet of a multi-domain ontology



     Consider a scenario where an ontologist wants to extract from a multi-domain
ontology a fragment that describes football and closely related information; see
Figure 1. With the central term “Football” being selected as a single seed in the
primitive signature, an extension Σ = {Football, BallGame, Sports, Player,
FootballPlayer} is obtained using the above signature extension algorithm.
Terms in other domains such as MentholSpray will not be collected in Σ, because
it is geographically isolated from the domain of Sports. However, the annotated
information of MentholSpray explains that “MentholSpray can be used as pain
reliever for sports players”. In this sense the term MentholSpray is supposed
to be strongly relevant to the topic. Collecting MentholSpray in the extended
signature may enable the expanded knowledge base to answer queries regarding
the treatment of an injury in a football match. This is a good example showing
that the relevance between a term and the expanding seed signature in the
context of the host ontology could be established based on important metadata
of the participating terms, for example, based on their lexical information.
     In this paper, we propose a novel term selection approach to discovering
semantic relationships between two isolated groups of terms. The idea is to mea-
sure the relevance of non-Σ terms with Σ terms based on their D-dimensional
vector representation computed from important metadata of the ontology using
OWL2Vec* [2], a random walk- and word embedding-based OWL ontology em-
bedding framework that encodes the semantics of OWL ontologies in a vector
space by taking into account their graph structure, lexical information, as well as
the logical constructors used therein. The work is intended to enhance existing
logic-based ontology abstraction techniques as practical tools for many ontology-
based knowledge processing tasks by exploiting non-logical approaches to facil-
itate this transfer. Previously, not much work has considered tightly coupled
logical and data-driven techniques and exploited the complementary strengths
of them to open up an application pipeline. Our empirical evaluation showed that
the proposed approach significantly outperformed other term selection baselines
in recommending good seed signatures, and with this approach, more precise
fragments could be produced using two existing modularization and uniform
interpolation tools.


2     Metadata-based Term Selection

For space reasons, we have to assume readers’ familiarity with the notions of
ontology modularization [9] and uniform interpolation [17]. Our term selection
approach accommodate ontologies described in OWL 2, which are based on the
description logic SROIQ [11]; see the Description Logic Handbook [1] for a
detailed description of the syntax and semantics of description logics.
    Arguably, most topics can satisfactorily be summarized or defined by a set
of concept names, but do not depend too much on role names. Hence, in this
paper, we only consider the seed signature to be a set of concept names.
    The signature sig(O) of an ontology O is the set of all concept names in O.
Given an ontology O and a seed signature Σ ⊆ sig(O) containing a single or
a few concept names suggested by domain experts or simply selected by users,
which are believed to be the central term or terms that can best summarize
the topic of interest, our approach computes an extension Σ 0 of Σ in three
steps, namely concept representation learning, computing relevance value, and
signature extension based on relevance value. Σ 0 is the seed signature to be fed
to modularization and uniform interpolation procedures.


2.1   Concept Representation Learning

The first step is to transform all concept names A in O into D-dimensional
vectors in a vector space where the relevance of each concept name (to Σ) is
computed based on important metadata of O.
    Our concept representation learning model is based on OWL2Vec* [2], an
ontology embedding framework, which computes the vector representations for
concept names in OWL ontologies as expressive as SROIQ. OWL2Vec* com-
putes the embedding of an OWL ontology based on a corpus of sequences of
tokens, which are encoded from the metadata of the ontology. Such metadata
includes the graph structure of the ontology, i.e., an RDF graph (a set of RDF
Algorithm 1 Nearest Neighbour Ranking
Input: A set of concepts NC , A set of seed signatures Σ s.t. Σ ⊆ NC ,
       A set of concept embedding {eA : A ∈ NC },
       A distance function d : RD × RD → [0, ∞].
Output: A relevance function f : NC → [0, 1],

 1: Let g be a mapping of NC → [0, ∞].
 2: for all A ∈ NC do
 3:   g(A) := ∞
 4:   for all A0 ∈ Σ do
 5:      g(A) := min (d (eA , eA0 ) , g(A)).
 6:   end for
 7: end for
 8: Let f be a mapping of NC → [0, 1].
 9: for all A ∈ NC do
10:   Find i, s.t. A has the i-th smallest g(A) in NC .
11:   f (A) := 1 − (i − 1)/|NC |.
12: end for
13: return f



triples) converted from the OWL ontology by OWL2Vec*, the so-called lexical
information about the ontology, i.e., annotations, and the so-called logical infor-
mation about the concepts and roles in the ontology, i.e, subsumption, equiva-
lence, disjointness, etc.
    We note that OWL2Vec* was not meant for term selection tasks, so we make
modificationstothe original OWL2Vec* model to maximize the performance of
the downstream term selection models. In particular,we designed a fine-tuning
process to further improve ontology embedding, which was task-specific and
further discussed in section 3. In the end, every concept name A is represented
as a D-dimensional vector eA .


2.2   Computing Relevance of Concept Names w.r.t Σ

The second step is to compute the relevance value of every (non-Σ) concept
name A in O w.r.t. Σ. The computation is based on the relative distance of e A
to its nearest seed neighbour (the nearest seed name) in the vector space. The
range of the relevance value is [0, 1] with 1 standing for the strongest relevance
and 0 for the weakest relevance. The relevance value is computed by a newly
developed algorithm called Nearest Neighbor Ranking algorithm (NN-RANK),
shown in Algorithm 1.
    NN-RANK first computes the distance from each concept name to each seed
name in the vector space. In principle, many distance functions d : RD × RD →
[0, ∞] can be used to achieve this, but the Consine distance, formulated as
                                                 eA · eB
                           d(eA , eB ) = 1 −
                                               keA k2 keB k2
has made the best measure of relevance in our experiments. |Σ| distance values
are computed in this way for each concept name A, while the smallest distance
value, which denotes the shortest distance, is identified as a valid distance value
of A to Σ. NN-RANK then sorts all concepts names in O by their valid distance
value. Concept names with smaller valid distance values are considered to be
semantically more relevant to the seed signature, and thus to the central topic.
These valid distance values (and the corresponding concept names) are then
uniformly distributed between 0 and 1. The result is the relevance value of each
A w.r.t. Σ.

2.3    Relevance-based Seed Signature Extension
A natural question arises: how to use the computed relevance values to guide the
selection of terms for ontology abstraction? Upon different application demands,
the strategies may vary. Without a well-acknowledged gold standard, a feasible
solution could be to measure the “degree” of relevance and define to what degree
the relevance is a concept name can be thought of as “relevant” to the seeds in
Σ. For example, one could set a numerical threshold on the relevance value at
0.9 if she wants to gain a more cohesive abstraction of ontology and at 0.5 if she
wants to have a looser one. We leave this flexibility to users. Given a threshold
σ at the scale of 0 to 1, our approach extends the primitive seed signature Σ by
adding to Σ the concept names with relevance value no less than σ. The result
is Σ 0 = Σ ∪ {A | A ∈ sig(O) ∧ f (A, Σ) ≥ σ}.
    Computing | sig(O)| × |Σ| distances requires linear time to | sig(O)|, and the
subsequent sorting requires linear time to | sig(O)|. Hence, we have the following
lemma regarding the time complexity of NN-RANK.

Lemma 1. Given any OWL ontology O in SROIQ and a primitive seed signa-
ture Σ ⊆ sig(O) with n = | sig(O)| and k = |Σ|, our term selection approach al-
ways computes an extended seed signature Σ 0 such that Σ ⊆ Σ 0 in O(n log n+kn)
time.


3     Empirical Evaluation of NN-RANK
In this experiment, we used NN-RANK to predict SNOMED CT Refset compo-
nents. The aim was to show that the algorithm could enrich a given primitive
seed signature Σ with concept names highly relevant to the initial seeds (in a
vector space). The experiment was conducted on a work station with an Intel
Xeon CPU @ 2.60GHz and 32 GB memory.
    SNOMED CT5 is currently the most comprehensive, multilingual clinical
healthcare ontology in the world. A SNOMED CT Refset6 is a collection of
SNOMED CT components sharing specific characteristics (e.g., a specific do-
main). An example of SNOMED CT Refset is the Malaria refset released by the
5
    https://www.snomed.org/
6
    https://confluence.ihtsdotools.org/display/DOCGLOSS/refset
National Resource Centre for EHR Standards in India, which includes findings,
disorders, and organisms related to Malaria. Arguably, the refset published offi-
cially by a group of ontology engineers and domain experts, can be considered
as a complete and precise standard of an Malaria abstract of SNOMED CT.
    Our task was to predict concepts in SNOMED CT Refsets based on a seed
signature (randomly or manually) selected from the refsets. This task was de-
signed to fit with realistic scenarios where we needed to develop a new refset
with least intervention from domain experts. We assumed that refsets developed
by the domain experts were complete and precise fragments, containing concepts
that were highly interconnected on the semantic level (e.g., in the same clini-
cal domain). Therefore, the task of predicting SNOMED CT Refset components
could be used to evaluate the performance of term selection models.
    To better position our algorithm, we compared NN-RANK with two other
term selection strategies, namely, a strategy adapted from locality-based mod-
ularization [10] (denoted as Star-modularization), and the signature-extension
based on geographic connections [3] (denoted as Sig-Ext, configured with depth
d). We treated them as baselines. The idea of the locality-based modularity
strategy was to take all concept names in the computed module as the extended
signature of the seed. This may not be ideal but was nevertheless a means to
extend the seed signature. In this way, the relevance value f (A, Σ) of A was 1
if A was in the signature of the computed module, and 0 otherwise. We also
considered a comparison of NN-RANK with Meta-SVDD [7], a model designed
for few-shot one-class-classification problems. Using Meta-SVDD, we learnt pat-
terns about refsets from existing refsets, in order to enhance its performance in
predicting new refset components.
    We considered the International Edition of SNOMED CT (version July 2020),
which contains 354,256 concepts, 355,214 logical axioms, and 1,506,185 descrip-
tion axioms. We used two sets of publicly accessible and in-use term collections,
NHS refsets 7 and NRC refsets 8 , as the target refsets.
    The NHS refsets, issued by the National Health Service (NHS) in the UK,
offered from the full Edition of SNOMED CT a set of components defined by a
particular requirement. The NRC refsets were released by the National Resource
Centre for EHR Standards (NRCeS) in India, which contained 30 standalone
refsets covering concepts related to common diseases.
    We adopted two metrics widely used in classification and ranking tasks,
namely the Normalized Discounted Cumulative Gain (NDCG) and the Area
under the ROC Curve (AUC), to evaluate the performance of term selection
models. Both measures returned high values if a model made accurate predic-
tions, i.e. they measured the similarity between the approximations and the
refset components.
    Ontology embedding generated by OWL2Vec* on SNOMED CT was used for
the concept embedding, where each concept was represented by a 200-dimensional
vector. Different from the original OWL2Vec* model, we used a fine-tuning pro-
7
    https://dd4c.digital.nhs.uk/dd4c/
8
    https://www.nrces.in/resources#snomedct releases
cess specially designed for this task, to further improve the ontology embedding.
Specifically, refsets in this process were transformed to documents containing
(concept uri, refset identifier, concept uri) triples, then a Word2Vec model was
used to fine-tune the pre-computed concept embedding on these documents. The
fine-tuning process was done in a 10-fold cross validation manner, which meant
that evaluations on any refset is based on a concept embedding fine-tuned on
90% refsets other than itself.
    For NRC refsets, two seed signatures Σr and Σs consisted of K concepts
respectively were used throughout the experiment. Σr was randomly selected
among all the refset concepts, while Σs was manually selected with the aim that
the K concepts it contained could describe the topic from different aspects. For
NHS refsets, we only used a different set of Σr generated in the same way. It was
crucial to be able to set the size of the primitive seed signature K accordingly
to the application. In realistic use cases, the seed signature may be manually
selected, where smaller K means less manual cost, so K = 5 is used in the
experiments.
    We used the OWL API syntactic locality module extraction tool9 as the
implementation of the locality-based module, and the official implementation
of Sig-Ext. For Meta-SVDD, our implementation was based on the source code
provided by [4].


         Table 1: Results on NHS, NRC refsets using Σr (the higher the better).

       Methods                               NHS refsets                                          NRC refsets
                                  NDCG                           AUC                     NDCG                       AUC
                            K=1           K=5              K=1         K=5         K=1          K=5           K=1         K=5
  Star-modularization  40.93 ± 14.61 47.33 ± 13.36 50.84 ± 1.10 54.73 ± 5.56 49.10 ± 16.23 51.83 ± 14.62 50.64 ± 0.98 54.58 ± 7.82
     Sig-Ext (d=1)           -       49.14 ± 10.92       -        54.31 ± 5.58      -       55.68 ± 11.06       -      53.60 ± 6.73
     Sig-Ext (d=2)           -       47.99 ± 11.66       -        54.31 ± 5.58      -       54.31 ± 11.81       -      53.78 ± 6.91
      Meta-SVDD              -       67.72 ± 23.26       -       91.55 ± 10.43      -       71.65 ± 16.62       -      88.81 ± 8.87
      NN-RANK          68.57 ± 20.36 77.93 ± 14.91 92.19 ± 11.19 96.49 ± 5.11 71.32 ± 14.33 77.25 ± 10.03 89.66 ± 8.42 94.29 ± 5.51
NN-RANK + fine-tuning 69.50 ± 20.13 78.76 ± 14.62 93.33 ± 9.57 96.98 ± 4.62 73.57 ± 12.54 80.19 ± 9.14 90.40 ± 8.43 94.79 ± 5.69




               Table 2: Results on NRC refset using Σs (the higher the better).

         Methods                                    NDCG                                               AUC
                                   K=1               K=3                K=5              K=1            K=3             K=5
  Star-modularization  48.85 ± 16.68 50.65 ± 15.26 52.25 ± 14.42 50.82 ± 2.01 53.21 ± 5.53 54.68 ± 7.50
     Sig-Ext (d=1)     49.97 ± 15.36 53.42 ± 12.16 55.76 ± 10.48 50.80 ± 1.53 52.14 ± 3.89 53.34 ± 5.81
     Sig-Ext (d=2)     49.56 ± 15.82 52.33 ± 13.06 54.38 ± 11.36 50.87 ± 1.60 52.28 ± 4.01 53.48 ± 5.93
      Meta-SVDD        71.28 ± 12.25 74.91 ± 16.48 75.24 ± 13.73 72.4 ± 16.23 86.83 ± 10.05 92.01 ± 6.45
      NN-RANK          79.77 ± 11.79 83.67 ± 10.74 84.83 ± 9.95 94.07 ± 5.11 96.09 ± 3.73 96.64 ± 3.09
NN-RANK + fine-tuning 80.39 ± 12.02 84.41 ± 10.95 85.53 ± 10.20 94.65 ± 5.01 96.49 ± 3.73 96.97 ± 3.06




9
    https://github.com/owlcs/owlapi
3.1   Results and Analysis

The results (mean value ± standard deviation of the two measures) in Table 1
and 2 show that embedding-based methods outperformed logical approaches in
the above settings. This was because logical methods were not designed for this
task, and it did not capture lexical information of the ontology, which was crucial
in determining the semantic relevance between concepts.
    Besides, NN-RANK slightly outperformed Meta-SVDD, particularly when
using Σs . We will conduct a case study on the aforementioned Malaria refset to
explain the mechanism and effectiveness of NN-RANK in this task.
    Figure 2 shows the distribution of the Malaria refset components and other
SNOMED CT concepts in a 2-dimensional vector space. As illustrated in the
figure, refset components tended to form a number of minor clusters, with each
containing some highly semantically relevant concepts. The whole refset was
composed of several concept clusters instead of a giant cluster. This meant that
when two seed concepts A1 and A2 were given, any concept A that was similar
to A1 or A2 , i.e. d(eA , eA1 ) <  or d(eA , eA2 ) <  with  being a small value
greater than 0, were more likely to be a refset component compared to another A
which was similar to the average of eA1 and eA2 , i.e., d (eA , (eA1 + eA2 )/2) < .
NN-RANK was designed to fit in this multi-clusters pattern, and achieved better
performance compared to other models utilizing concept embedding.
    The performance of NN-RANK could be significantly enhanced when seed
signatures described the topic from different aspects. For a high quality primitive
seed signature like Σs , an increased seed signature size would generally led to
more accurate selection results.


3.2   Time Efficiency

For the current setting of N = 354, 256, K = 5, D = 200 and using Cosine
distance as the distance function, NN-RANK generated Σ 0 within 5 seconds. For
comparison, it usually takes minutes to hours for other approaches (e.g., Star-
modularization and Sig-Ext) to compute on a large-scale ontology like SNOMED
CT, and five minutes for the Meta-SVDD model to converge in the same setting.
    It is true that our approach takes around 2 hours to build embedding vectors
on SNOMED CT, but this cost is acceptable in real-life scenarios since the
training is conducted only once but can be meaningfully used many times and
forever. Also, the training time can be adjusted. When the ontology contains
less than 100K logical and annotation axioms, it is typically less than one hour.


4     Case Study: Ontology Abstraction

In this part, we explored how input signature extended by NN-RANK bene-
fits differently between modularization and uniform interpolation in the OWL
ontology abstraction task.
Fig. 2: Distribution of malaria refset components and other SNOMED CT concepts (
170 concepts from the malaria refset and 1700 random concepts outside of the refset).
Each point corresponds to a SNOMED CT concept, whose colour shows its relevance
with the seed signatures computed by NN-RANK (the higher, the deeper), and shape
denotes its type (cross for being refset components, and circles for not). Seed concepts
are depicted as blue stars accompanied with tags. The mappings between tag and
label are: A - Malaria (disorder), B - Allergy to primaquine (finding), C - Accidental
pyrimethamine poisoning (disorder), D - Malaria outbreak education (procedure), E -
Antimalarial drug adverse reaction (disorder)




   We considered HeLiS10 , an ALCHIQ(D) ontology integrating knowledge
about food and activity from a nutritional point of view. The experiment was
based on HeLiS v1.10 which has 172,213 axioms, 277 concepts, and 50 roles.

4.1      Setup Details
First, we randomly generated 10 concept subsets from sig(OHeLiS ) with the size
of subsets ranged from 1 to 5. These randomly generated concept sets, denoted
as Σr , could be the approximations of seed signatures around random topics.
Then NN-RANK returned the ordered sets Σ 0 .
    As the abstractions in real-life are usually small in size, we chose the top 10%
of Σ 0 (i.e., set the threshold as 0.9) to be the input signature for modularization
and uniform interpolation. We used UI-FAME [19] to compute uniform inter-
polants, and Star-modularization to compute locality-based modules as they are
publicly accessible. Both preserved full logical entailments of the input signa-
ture Σ 0 in OHeLiS [10,14]. Then the abstraction results computed by these two
10
     https://horus-ai.fbk.eu/helis/
tools with the input of Σ 0 (denoted as Σ 0 +UI-FAME, Σ 0 +Star-modularization)
were assessed with four metrics [12]: module size |M|, module inherent rich-
ness InhRich, module intra distance IntraDist and module cohesion Cohesion.
A module with relative smaller size, higher inherent richness, relative smaller
intra distance, and higher cohesion was said to be more compact. We also test
Σr +Star-modularization and compared it with Σ 0 +Star-modularization.


4.2   Results and Analysis

We compared Σ 0 +UI-FAME and Σ 0 +Star-modularization to see the effective-
ness of NN-RANK to different abstraction methods. From table 3, we can see
that UI-FAME generated more compact abstractions. Besides, UI-FAME was
sensitive to the input signature. These results make sense because locality-based
modularization introduced other terms which were not in Σ 0 but uniform inter-
polation stuck to Σ 0 . Experiments with thresholds setting as 0.3, 0.5, and 0.7
show that the size of Σ 0 did not affect the compactness of the locality-based
module abstraction.



Table 3: Module compactness Evaluation (Use top 10% Σ 0 as input. |M| is the sum
of quantities of concepts, roles and individuals in M . InhRich: the average number of
subclasses per class. IntraDist: the overall distance between the entities in the module.
Cohesion: the extent to which entities are related to each other in the module.)
         Metrics                      K=1                                  K=5
                     Star-modularization       UI-FAME     Star-modularization     UI-FAME
            |M|             171 ± 14            20 ± 7          174 ± 15           18 ± 8
          InhRich         2.92 ± 0.12          2.1 ± 1.25     4.08 ± 0.17        3.75 ± 0.49
         IntraDist      49683.90 ± 94.61    618.75 ± 617.87 49798.70 ± 278.77 289.50 ± 344.26
         Cohesion          0.08 ± 0.01        0.19 ± 0.09      0.08 ± 0.00      0.15 ± 0.10




                Table 4: Term selection for SpecialBread topic in HeLiS

                   Σr      {SpecialBread}
              Of ragment SpecialBread v Bread
                         {SoyBread, OliveBread, MilkBread, OilBread, RyeBread}v SpecialBread
               Σ 0 @10     SpecialBread, Bread, WhiteBread, PizzaAndFocacciaBread,
                           OlivesAndOliveProducts, SoyProducts, LegumesAndLegumeProducts,
                           WheatFlour, WholeWheatFlour, MilkAndDairyProducts




   Term selection allowed users to extend the seed signature in an adjustable
way. For uniform interpolation, it is a key step to select suitable terms for the
specified topic, because the semantics of the topic is mainly captured by the in-
put terms. We observe that once if the input terms were not sufficient enough for
uniform interpolation, the module could be very small, containing many mean-
ingless axioms like A v > or concept assertion axioms. NN-RANK+UI-FAME
generated knowledge highly relative to the topic. For instance, in Table 4, the
topic was “SpecialBread”. The related axioms in OHeLiS were contained in
Of ragment . Clearly, “SpecialBread” had five individuals. Besides, these indi-
viduals had no other super-classes except “SpecialBread’. As commonsense
knowledge, “OliveBread” can be
“OlivesAndOliveProducts”, “SoyBread” can be “SoyProducts”, “MilkBread”
can be “MilkAndDairyProducts”, which were missing in OHeLiS . So without
the extension of NN-RANK, these related concepts could not be preserved in
Σr + Star-modularization or Σr + UI-FAME. While NN-RANK could pre-
serve them according to that “OlivesAndOliveProducts”, “SoyProducts”, and
“MilkAndDairyProducts” were lexically close to the individuals of the topic con-
cept “SpecialBread”.
    To sum up, with NN-RANK modules and uniform interpolants produced
more complete fragments. In addition, Σ 0 +uniform interpolation produced more
precise fragments than Σ 0 +modularization.



5   Conclusion and Future Work

This paper makes a preliminary attempt to address the problem of extending the
given seed signature with new terms selected sophisticatedly through embedding-
based computation of important metadata of an OWL ontology. An evaluation
of the approach on a predication task of a SNOMED CT refset shows that our
approach makes accurate selections compared with other term selection base-
lines. A case study shows that high-quality modules and uniform interpolants of
OWL ontologies can be produced using our term selection approach.
    The absence of standardized benchmarks remains the main bottleneck in
evaluating the performance of term selection methods. Hence, a number of pre-
defined question answering instances that are generated based on the input ontol-
ogy might be helpful in deciding the completeness and precision of the generated
abstracts of OWL ontologies. For a problem Q that can be answered by querying
an ontology O, a satisfactory abstract M of O regarding a input signature Σ
should be able to answer Q if Q is relevant to Σ, and should not be able to
answer Q if Q is not relevant to Σ.


Acknowledgements

The authors would like to thank the reviewers for their insightful comments
and good suggestions. This work was supported by National Natural Science
Foundation of China (grant 62006114) and Open Research Projects of Zhejiang
Lab (grant 2021KE0AB08).
References
 1. Baader, F., Horrocks, I., Lutz, C., Sattler, U.: An Introduction to Description
    Logic. Cambridge University Press (2017)
 2. Chen, J., Hu, P., Jimenez-Ruiz, E., Holter, O.M., Antonyrajah, D., Horrocks, I.:
    Owl2vec*: Embedding of owl ontologies. arXiv preprint arXiv:2009.14654 (2020)
 3. Chen, J., Alghamdi, G., Schmidt, R.A., Walther, D., Gao, Y.: Ontology Extraction
    for Large Ontologies via Modularity and Forgetting. In: Kejriwal, M., Szekely, P.A.,
    Troncy, R. (eds.) Proc. K-CAP’19. pp. 45–52. ACM (2019)
 4. Dahia, G., Segundo, M.P.: Meta learning for few-shot one-class classification. arXiv
    preprint arXiv:2009.05353 (2020)
 5. d’Aquin, M.: Modularizing ontologies. In: Ontology Engineering in a Networked
    World, pp. 213–233. Springer (2012)
 6. Eiter, T., Ianni, G., Schindlauer, R., Tompits, H., Wang, K.: Forgetting in manag-
    ing rules and ontologies. In: Web Intelligence. pp. 411–419. IEEE Computer Society
    (2006)
 7. Gamper, J., Chan, B., Tsang, Y.W., Snead, D., Rajpoot, N.: Meta-svdd: Proba-
    bilistic meta-learning for one-class classification in cancer histology images. arXiv
    preprint arXiv:2003.03109 (2020)
 8. Gatens, W., Konev, B., Wolter, F.: Lower and upper approximations for depleting
    modules of description logic ontologies. In: Proc. ECAI’14. Frontiers in Artificial
    Intelligence and Applications, vol. 263, pp. 345–350. IOS Press (2014)
 9. Grau, B.C., Horrocks, I., Kazakov, Y., Sattler, U.: Modular Reuse of Ontologies:
    Theory and Practice. J. Artif. Intell. Res. 31, 273–318 (2008)
10. Grau, B.C., Parsia, B., Sirin, E., Kalyanpur, A.: Modularity and web ontologies.
    In: KR. pp. 198–209 (2006)
11. Horrocks, I., Kutz, O., Sattler, U.: The even more irresistible SROIQ. In: Proc.
    KR’06. pp. 57–67. AAAI Press (2006)
12. Khan, Z.C.: Evaluation metrics in ontology modules. In: Description Logics (2016)
13. Konev, B., Lutz, C., Walther, D., Wolter, F.: Model-theoretic inseparability and
    modularity of description logic ontologies. Artif. Intell. 203, 66–103 (2013)
14. Kontchakov, R., Wolter, F., Zakharyaschev, M.: Logic-based ontology compari-
    son and module extraction, with an application to dl-lite. Artificial Intelligence
    174(15), 1093–1141 (2010)
15. Koopmann, P., Chen, J.: Deductive Module Extraction for Expressive Description
    Logics. In: Proc. IJCAI’20. pp. 1636–1643. ijcai.org (2020)
16. Lang, J., Liberatore, P., Marquis, P.: Propositional independence: Formula-variable
    independence and forgetting. J. Artif. Intell. Res. 18, 391–443 (2003)
17. Lutz, C., Wolter, F.: Foundations for Uniform Interpolation and Forgetting in
    Expressive Description Logics. In: Proc. IJCAI’11. pp. 989–995. IJCAI/AAAI Press
    (2011)
18. Visser, A.: Bisimulations, Model Descriptions and Propositional Quantifiers. Logic
    Group Preprint Series, Utrecht University (1996)
19. Zhao, Y., Alghamdi, G., Schmidt, R.A., Feng, H., Stoilos, G., Juric, D., Kho-
    dadadi, M.: Tracking logical difference in large-scale ontologies: a forgetting-based
    approach. In: Proceedings of the AAAI Conference on Artificial Intelligence. vol. 33,
    pp. 3116–3124 (2019)