<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>In the Eye of the Beholder: Which Proofs are Best?</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Stefan Borgwardt</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Anke Hirsch</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Alisa Kovtunova</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Frederik Wiehr</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>German Research Center for Artificial Intelligence (DFKI)</institution>
          ,
          <addr-line>Saarland Informatics Campus, Saarbrücken</addr-line>
          ,
          <country country="DE">Germany</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Institute of Theoretical Computer Science</institution>
          ,
          <addr-line>TU Dresden</addr-line>
          ,
          <country country="DE">Germany</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Although logical entailments are often considered “explainable”, experience with justifications in DLs has shown that explaining why a logical consequence holds still requires some effort. However, a full formal proof of a DL entailment may be considered too long-winded, and a textual representation of a proof may be preferred. It may also depend on the user's experience and individual preferences which representation of a proof constitutes a good explanation for them. Building on previous work on explaining DL consequences to users, we ran an experiment to compare 4 different forms of proofs: formal proofs and textual proofs, which are either very detailed or condensed. A multiple linear regression with contrast coding revealed that the participants rated short proofs as being easier than long proofs, independent of their representation. On the other hand, we could not verify any influence of prior experience with logic on how easy or difficult the different kinds of proofs were considered.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>
        Explanations of automated decisions are currently an important topic of research.
However, apart from the popular discussion about how “explainable” different
AI methods are, the main task of explanations is understanding, i.e., that the
information transmitted is actually received by the human user, and hence is a
topic for social science research [
        <xref ref-type="bibr" rid="ref21">21</xref>
        ]. This large existing body of research needs to
be taken into account when designing explainable AI systems. Particular lessons
are that the context of an explanation is important, i.e., an explanation depends
on the environment as well as the user, and that an effective explanation should
be a dialog between the explainer and the explainee, allowing for a cycle of
questions and clarifications.
      </p>
      <p>
        In the DL community, research on explanations first focused on proofs for
explaining entailments [
        <xref ref-type="bibr" rid="ref19 ref7">7, 19</xref>
        ], but it was quickly realized that often it is enough
to point out the responsible axioms from the ontology, i.e., so-called
justifications [
        <xref ref-type="bibr" rid="ref10 ref28 ref5">5, 10, 28</xref>
        ]. There have been some experiments on determining factors
Copyright c 2020 for this paper by its authors. Use permitted under Creative
Commons License Attribution 4.0 International (CC BY 4.0).
that make a justification hard to understand [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ]. However, recent research has
been going back towards providing (partial) proofs [
        <xref ref-type="bibr" rid="ref12 ref15">12, 15</xref>
        ]. Explanations using
proofs with intermediate steps have been investigated through several approaches,
for example modified sequent calculi [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ], an empirical study using around 500
OWL ontologies to find new “understandable” inference rules [
        <xref ref-type="bibr" rid="ref22">22</xref>
        ], a probabilistic
model for measuring the understandability of proofs consisting of a few inference
rules [
        <xref ref-type="bibr" rid="ref23">23</xref>
        ], and an approach for designing inference rules keeping the bounded
cognitive resources of the user in mind [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]. In parallel, several approaches for
converting description logic axioms and proofs into textual representations have
been developed and evaluated [
        <xref ref-type="bibr" rid="ref18 ref2 ref24 ref25">2, 18, 24, 25</xref>
        ].
      </p>
      <p>
        In an effort to understand which of these approaches are most promising
for improving explainability, our experiment continues in this line of research.
We investigated which representations of DL proofs are preferred by users (who
had some prior experience in logic). The main formats we used are full formal
proofs in a tree-shaped representation, e.g. based on consequence-based reasoning
procedures [
        <xref ref-type="bibr" rid="ref16 ref30">16, 30</xref>
        ], and linearized translations of these proofs into text, e.g. as
produced by various verbalization techniques [
        <xref ref-type="bibr" rid="ref18 ref2 ref25">2, 18, 25</xref>
        ]. In addition, to find out
how detailed proofs should be, we added shortened representations for each of
these two versions, in which some (easy) reasoning steps were omitted or merged.
We chose these four combinations since they are representative of the
state-of-theart in DL explanations. We did not consider plain justifications as explanations in
our experiment, since we chose examples that were identified as more challenging
in the DL literature, and thus require more in-depth explanation. Differently
from previous studies, we directly compared textual and formal proof formats.
      </p>
      <p>Our main finding is that shorter proofs were rated as being easier than
ones that contained detailed reasoning chains. In particular, the long textual
proof representation was often considered to be too long-winded and repetitive.
However, the experience level of our participants did not seem to influence their
preferences. Nevertheless, there clearly are differences in how different individuals
were working with the different proof representations.
2</p>
    </sec>
    <sec id="sec-2">
      <title>Proofs</title>
      <p>
        We assume a basic familiarity with DLs, in particular ALC [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. Let O be an
ontology and α be an axiom that follows from O (O |= α), which represents a
surprising or unintuitive consequence for the user. A justification is a minimal
subset J ⊆ O such that J |= α, which already points to some of the axioms
from O that are responsible for α. However, actually understanding why α follows
may require a more detailed proof. Informally, a proof is a tree of connected
inference steps α1...αn , where each step is sound, i.e., {α1, . . . αn} |= α holds
α
(see Fig. 1). Usually, one imagines such a proof to be built using inference rules
from an appropriate calculus [
        <xref ref-type="bibr" rid="ref3 ref30">3, 30</xref>
        ]. However, there also exist approaches to
generate DL proofs that are not based on proof calculi. They usually start
with a justification, and extend it with intermediate axioms (lemmas) using
heuristics [
        <xref ref-type="bibr" rid="ref10 ref11">10, 11</xref>
        ], concept interpolation [
        <xref ref-type="bibr" rid="ref27">27</xref>
        ], or forgetting [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ].
A v ⊥
      </p>
      <p>O = { A v ∃r.&gt;,</p>
      <p>C u B v ⊥,
A v ∀r.(B u C) }</p>
      <p>Since the only requirement is soundness of each proof step, adjacent steps can
be merged to obtain shorter proofs, e.g. for the proof in Fig. 1 we could merge
the two steps:</p>
      <p>A v ∃r.&gt;</p>
      <p>A v ∀r.(B u C)</p>
      <p>A v ⊥</p>
      <p>C u B v ⊥
This directly corresponds to the justification J = O. Thus, a justification can also
be seen as a very short proof. However, it can be useful to highlight intermediate
inference steps, for example to pinpoint the precise cause of an unintended
entailment α: even if all axioms from the ontology seem reasonable to the user, at
some point in the proof they will start to have unintuitive consequences, which
can be used to track down the problem.</p>
      <p>A textual representation of a proof is necessarily a linearization, where the
inference steps are explained in a sequence, for example in a top-down left-right
order. A text corresponding to the formal proof in Fig. 1 could be the following:
Since every A has an r-successor and every A has only r-successors that
are B and C, every A has an r-successor that is B and C. Since every A
has an r-successor that is B and C and there is no object which is C and
B at the same time, there is no A.
3</p>
    </sec>
    <sec id="sec-3">
      <title>Experiment</title>
      <p>We conducted an experiment where we assessed participants’ understanding
of these different proof representations. The full details of the experiment are
available online1.
3.1</p>
      <sec id="sec-3-1">
        <title>Description of the experiment</title>
        <p>Introduction and Goals. With the current experiment, we attempt to find
which proof representation is most understandable, also considering experience
with logic. The objective is to find a difference in the difficulty rating of longer
or shorter proofs with either textual or a formal representation.
1 https://cloud.perspicuous-computing.science/s/Wmtmyp8JQNaF2AD
Participants. 16 participants (four female) were assessed with a mean age of 23
(standard deviation = 1.71). Our international participants were recruited from
undergraduate and graduate university students with basic knowledge of logic,
which was required to understand the proofs. Participants were recruited via
advertisements on mailing lists. Screening criteria were familiarity with first-order
logic (e.g. through a lecture), a stable Internet connection, installing a video
conference app with video access (on their mobile device or computer) and the
permission to record their handwriting and voice during the experiment.</p>
        <p>
          In Table 1, the descriptive statistics of the participants’ self-rated experience
with logic can be found.
Conditions and Design. We used two different conditions (factors) with two
levels each. One condition was the representational form of the proof, which
was either textual or formal. The other condition was the length of the proof,
which was either short or long. Thus, there were the four following condition
combinations: Long Text, Short Text, Long Formal, and Short Formal (see Table 2).
We used a 2 × 2 within-subjects design, which means that each participant saw
all four combinations.
Material. The proofs were chosen such that they represent an unintuitive
consequence, e.g. the unsatisfiability of a concept name, or that any amputation
of a finger is also an amputation of the whole hand [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ]. We chose to employ
real-domain examples rather than versions with abstract concept and role names
to increase the readability and motivation, as was also done in [
          <xref ref-type="bibr" rid="ref25">25</xref>
          ]. Additionally,
the domain for the experiment phase must not be too familiar to participants:
with background knowledge they can easily spot controversial axiom(s) in an
ontology and do not require a proof for the unintuitive entailment. Therefore,
all four examples were chosen from the medical domain and were adapted from
examples in the literature on DL explanations, in particular [
          <xref ref-type="bibr" rid="ref14 ref20 ref29 ref6">6, 14, 20, 29</xref>
          ]. For
each of them, four different proof representations (see Table 2) were manually
created, not automatically generated, to make them comparable in difficulty.
To keep consistency among the long formal proofs, we used a fix set of basic
inference rules based on the one for Elk [
          <xref ref-type="bibr" rid="ref16">16</xref>
          ]. Short formal proofs were obtained
by omitting tautologies and merging two transitivity or transitivity and additive
rule instances into one. Textual proofs were obtained by (naively) applying a set
of phrase templates to every inference step in corresponding short or long formal
proofs. For example, for a rule instance with the premises prem1, prem2 and
the conclusion concl, verbalisation of this instance can be as follows: “From the
facts that verb(prem1 ) and verb(prem2 ), it follows that verb(concl)”, where
verb is a unified template verbalization of DL axioms according to their logical
constructors.
        </p>
        <p>In Figure 2, we depict a short formal and a short textual representation for one
of the examples; the longer variants can be found in the appendix. Each proof was
shown below a list of the involved ontology axioms on the left (Cell v Compound
etc.), a textual translation of these axioms on the right (e.g. “Every cell is a
compound.”), as well as a short statement of the entailment (“The ontology above
implies that there is no cell culture.”).</p>
        <p>Procedure and Tasks. Due to the Corona pandemic, we could not conduct
the experiment in person as planned. Instead we used video conferences with
Zoom2 and GoToMeeting3 to communicate with our participants.</p>
        <p>The experiment was conducted in English. Every participant got an individual
appointment. They were asked via e-mail whether they had a printer. If this
was not the case, then the experiment sheets were send to them via post. If
they had a printer, then they received the sheets as a PDF via e-mail with
specific instructions on how to print them. One day before their appointment,
the participants received an e-mail with a reminder and the link to the video
conference meeting. To start the experiment, the participant was welcomed by the
experimenter. The purpose and the procedure of the experiment was explained.
They were then instructed to put the printed or sent papers in front of them and
a link to the questionnaire and a participant number were sent to them. They
went through the first six pages of the questionnaire asking them about their
demographic data and experience with logic. To refresh participant’s memory
and to make sure everyone started with at least the same basic knowledge about
2 https://zoom.us
3 https://www.gotomeeting.com</p>
        <sec id="sec-3-1-1">
          <title>CClt v ∃ct.C u ∀ct.C</title>
        </sec>
        <sec id="sec-3-1-2">
          <title>CClt v ∃ct.(At u C)</title>
        </sec>
        <sec id="sec-3-1-3">
          <title>CClt v MaObj MaObj v ∃ct.At</title>
        </sec>
        <sec id="sec-3-1-4">
          <title>CClt v ∃ct.At C v Cmp At u C v At u Cmp</title>
          <p>At u C v ⊥
At u Cmp v ⊥</p>
        </sec>
        <sec id="sec-3-1-5">
          <title>CClt v ⊥</title>
          <p>Since every cell culture is a material object and every material object contains an
atom, every cell culture contains an atom. From the facts that every cell culture
contains an atom and that every cell culture contains a cell and contains only cells, it
follows that every cell culture contains something which is both an atom and a cell.
Every cell is a compound. Thus, any object which is an atom and a cell at the same
time is also an atom and a compound. There is no object which is an atom and a
compound at the same time. Therefore, there is no object which is both an atom
and a cell.</p>
          <p>Furthermore, since every cell culture contains something which is both an atom and
a cell and there is no object which is both an atom and a cell, there is no cell culture.
description logics, they were shown a 10-minute video with a crash course on
description logics.</p>
          <p>Afterwards, they were asked to arrange the camera setup in such a way that
the papers in front of them and their hands were visible. Then the experimenter
started recording the participant. To demonstrate both proof representations
(textual and formal), a training phase was conducted. The participant saw a
textual proof and a formal proof consecutively. The experimenter explained
that they would see an ontology which implies an unintuitive statement. We
used the think-aloud technique, in which the participant was asked to read the
ontology and the proof carefully and simultaneously summarize the proof out
loud. They were encouraged to take notes on the paper, point or draw connections
between the axioms and the proof and to explain their line of reasoning. After
each proof they were asked to answer the following two questions and to mark
their answers on the sheets: “Which part of the proof was the most difficult to
understand and why?” and “Which axioms in the ontology do you think are most
responsible for the unintuitive consequence and why?” During the training phase
the participant was allowed to ask questions. After the two training examples,
the experiment phase began. One after another, they saw four ontologies, each of
which implied an unintuitive statement. The instructions were the same as in
the training phase except that no questions from the participant were answered
by the experimenter. They were reminded to think aloud while working through
the proofs. To minimize the impact of the order in which they saw the different
proof representations, we used a Latin square, so every combination of the two
conditions was seen first by one fourth of the participants (see Table 2).</p>
          <p>After each proof, the participant filled out one question in the questionnaire
about the perceived difficulty of the proof and received the same two questions as
in the training phase. Subsequently, the final question in the questionnaire asked
the participant to rank the proofs based on their comprehensibility (first rank =
very easy, fourth rank = very difficult). It was possible to give several proofs the
same rank. They were asked to comment on the ranking, what they liked and
disliked about the proofs and afterwards also if they noticed any differences or
similarities between the proofs. They were then told that there were two different
representational forms of proofs and asked which they preferred and why.</p>
          <p>Finally, the experimenter made sure that the participant submitted the
questionnaire, stopped the recording and told them to fill in their participant
number on each sheet, scan it or take a photo and send us the final sheets.</p>
          <p>To assure a stable video conference connection, the experimenter’s video was
off the whole time. To match the video and the questionnaire after the experiment,
each participant was given the participant number, i.e., an individual code which
they had to use in the questionnaire and on the sheets and under which we saved
the video. Overall the experiment lasted around one and a half hours. After
receiving all six pages (two for the training and four for the experiment phase),
we sent them a code for an Amazon voucher worth 20 e.</p>
          <p>To make sure the participants really understood the proofs a logic expert
reviewed the video of each participant after each session. Due to the
thinkaloud technique the expert was able to follow the participant’s thought and
rated the video based on the participant’s understanding on a scale from 1 (no
understanding) to 3 (complete understanding).</p>
          <p>Independent and Dependent Variables. To assess participants’ logic
experience we asked them how they would rate their experience with propositional,
description, first order logic on a Likert-like rating scale from 1 (no knowledge)
to 5 (expert). We further evaluated how they rated the difficulty of each proof
on a Likert-like rating scale from 1 (very easy) to 5 (very difficult). To compare
the different proof representations, we asked the participants to rank the proofs
at the end of the experiment based on their comprehensibility.</p>
          <p>Hypotheses. We stated three hypotheses concerning the participants’ self-rating
of the difficulty of the proofs.</p>
          <p>Hypothesis 1 : It is easier to understand a short, concise explanation than a
longer version (in the same representation format). This will be shown by a lower
difficulty rating for the short proofs than for the long proofs.</p>
          <p>Hypothesis 2 : Users with less experience in logic can understand the longer
text better than a short formal proof. This will be shown by a lower difficulty
rating of the long textual proof.</p>
          <p>Hypothesis 3 : Users with more experience in logic can understand a long
formal proof better than a long text. This will be shown by a lower difficulty
rating of the long formal proof.
Problems and Limitations. Originally, this experiment was planned as
inperson: all the technical equipment as well as a suitable room would have been
arranged by us to minimize possible distractions. However, due to the Corona
pandemic, we had to re-design the experiment to an online setting. As immediate
drawbacks of a less controlled environment, we encountered i) technical problems
with the video conference due to unstable Internet connections, ii) poor quality of
video and audio recordings, e.g. a low-resolution phone camera, iii) compatibility
issues with different browsers, e.g. for video playback, iv) concerns about privacy
for receiving and storing the video and audio recordings of participants and their
homes, v) various interruptions related to the environment, i.e., family members
or domestic animals.</p>
          <p>Additionally, there was a more conceptual problem concerning prior knowledge
about the example domains. For example, if a participant sees a (for them)
contradictory axiom in the ontology (e.g. “every vegetarian is a person,” because
they would consider herbivorous animals to also be vegetarians; or “every artist
is a professional,” since there exist also amateur artists),4 they would not focus
anymore on the actual task of understanding why this ontology implies the goal
axiom.
3.2</p>
        </sec>
      </sec>
      <sec id="sec-3-2">
        <title>Results</title>
        <p>
          Quantitative Results. To compute the quantitative analyses IBM SPSS
Statistics (Version 26) predictive analytics software for Windows [
          <xref ref-type="bibr" rid="ref13">13</xref>
          ] and the Macro
PROCESS [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ] was used. For all hypotheses, we used a p-value threshold of 0.05.
        </p>
        <p>For Hypothesis 1, a multiple linear regression with contrast coding (K1, K2,
K3) was conducted. K1 contrasted the textual representation against the formal
one. K2 contrasted the short vs. long proofs and K3 the interaction between the
two general conditions. The three contrasts explained 14.2% of variance in the
rating after each proof, R2 = .14, F(3, 60) = 3.30, p &lt; .05. Only K2 was found
to be a significant predictor in the linear regression, β = −.29, t(60) = −2.42,
p &lt; .05. This means that the participants rated the shorter proofs as being easier
than the longer ones, which was independent of the presentation format. Thus,
Hypothesis 1 could be supported by our data.</p>
        <p>For Hypotheses 2 and 3, we want to assess if the strength of the relationship
between the independent variables (condition combinations) and the dependent
variable (rating after each proof) changes with the third variable (experience),
i.e., whether there is an interaction between the conditions and the experience.
We computed moderator analyses with the two condition combinations as a
predictor, the experience as a moderator variable and the rating after each
proof as the criterion. To put the conditions into the moderator analysis for
Hypothesis 2, the Long Text was coded with +1 and Short Formal with -1 (and
similarly for Hypothesis 3 with Long Text and Long Formal). For each self-rated
measure of experience (propositional logic, first-order logic, description logic),
one moderation analysis was computed.
4 These occurred in the training examples.</p>
        <p>For Hypothesis 2, the moderation analyses for the self-rated experience
with propositional and first-order logic revealed a significant model, R2 = .25,
F(3, 28) = 3.07, p &lt; .05 and R2 = .31, F(3, 28) = 4.21, p &lt; .05, respectively, but
only the two main effects were shown to be significant and not the interaction
(see Table 3). Thus, there was no significant moderation. The main effect of
the condition combinations showed that the Short Formal representation was
rated as being easier than the Long Text. The main effect of experience means
that, the more experience the participants had, the easier they rated the proof.
However, the two main effects were independent from each other and showed
no interaction, which means that neither the experience with propositional logic
nor the experience with first-order logic influenced the relationship between the
condition and the rating. For the self-rated experience with description logic, the
moderation model was not found to be significant. Thus, Hypothesis 2 could not
be supported by our data. Experience with logic did not make a difference on
the understanding of the different proof representations.</p>
        <p>For Hypothesis 3, from the three moderation models, only the model with
the self-rated experience with propositional logic turned out to be significant,
R2 = .25, F(3, 28) = 3.07, p &lt; .05. Only the main effect of experience was
significant here, β = −.43, t(28) = −2.10, p &lt; .05. Thus, also Hypothesis 3 could
not be supported by our data. Experience with logic did not make a difference
on the understanding of the different long conditions.</p>
        <p>Additionally to the three hypotheses, we used Friedman’s ANOVA for
comparing the comprehensibility ranking of the proof representations at the end of
the experiment (first rank = very easy, fourth rank = very difficult). It revealed a
significant difference in the ranking of the condition combinations, χ2(3) = 15.29,
p &lt; .01 with a moderate effect size (Kendall’s W = .32). For the post-hoc pairwise
comparisons Bonferroni correction was used which resulted in a p-threshold of
0.008, resulting in only two significant comparisons.</p>
        <p>The participants’ ranking of condition combinations is shown in Figure 3.
The combination Short Text was preferred over Long Text, Z = 1.53, p &lt; .008.
The median ranking for Short Text and Long Text was 2 and 3.5, respectively.
Additionally, Short Formal was preferred over Long Text, Z = 1.50, p &lt; .008.
8
ts7
n6
a
p
ic5
itr
fp4
a
o
r3
e
b
m2
u
N
1
0
8
ts7
itrcapn6
i5
a
fsep4
o
r3
b
2
m
u
N
1
0
8
ts7
n6
a
p
ic5
itr
fp4
a
o
r3
e
b
m2
u
N
1
0
8
ts7
itrcapn6
i5
a
fsep4
o
r3
b
2
m
u
N
1
0
1.Rank
2.Rank 3.Rank</p>
        <p>Chosen rank
Long Formal
4.Rank
1.Rank
2.Rank
3.Rank</p>
        <p>4.Rank</p>
        <p>Chosen rank
Short Formal
Short Formal had the lowest median ranking with 1.50. Both comparisons showed
moderate effect sizes, r = 0.38 for both. Median ranking for Long Formal was 2.</p>
        <p>Only one participant chose Long Text on the first rank. Their statement can
be found in the next subsection. However, nobody put Long Text on the second
rank, but 15 chose the third or fourth rank for it. Thus, most participants ranked
it as (very) difficult. Short Text was never assigned the fourth rank, but by 13
participants it was considered very easy or easy.</p>
        <p>Long Text
Short Text
1.Rank
2.Rank 3.Rank</p>
        <p>Chosen rank
4.Rank
1.Rank
2.Rank
3.Rank</p>
        <p>4.Rank
Chosen rank
Qualitative Results. For the qualitative evaluation, we transcribed the audio
recordings into text and arranged the texts according to the condition or condition
combination. Then, we compared the participants’ statements on each condition
based on similarity and differences according to the hypotheses and grouped
them contentwise.</p>
        <p>In general, the participants’ statements also supported the view that the
formal proofs were preferred over the textual proofs. Participants described the
formal proofs as “easier to understand” and also as “clearer”. One participant
mentioned that for them it was “easier to find certain parts” of the proof. Another
one that that their “orientation is better” within the proof, in order to go through
it step-by-step, and that it is “easy to follow the proof”. The textual proofs were
often characterized as inconvenient, “less understandable” or “hard to understand”
or even “annoying”.</p>
        <p>On the other hand, one participant stated that “for short proofs text is ok.”
Only one participant said that they “preferred a proof explained through words
rather than through a schema.” The participant commented on it as follows:
“Texts add more redundancy and this redundancy helps with connecting. My
intuition for language can spot inconsistencies better. So, if a domain is abstract
or unknown and a proof is complex (like in Mathematics) and requires to put
together several pieces and arguing about them, wording might help a lot.”</p>
        <p>
          Many participants were struggling with inferences involving both ∀- and
∃quantifiers in formal proofs. However, its textual equivalent or a second encounter
of ∀+∃ were perceived much better (this combination appeared already in the
training examples). Several participants mentioned that a graphical presentation
is better when there are parallel threads: “I could clearly see what and where
they come from. And then forget about what I just learned so I could go to
another part. And once I again need this part, I know exactly where it came
from later.” Also, some participants found it very difficult to deal with longer
domain-specific terms, since it requires more time to understand and to keep
track of them. These opinions confirm the motivation behind [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ] for restricting
the length of axioms and developing a proof system that takes into account
bounded cognitive resources.
4
        </p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Discussion and Conclusions</title>
      <p>
        Short proofs were rated as being easier than long proofs, independent of the
presentation format. Thus, future experiments and theoretical approaches should
focus on shortening proofs. With our data, Hypotheses 2 and 3 could not be
supported. However, the rankings and discussions clearly showed individual user
preferences between formal and textual representations. For further research in
this direction, the questions from [
        <xref ref-type="bibr" rid="ref26">26</xref>
        ] could be used to get another quantitative
approach. One possibility to further assess the difference between formal and
textual representation could be to include experts working in the field of logic,
like computer scientists and mathematicians teaching logic at a university. This
way, there could be a clearer distinction between novices, e.g. students having
attended a single lecture about logic, vs. experts. Maybe then one could find an
influence of experience on the perception of difficulty of the proofs.
      </p>
      <p>
        On the other hand, the ultimate goal is to explain logical conclusions to
domain experts who are not familiar with logic. Here, an interesting direction
of study is to generate (concise) textual explanations [
        <xref ref-type="bibr" rid="ref18 ref2 ref25">2, 18, 25</xref>
        ], or perhaps a
combination of graphical and textual elements to better convey the structure of
a proof while still providing each (derived) axiom in a readable form.
      </p>
      <p>
        From a procedural point of view, it would be preferable to use a
betweensubjects design (different people test each condition) instead of within-subjects
(when the same person tests all the conditions), to minimize learning effects, which
however requires more participants. Of course we would also like to compare other
proof representations, e.g. pure justifications, linear vs. non-linear formats, mixed
formal/textual presentations as mentioned above, incorporating annotations such
as axiom numbering or coloring, and most importantly interactive approaches
such as the proof plugin for the Protégé ontology editor [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ]. The main goal
with these different representations should always be usability, which has to be
assessed experimentally.
      </p>
      <p>
        As was demonstrated by the participants’ different opinions and preferences
about proof representations, it makes sense to incorporate the user as an active
element in the design of a suitable presentation. User modeling [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ] can help
make automatic design decisions, by taking into account user preferences or the
user’s existing knowledge, e.g. in the form of a background ontology that the user
is assumed to know intimately.
      </p>
      <p>Another take-away message we learned from our experiment is that preparing
such experiments to take place online from the beginning also has an upside since
then they are more resilient against unforeseen events.</p>
      <p>Acknowledgements This work was partially supported by DFG grant 389792660
as part of TRR 248 (https://perspicuous-computing.science). We are also
grateful to our colleagues who forwarded the advertisement for the experiment
to their students, and of course to the participants.</p>
    </sec>
    <sec id="sec-5">
      <title>Long Proofs</title>
      <p>On the following two pages we depict the long versions of the proofs in Figure 2:
Every cell culture contains a cell and contains only cells. Thus, every cell culture
contains only cells. Since every cell culture is a material object and every material
object contains an atom, every cell culture contains an atom. From the facts that
every cell culture contains an atom and that every cell culture contains only cells, it
follows that every cell culture contains something which is both an atom and a cell.
Trivially, every object which is both an atom and a cell is an atom and a cell. Thus,
every object which is both an atom and a cell is a cell. Every cell is a compound.
Therefore, every object which is both an atom and a cell is a compound. Trivially,
every object which is both an atom and a cell is an atom and a cell. Thus, every
object which is both an atom and a cell is an atom. From the facts that every object
which is both an atom and a cell is an atom and that every object which is both an
atom and a cell is a compound, it follows that every object which is both an atom
and a cell is both an atom and a compound. There is no object which is both a
compound and an atom at the same time. Therefore, there is no object which is both
an atom and a cell.</p>
      <p>v
oC lle
lle llv Cu
C e</p>
      <p>l C m
ule</p>
      <p>u nd to
vC
ll
e u
C m</p>
      <p>m
uto o</p>
      <p>t
m A A
o
t</p>
      <p>v
A ll
e</p>
      <p>C
ll u
e m
C o</p>
      <p>m t
uo A</p>
      <p>t
m A
o
t
A lv</p>
      <p>l
ve
ll C
e u
C</p>
      <p>m
uo</p>
      <p>t
m A
o
t
A
⊥
M
m
o
v t</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Alrabbaa</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Baader</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Borgwardt</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Koopmann</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kovtunova</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Finding small proofs for description logic entailments: Theory and practice</article-title>
          . In: Albert,
          <string-name>
            <given-names>E.</given-names>
            ,
            <surname>Kovacs</surname>
          </string-name>
          ,
          <string-name>
            <surname>L</surname>
          </string-name>
          . (eds.) LPAR-
          <volume>23</volume>
          : 23rd International Conference on Logic for Programming,
          <source>Artificial Intelligence and Reasoning</source>
          . EPiC Series in Computing, vol.
          <volume>73</volume>
          , pp.
          <fpage>32</fpage>
          -
          <lpage>67</lpage>
          . EasyChair (
          <year>2020</year>
          ). https://doi.org/10.29007/nhpp
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Androutsopoulos</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lampouras</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Galanis</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          :
          <article-title>Generating natural language descriptions from OWL ontologies: The NaturalOWL system</article-title>
          .
          <source>Journal of Artificial Intelligence Research</source>
          <volume>48</volume>
          ,
          <fpage>671</fpage>
          -
          <lpage>715</lpage>
          (
          <year>2013</year>
          ). https://doi.org/10.1613/jair.4017
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Baader</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Brandt</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lutz</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          :
          <article-title>Pushing the EL envelope</article-title>
          . In: Kaelbling,
          <string-name>
            <given-names>L.P.</given-names>
            ,
            <surname>Saffiotti</surname>
          </string-name>
          ,
          <string-name>
            <surname>A</surname>
          </string-name>
          . (eds.)
          <source>Proc. of the 19th Int. Joint Conf. on Artificial Intelligence (IJCAI'05)</source>
          . pp.
          <fpage>364</fpage>
          -
          <lpage>369</lpage>
          . Professional Book Center (
          <year>2005</year>
          ), http://ijcai.org/ Proceedings/09/Papers/053.pdf
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Baader</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Horrocks</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lutz</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sattler</surname>
            ,
            <given-names>U.</given-names>
          </string-name>
          :
          <article-title>An Introduction to Description Logic</article-title>
          . Cambridge University Press (
          <year>2017</year>
          ). https://doi.org/10.1017/9781139025355
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Baader</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Peñaloza</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Suntisrivaraporn</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          :
          <article-title>Pinpointing in the description logic EL+</article-title>
          .
          <source>In: Proc. of the 30th German Annual Conf. on Artificial Intelligence (KI'07). Lecture Notes in Computer Science</source>
          , vol.
          <volume>4667</volume>
          , pp.
          <fpage>52</fpage>
          -
          <lpage>67</lpage>
          . Springer, Osnabrück, Germany (
          <year>2007</year>
          ). https://doi.org/10.1007/978-3-
          <fpage>540</fpage>
          -74565-
          <issue>5</issue>
          _
          <fpage>7</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Baader</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Suntisrivaraporn</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          :
          <article-title>Debugging SNOMED CT using axiom pinpointing in the description logic EL+</article-title>
          .
          <source>In: Proc. of the 3rd Conference on Knowledge Representation in Medicine (KR-MED</source>
          '08):
          <article-title>Representing and Sharing Knowledge Using SNOMED. CEUR-WS</article-title>
          , vol.
          <volume>410</volume>
          (
          <year>2008</year>
          ), http://ceur-ws.
          <source>org/</source>
          Vol-
          <volume>410</volume>
          /Paper01. pdf
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Borgida</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Franconi</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Horrocks</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          :
          <article-title>Explaining ALC subsumption</article-title>
          .
          <source>In: ECAI 2000, Proceedings of the 14th European Conference on Artificial Intelligence</source>
          , Berlin, Germany,
          <source>August 20-25</source>
          ,
          <year>2000</year>
          . pp.
          <fpage>209</fpage>
          -
          <lpage>213</lpage>
          (
          <year>2000</year>
          ), http://www.frontiersinai. com/ecai/ecai2000/pdf/p0209.pdf
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>Engström</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Nizamani</surname>
            ,
            <given-names>A.R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Strannegård</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          :
          <article-title>Generating comprehensible explanations in description logic</article-title>
          .
          <source>In: Informal Proceedings of the 27th International Workshop on Description Logics</source>
          , Vienna, Austria,
          <source>July 17-20</source>
          ,
          <year>2014</year>
          . pp.
          <fpage>530</fpage>
          -
          <lpage>542</lpage>
          (
          <year>2014</year>
          ), http://ceur-ws.
          <source>org/</source>
          Vol-
          <volume>1193</volume>
          /paper_17.pdf
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Hayes</surname>
            ,
            <given-names>A.F.</given-names>
          </string-name>
          :
          <article-title>Introduction to mediation, moderation, and conditional process analysis: A regression-based approach</article-title>
          .
          <source>Guilford Publications</source>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Horridge</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Justification Based Explanation in Ontologies</article-title>
          .
          <source>Ph.D. thesis</source>
          , University of Manchester, UK (
          <year>2011</year>
          ), https://www.research.manchester.ac.uk/portal/ files/54511395/FULL_TEXT.PDF
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>Horridge</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bail</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Parsia</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sattler</surname>
            ,
            <given-names>U.</given-names>
          </string-name>
          :
          <article-title>Toward cognitive support for OWL justifications</article-title>
          .
          <source>Knowledge-Based Systems 53</source>
          ,
          <fpage>66</fpage>
          -
          <lpage>79</lpage>
          (
          <year>2013</year>
          ). https://doi.org/10. 1016/j.knosys.
          <year>2013</year>
          .
          <volume>08</volume>
          .021
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <string-name>
            <surname>Horridge</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Parsia</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sattler</surname>
            ,
            <given-names>U.</given-names>
          </string-name>
          :
          <article-title>Justification oriented proofs in OWL</article-title>
          .
          <source>In: The Semantic Web - ISWC 2010 - 9th International Semantic Web Conference, ISWC 2010</source>
          , Shanghai, China, November 7-
          <issue>11</issue>
          ,
          <year>2010</year>
          ,
          <string-name>
            <given-names>Revised</given-names>
            <surname>Selected</surname>
          </string-name>
          <string-name>
            <surname>Papers</surname>
          </string-name>
          , Part I. pp.
          <fpage>354</fpage>
          -
          <lpage>369</lpage>
          (
          <year>2010</year>
          ). https://doi.org/10.1007/978-3-
          <fpage>642</fpage>
          -17746-0_
          <fpage>23</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>13. IBM Corp.: IBM SPSS Statistics for Windows [computer software], https://www. ibm.com/products/spss-statistics</mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          14.
          <string-name>
            <surname>Kalyanpur</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Debugging and Repair of OWL Ontologies</article-title>
          .
          <source>Ph.D. thesis</source>
          , University of Maryland, College Park, USA (
          <year>2006</year>
          ), http://hdl.handle.net/
          <year>1903</year>
          /3820
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          15.
          <string-name>
            <surname>Kazakov</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Klinov</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Stupnikov</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Towards reusable explanation services in Protege</article-title>
          . In: Artale,
          <string-name>
            <given-names>A.</given-names>
            ,
            <surname>Glimm</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            ,
            <surname>Kontchakov</surname>
          </string-name>
          ,
          <string-name>
            <surname>R</surname>
          </string-name>
          . (eds.)
          <source>Proc. of the 30th Int. Workshop on Description Logics (DL'17)</source>
          .
          <source>CEUR Workshop Proceedings</source>
          , vol.
          <source>1879</source>
          (
          <year>2017</year>
          ), http://www.ceur-ws.
          <source>org/</source>
          Vol-1879/paper31.pdf
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          16.
          <string-name>
            <surname>Kazakov</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Krötzsch</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Simancik</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          :
          <article-title>The incredible ELK - from polynomial procedures to efficient reasoning with EL ontologies</article-title>
          .
          <source>J. Autom. Reasoning</source>
          <volume>53</volume>
          (
          <issue>1</issue>
          ),
          <fpage>1</fpage>
          -
          <lpage>61</lpage>
          (
          <year>2014</year>
          ). https://doi.org/10.1007/s10817-013-9296-3
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          17.
          <string-name>
            <surname>Kobsa</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wahlster</surname>
            ,
            <given-names>W.:</given-names>
          </string-name>
          <article-title>User models in dialog systems</article-title>
          . Springer (
          <year>1989</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          18.
          <string-name>
            <surname>Kuhn</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          :
          <article-title>The understandability of OWL statements in controlled english</article-title>
          .
          <source>Semantic Web</source>
          <volume>4</volume>
          (
          <issue>1</issue>
          ),
          <fpage>101</fpage>
          -
          <lpage>115</lpage>
          (
          <year>2013</year>
          ). https://doi.org/10.3233/SW-2012-0063
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          19.
          <string-name>
            <surname>McGuinness</surname>
            ,
            <given-names>D.L.</given-names>
          </string-name>
          :
          <article-title>Explaining Reasoning in Description Logics</article-title>
          .
          <source>Ph.D. thesis</source>
          , Rutgers University, NJ, USA (
          <year>1996</year>
          ). https://doi.org/10.7282/t3-q0c6-5305
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          20.
          <string-name>
            <surname>Meehan</surname>
            ,
            <given-names>T.F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Masci</surname>
            ,
            <given-names>A.M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Abdulla</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cowell</surname>
            ,
            <given-names>L.G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Blake</surname>
            ,
            <given-names>J.A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mungall</surname>
            ,
            <given-names>C.J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Diehl</surname>
            ,
            <given-names>A.D.</given-names>
          </string-name>
          :
          <article-title>Logical development of the cell ontology</article-title>
          .
          <source>BMC Bioinformatics</source>
          <volume>12</volume>
          (
          <issue>1</issue>
          ),
          <volume>6</volume>
          (
          <year>2011</year>
          ). https://doi.org/10.1186/
          <fpage>1471</fpage>
          -2105-12-6
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          21.
          <string-name>
            <surname>Miller</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          :
          <article-title>Explanation in artificial intelligence: Insights from the social sciences</article-title>
          .
          <source>Artif. Intell</source>
          .
          <volume>267</volume>
          ,
          <fpage>1</fpage>
          -
          <lpage>38</lpage>
          (
          <year>2019</year>
          ). https://doi.org/10.1016/j.artint.
          <year>2018</year>
          .
          <volume>07</volume>
          .007
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          22.
          <string-name>
            <surname>Nguyen</surname>
            ,
            <given-names>T.A.T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Power</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Piwek</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Williams</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          :
          <article-title>Measuring the understandability of deduction rules for OWL</article-title>
          .
          <source>In: Proceedings of the First International Workshop on Debugging Ontologies and Ontology Mappings</source>
          ,
          <source>WoDOOM</source>
          <year>2012</year>
          , Galway, Ireland, October 8,
          <year>2012</year>
          . pp.
          <fpage>1</fpage>
          -
          <lpage>12</lpage>
          (
          <year>2012</year>
          ), http://www.ida.liu.se/~patla/conferences/ WoDOOM12/papers/paper4.pdf
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          23.
          <string-name>
            <surname>Nguyen</surname>
            ,
            <given-names>T.A.T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Power</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Piwek</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Williams</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          :
          <article-title>Predicting the understandability of OWL inferences</article-title>
          . In: Cimiano,
          <string-name>
            <given-names>P.</given-names>
            ,
            <surname>Corcho</surname>
          </string-name>
          , Ó.,
          <string-name>
            <surname>Presutti</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hollink</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rudolph</surname>
          </string-name>
          , S. (eds.)
          <source>The Semantic Web: Semantics and Big Data</source>
          , 10th International Conference, ESWC 2013, Montpellier, France, May
          <volume>26</volume>
          -30,
          <year>2013</year>
          .
          <source>Proceedings. Lecture Notes in Computer Science</source>
          , vol.
          <volume>7882</volume>
          , pp.
          <fpage>109</fpage>
          -
          <lpage>123</lpage>
          . Springer (
          <year>2013</year>
          ). https://doi.org/10. 1007/978-3-
          <fpage>642</fpage>
          -38288-
          <issue>8</issue>
          _
          <fpage>8</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          24.
          <string-name>
            <surname>Schiller</surname>
            ,
            <given-names>M.R.G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Glimm</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          :
          <article-title>Towards explicative inference for OWL</article-title>
          .
          <source>In: Informal Proceedings of the 26th International Workshop on Description Logics</source>
          , Ulm, Germany,
          <source>July 23 - 26</source>
          ,
          <year>2013</year>
          . pp.
          <fpage>930</fpage>
          -
          <lpage>941</lpage>
          (
          <year>2013</year>
          ), http://ceur-ws.
          <source>org/</source>
          Vol-
          <volume>1014</volume>
          / paper_36.pdf
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          25.
          <string-name>
            <surname>Schiller</surname>
            ,
            <given-names>M.R.G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Schiller</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Glimm</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          :
          <article-title>Testing the adequacy of automated explanations of EL subsumptions</article-title>
          .
          <source>In: Proceedings of the 30th International Workshop on Description Logics</source>
          , Montpellier, France,
          <source>July 18-21</source>
          ,
          <year>2017</year>
          . (
          <year>2017</year>
          ), http://ceur-ws.
          <source>org/</source>
          Vol-1879/paper43.pdf
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          26.
          <string-name>
            <surname>Schiller</surname>
            ,
            <given-names>M.R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Schiller</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Glimm</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          :
          <article-title>Testing the adequacy of automated explanations of el subsumptions</article-title>
          .
          <source>Description Logics</source>
          <year>1879</year>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          27.
          <string-name>
            <surname>Schlobach</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          :
          <article-title>Explaining subsumption by optimal interpolation</article-title>
          . In: Alferes,
          <string-name>
            <given-names>J.J.</given-names>
            ,
            <surname>Leite</surname>
          </string-name>
          ,
          <string-name>
            <surname>J.A</surname>
          </string-name>
          . (eds.)
          <source>Proc. of the 9th Eur. Conf. on Logics in Artificial Intelligence (JELIA'04). Lecture Notes in Computer Science</source>
          , vol.
          <volume>3229</volume>
          , pp.
          <fpage>413</fpage>
          -
          <lpage>425</lpage>
          . SpringerVerlag (
          <year>2004</year>
          ). https://doi.org/10.1007/978-3-
          <fpage>540</fpage>
          -30227-8_
          <fpage>35</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          28.
          <string-name>
            <surname>Schlobach</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cornet</surname>
          </string-name>
          , R.:
          <article-title>Non-standard reasoning services for the debugging of description logic terminologies</article-title>
          . In: Gottlob,
          <string-name>
            <given-names>G.</given-names>
            ,
            <surname>Walsh</surname>
          </string-name>
          , T. (eds.)
          <source>Proc. of the 18th Int. Joint Conf. on Artificial Intelligence (IJCAI</source>
          <year>2003</year>
          ). pp.
          <fpage>355</fpage>
          -
          <lpage>362</lpage>
          . Morgan Kaufmann, Acapulco, Mexico (
          <year>2003</year>
          ), http://ijcai.org/Proceedings/03/Papers/053.pdf
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          29.
          <string-name>
            <surname>Schulz</surname>
            ,
            <given-names>S.:</given-names>
          </string-name>
          <article-title>The role of foundational ontologies for preventing bad ontology design</article-title>
          .
          <source>In: Proc. of the 1st Int. Workshop on BadOntoloGy (BOG'18)</source>
          <article-title>, part of The Joint Ontology Workshops (JOWO'18)</article-title>
          .
          <source>CEUR Workshop Proceedings</source>
          , vol.
          <volume>2205</volume>
          . CEURWS.org (
          <year>2018</year>
          ), http://ceur-ws.
          <source>org/</source>
          Vol-
          <volume>2205</volume>
          /paper22_bog1.pdf
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          30.
          <string-name>
            <surname>Simancik</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kazakov</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Horrocks</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          :
          <article-title>Consequence-based reasoning beyond horn ontologies</article-title>
          .
          <source>In: IJCAI 2011, Proceedings of the 22nd International Joint Conference on Artificial Intelligence</source>
          , Barcelona, Catalonia, Spain,
          <source>July 16-22</source>
          ,
          <year>2011</year>
          . pp.
          <fpage>1093</fpage>
          -
          <lpage>1098</lpage>
          (
          <year>2011</year>
          ). https://doi.org/10.5591/978-1-
          <fpage>57735</fpage>
          -516-8/
          <fpage>IJCAI11</fpage>
          -187
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>