<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Probabilistic Implicational Bases in FCA and Probabilistic Bases of GCIs in EL⊥</article-title>
      </title-group>
      <contrib-group>
        <aff id="aff0">
          <label>0</label>
          <institution>Institute for Theoretical Computer Science</institution>
          ,
          <addr-line>TU Dresden</addr-line>
          ,
          <country country="DE">Germany</country>
        </aff>
      </contrib-group>
      <fpage>193</fpage>
      <lpage>204</lpage>
      <abstract>
        <p>A probabilistic formal context is a triadic context whose third dimension is a set of worlds equipped with a probability measure. After a formal definition of this notion, this document introduces probability of implications, and provides a construction for a base of implications whose probability satisfy a given lower threshold. A comparison between confidence and probability of implications is drawn, which yields the fact that both measures do not coincide, and cannot be compared. Furthermore, the results are extended towards the light-weight description logic EL⊥ with probabilistic interpretations, and a method for computing a base of general concept inclusions whose probability fulfill a certain lower bound is proposed.</p>
      </abstract>
      <kwd-group>
        <kwd>Formal Concept Analysis</kwd>
        <kwd>Description Logics</kwd>
        <kwd>Probabilistic Formal Context</kwd>
        <kwd>Probabilistic Interpretation</kwd>
        <kwd>Implication</kwd>
        <kwd>General Concept Inclusion</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1 Introduction</title>
      <p>
        Most data-sets from real-world applications contain errors and noise. Hence, for mining
them special techniques are necessary in order to circumvent the expression of the
errors. This document focuses on rule mining, especially we attempt to extract rules
that are approximately valid in data-sets, or families of data-sets, respectively. There
are at least two measures for the approximate soundness of rules, namely confidence
and probability. While confidence expresses the number of counterexamples in a single
data-set, probability expresses somehow the number of data-sets in a data-set family
that do not contain any counterexample. More specifically, we consider implications
in the formal concept analysis setting [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ], and general concept inclusions (GCIs) in the
description logics setting [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] (in the light-weight description logic EL⊥).
      </p>
      <p>
        Firstly, for axiomatizing rules from formal contexts possibly containing wrong
incidences or having missing incidences the notion of a partial implication (also called
association rule) and confidence has been defined by Luxenburger in [12].
Furthermore, Luxenburger introduced a method for the computation of a base of all partial
implications holding in a formal context whose confidence is above a certain threshold.
In [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] Borchmann has extended the results to the description logic EL⊥ by adjusting
the notion of confidence to GCIs, and also gave a method for the construction of a base
of confident GCIs for an interpretation.
      </p>
      <p>
        Secondly, another perspective is a family of data-sets representing different views
of the same domain, e.g., knowledge of different persons, or observations of an
experiment that has been repeated several times, since some effects could not be observed
in every case. In the field of formal concept analysis, Vityaev, Demin, and Ponomaryov,
have introduced in probabilistic extensions of formal contexts and their formal concepts
and implications, and furthermore gave some methods for their computation, cf. [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ].
In [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] the author has shown some methods for the computation of a base of GCIs in
probabilistic description logics where concept and role constructors are available to
express probability directly in the concept descriptions. Here, we want to use another
approach, and do not allow for probabilistic constructors, but define the notion of a
probability of general concept inclusions in the light-weight description logic EL⊥.
Furthermore, we provide a method for the computation of a base of GCIs satisfying
a certain lower threshold for the probability. More specifically, we use the description
logic EL⊥ with probabilistic interpretations that have been introduced by Lutz and
Schröder in [11]. Beforehand, we only consider conjunctions in the language of formal
concept analysis, and define the notion of a probabilistic formal context in a more
general form than in [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ], and provide a technique for the computation of base of
implications satisfying a given lower probability threshold.
      </p>
      <p>The document is structured as follows. In Section 2 some basic notions for
probabilistic extensions of formal concept analysis are defined. Then, in Section 3 a method for the
computation of a base for all implications satisfying a given lower probability threshold
in a probabilistic formal context is developed, and its correctness is proven. The
following sections extend the results to the description logic EL⊥. In particular, Section 4
introduces the basic notions for EL⊥, and defines probabilistic interpretations. Section 5
shows a technique for the construction of a base of GCIs holding in a probabilistic
interpretation and fulfilling a lower probability threshold. Furthermore, a comparison
of the notions of confidence and probability is drawn at the end of Section 3.
2</p>
    </sec>
    <sec id="sec-2">
      <title>Probabilistic Formal Concept Analysis</title>
      <p>
        A probability measure P on a countable set W is a mapping P : 2W → [
        <xref ref-type="bibr" rid="ref1">0, 1</xref>
        ] such that
P(∅) = 0 and P(W) = 1 hold, and P is σ-additive, i.e., for all pairwise disjoint
countable families (Un)n∈N with Un ⊆ W it holds that P(Sn∈N Un) = ∑n∈N P(Un).
A world w ∈ W is possible if P{ w } &gt; 0 holds, and impossible otherwise. The set of
all possible worlds is denoted by Wε, and the set of all impossible worlds is denoted
by W0. Obviously, Wε ] W0 is a partition of W.
      </p>
      <p>Definition 1 (Probabilistic Formal Context). A probabilistic formal context K is a
tuple (G, M, W, I, P) that consists of a set G of objects, a set M of attributes, a countable set</p>
      <sec id="sec-2-1">
        <title>W of worlds, an incidence relation I ⊆ G × M × W, and a probability measure P on W.</title>
      </sec>
      <sec id="sec-2-2">
        <title>For a triple (g, m, w) ∈ I we say that object g has attribute m in world w. Furthermore, we</title>
        <p>define the derivations in world w as operators ·Iw : 2G → 2M and ·Iw : 2M → 2G where
AIw := { m ∈ M | ∀g ∈ A : (g, m, w) ∈ I }</p>
        <p>BIw := { g ∈ G | ∀m ∈ B : (g, m, w) ∈ I }
for object sets A ⊆ G and attribute sets B ⊆ M, i.e., AIw is the set of all common attributes of
all objects in A in the world w, and BIw is the set of all objects that have all attributes in B in w.
Definition 2 (Implication, Probability). Let K = (G, M, W, I, P) be a probabilistic
formal context. For attribute sets X, Y ⊆ M we call X → Y an implication over M, and its
probability in K is defined as the measure of the set of worlds it holds in, i.e.,</p>
        <p>P(X → Y) := P{ w ∈ W | XIw ⊆ YIw }.</p>
        <p>Furthermore, we define the following properties for implications X → Y:</p>
        <sec id="sec-2-2-1">
          <title>1. X → Y holds in world w of K if XIw ⊆ YIw is satisfied.</title>
        </sec>
      </sec>
      <sec id="sec-2-3">
        <title>2. X → Y certainly holds in K if it holds in all worlds of K.</title>
      </sec>
      <sec id="sec-2-4">
        <title>3. X → Y almost certainly holds in K if it holds in all possible worlds of K.</title>
      </sec>
      <sec id="sec-2-5">
        <title>4. X → Y possibly holds in K if it holds in a possible world of K.</title>
      </sec>
      <sec id="sec-2-6">
        <title>5. X → Y is impossible in K if it does not hold in any possible world of K.</title>
      </sec>
      <sec id="sec-2-7">
        <title>6. X → Y is refuted by K if does not hold in any world of K.</title>
        <p>It is readily verified that P(X → Y) = P{ w ∈ Wε | XIw ⊆ YIw } = ∑{ P{ w } | w ∈
Wε and XIw ⊆ YIw }. An implication X → Y almost certainly holds if P(X → Y) = 1,
possibly holds if P(X → Y) &gt; 0, and is impossible if P(X → Y) = 0. If X → Y
certainly holds, then it is almost certain, and if X → Y is refuted, then it is impossible.
3</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Probabilistic Implicational Bases</title>
      <p>
        At first we introduce the notion of a probabilistic implicational base. Then we will
develop and prove a construction for such bases w.r.t. probabilistic formal contexts. If
the underlying context is finite, then the base is computable. The reader should be aware
of the standard notions of formal concept analysis in [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]. Recall that an implication
follows from an implication set if, and only if, it can be syntactically deduced using the
so-called Armstrong rules as follows: 1. From X ⊇ Y infer X → Y. 2. From X → Y
and Y → Z infer X → Z. 3. From X1 → Y1 and X2 → Y2 infer X1 ∪ X2 → Y1 ∪ Y2.
Definition 3 (Probabilistic Implicational Base). Let K = (G, M, W, I, P) be a
probabilistic formal context, and p ∈ [
        <xref ref-type="bibr" rid="ref1">0, 1</xref>
        ] a threshold. A probabilistic implicational base for K
and p is an implication set B over M that satisfies the following properties:
      </p>
      <sec id="sec-3-1">
        <title>1. B is sound for K and p, i.e., P(X → Y) ≥ p holds for all implications X → Y ∈ B, and</title>
      </sec>
      <sec id="sec-3-2">
        <title>2. B is complete for K and p, i.e., if P(X → Y) ≥ p, then X → Y follows from B.</title>
        <p>A probabilistic implicational base is irredundant if none of its implications follows from the
others, and is minimal if it has minimal cardinality among all bases for K and p.</p>
        <p>It is readily verified that the above definition is a straight-forward generalization of
implicational bases as defined in [7, Definition 37], in particular formal contexts coincide
with probabilistic formal contexts having only one possible world, and implications
holding in the formal context coincide with implications having probability 1.</p>
        <p>We now define a transformation from probabilistic formal contexts to formal contexts.
It allows to decide whether an implication (almost) certainly holds, and furthermore it
can be utilized to construct an implicational base for the (almost) certain implications.
Definition 4 (Scaling). Let K be a probabilistic formal context. The certain scaling of K is
the formal context K× := (G × W, M, I×) where ((g, w), m) ∈ I× iff (g, m, w) ∈ I, and
the almost certain scaling of K is the subcontext Kε× := (G × Wε, M, Iε×) of K×.
Lemma 5. Let K = (G, M, W, I, P) be a probabilistic formal context, and let X → Y be a
formal implication. Then the following statements are satisfied:</p>
      </sec>
      <sec id="sec-3-3">
        <title>1. X → Y certainly holds in K if, and only if, X → Y holds in K×.</title>
      </sec>
      <sec id="sec-3-4">
        <title>2. X → Y almost certainly holds in K if, and only if, X → Y holds in Kε×.</title>
        <p>Proof. It is readily verified that the following equivalences hold:</p>
        <p>P(X → Y) = 1 ⇔ ∀w ∈ W : XIw ⊆ YIw
⇔ XI× = ]</p>
        <p>w∈W
⇔ K× |= X → Y.</p>
        <p>XIw × { w } ⊆ ]
w∈W</p>
        <p>YIw × { w } = YI×
The second statement can be proven analogously.</p>
        <p>
          Recall the notion of a pseudo-intent [
          <xref ref-type="bibr" rid="ref6 ref7 ref8">6–8</xref>
          ]: An attribute set P ⊆ M of a formal context
(G, M, I) is a pseudo-intent if P 6= PII, and QII ⊆ P holds for all pseudo-intents Q ( P.
Furthermore, it is well-known that the canonical implicational base of a formal context
(G, M, I) consists of all implications P → PII where P is a pseudo-intent, cf. [
          <xref ref-type="bibr" rid="ref6 ref7 ref8">6–8</xref>
          ].
Consequently, the next corollary is an immediate consequence of Lemma 5.
Corollary 6. Let K be a probabilistic formal context. Then the following statements hold:
1. An implicational base for K× is an implicational base for the certain implications of K,
in particular this holds for the following implication set:
tu
        </p>
        <p>BK := { P → PI× I× | P ∈ PsInt(K×) }.</p>
      </sec>
      <sec id="sec-3-5">
        <title>2. An implicational base for K× w.r.t. the background knowledge BK is an implicational</title>
        <p>ε
base for the almost certain implications of K, in particular this holds for the following
implication set:</p>
        <p>BK,1 := BK ∪ { P → PIε× Iε× | P ∈ PsInt(Kε×, BK) }.</p>
        <p>Lemma 7. Let K = (G, M, W, I, P) be a probabilistic formal context. Then the following
statements are satisfied:</p>
      </sec>
      <sec id="sec-3-6">
        <title>1. Y ⊆ X implies that X → Y certainly holds in K.</title>
        <p>2. X1 ⊆ X2 and Y1 ⊇ Y2 imply P(X1 → Y1) ≤ P(X2 → Y2).
3. X0 ⊆ X1 ⊆ . . . ⊆ Xn implies P(X0 → Xn) ≤ Vin=1 P(Xi−1 → Xi).
Proof.</p>
        <p>1. If Y ⊆ X, then XIw ⊆ YIw follows for all worlds w ∈ W.
2. Assume X1 ⊆ X2 and Y2 ⊆ Y1. Then X1Iw ⊇ X2Iw and Y2Iw ⊇ Y1Iw follow for all
worlds w ∈ W. Consider a world w ∈ W where X1Iw ⊆ Y1Iw . Of course, we may
conclude that X2Iw ⊆ Y2Iw . As a consequence we get P(X1 → Y1) ≤ P(X2 → Y2).
3. We prove the third claim by induction on n. For n = 0 there is nothing to show,
and the case n = 1 is trivial. Hence, consider n = 2 for the induction base,
and let X0 ⊆ X1 ⊆ X2. Then we have that X0Iw ⊇ X1Iw ⊇ X2Iw is satisfied in
all worlds w ∈ W. Now consider a world w ∈ W where X0Iw ⊆ X2Iw is true.
Of course, it then follows that X0Iw ⊆ X1Iw ⊆ X2Iw . Consequently, we conclude
P(X0 → X2) ≤ P(X0 → X1) and P(X0 → X2) ≤ P(X1 → X2).</p>
        <p>For the induction step let n &gt; 2. The induction hypothesis yields that</p>
        <p>P(X0 → Xn−1) ≤ ^in=−11 P(Xi−1 → Xi).</p>
        <p>Of course, it also holds that X0 ⊆ Xn−1 ⊆ Xn, and it follows by induction
hypothesis and the previous inequality that</p>
        <p>P(X0 → Xn) ≤ P(X0 → Xn−1) ∧ P(Xn−1 → Xn) ≤
^n
i=1 P(Xi−1 → Xi).</p>
        <p>tu
Lemma 8. Let K = (G, M, W, I, P) be a probabilistic formal context. Then for all
implications X → Y the following equalities are valid:</p>
        <p>P(X → Y) = P(XI× I× → YI× I× ) = P(XIε× Iε× → YIε× Iε× ).</p>
        <p>Proof. Let X → Y be an implication. Then for all worlds w ∈ W it holds that
g ∈ XIw ⇔ ∀m ∈ X : (g, m, w) ∈ I ⇔ ∀m ∈ X : ((g, w), m) ∈ I× ⇔ (g, w) ∈ XI× ,
and we conclude that XIw = π1(XI× ∩ (G × { w })). Furthermore, we then infer
XIw = XI× I× Iw , and thus the following equations hold:</p>
        <p>P(X → Y) = P{ w ∈ W | XIw ⊆ YIw }</p>
        <p>= P{ w ∈ W | XI× I× Iw ⊆ YI× I× Iw } = P(XI× I× → YI× I× ).</p>
        <sec id="sec-3-6-1">
          <title>X → Y follows from B ∪ { XIε× Iε× → YIε× Iε× }.</title>
          <p>it may be concluded that P(X → Y) = P(XIε× Iε× → YIε× Iε× ).</p>
          <p>In particular, for all possible worlds w ∈ Wε it holds that g ∈ XIw ⇔ (g, w) ∈ XIε× , and
thus XIw = π1(XIε× ∩ (G × { w })) and XIw = XIε× Iε× Iw are satisfied. Consequently,
Lemma 9. Let K be a probabilistic formal context. Then the following statements hold:</p>
        </sec>
      </sec>
      <sec id="sec-3-7">
        <title>1. If B is an implicational base for the certain implications of K, then the implication X → Y</title>
        <p>follows from B ∪ { XI× I× → YI× I× }.</p>
      </sec>
      <sec id="sec-3-8">
        <title>2. If B is an implicational base for the almost certain implications of K, then the implication</title>
        <p>Proof. Of course, the implication X → XI× I× holds in K×, i.e., certainly holds in K
by Lemma 5, and hence follows from B. Thus, the implication X → YI× I× is entailed
by B ∪ { XI× I× → YI× I× }, and because of Y ⊆ YI× I× the claim follows.</p>
        <p>The second statement follows analogously.</p>
        <p>Lemma 10. Let K be a probabilistic formal context. Then the following statements hold:
tu
tu</p>
        <p>YIε× Iε×
Proof. First note that (XIε× Iε× ∪ YIε× Iε× )Iε× Iε× = (X ∪ Y)Iε× Iε× . As YIε× Iε× is a subset of
(XIε× Iε× ∪ YIε× Iε× )Iε× Iε× , the implication (X ∪ Y)Iε× Iε× → YIε× Iε× certainly holds in K, cf.
Statement 1 in Lemma 7.</p>
        <p>Furthermore, we have that XIw ⊆ YIw if, and only if, XIw ⊆ XIw ∩ YIw = (X ∪ Y)Iw .
Hence, the implication X → Y has the same probability as X → X ∪ Y. Consequently,
we may conclude by means of Lemma 7 that
P(XIε× Iε× → YIε× Iε× ) = P(X → Y) = P(X → X ∪ Y) = P(XIε× Iε× → (X ∪ Y)Iε× Iε× ).
Obviously, { XIε× Iε× → (X ∪ Y)Iε× Iε× , (X ∪ Y)Iε× Iε× → YIε× Iε× } entails XIε× Iε× → YIε× Iε×. tu
Lemma 11. Let K be a probabilistic formal context, and X, Y be intents of Kε× such that</p>
      </sec>
      <sec id="sec-3-9">
        <title>X ⊆ Y and P(X → Y) ≥ p. Then the following statements are true:</title>
        <p>1. There is a chain of neighboring intents X = X0 ≺ X1 ≺ X2 ≺ . . . ≺ Xn = Y in Kε×,
2. P(Xi−1 → Xi) ≥ p for all i ∈ { 1, . . . , n }, and
3. X → Y is entailed by { Xi−1 → Xi | i ∈ { 1, . . . , n } }.</p>
        <p>Proof. The existence of a chain X = X0 ≺ X1 ≺ X2 ≺ . . . ≺ Xn−1 ≺ Xn = Y of
neighboring intents between X and Y in Kε× follows from X ⊆ Y.</p>
        <p>From Statement 3 in Lemma 7 it follows that all implications Xi−1 → Xi have a
probability of at least p in K. It is trivial that they entail X → Y. tu
Theorem 12 (Probabilistic Implicational Base). Let K be a probabilistic formal context,
and p ∈ [0, 1) a probability threshold. Then the following implication set is a probabilistic
implicational base for K and p:</p>
        <p>BK,p := BK,1 ∪ { X → Y | X, Y ∈ Int(Kε×) and X ≺ Y and P(X → Y) ≥ p }.
Proof. All implications in BK,1 hold almost certainly in K, and thus have probability 1.
By construction, all other implications X → Y in the second subset have a probability
≥ p. Hence, Statement 1 in Definition 3 is satisfied.</p>
        <p>Now consider an implication X → Y over M such that P(X → Y) ≥ p. We have
to prove Statement 2 of Definition 3, i.e., that X → Y is entailed by BK,p.</p>
        <p>Lemma 8 yields that both implications X → Y and XIε× Iε× → YIε× Iε× have the same
probability. Lemma 9 states that X → Y follows from BK,1 ∪ { XIε× Iε×
According to Lemma 10, the implication XIε× Iε×
→ YIε× Iε× follows from { XIε× Iε× →
→ YIε× Iε× }.
(X ∪ Y)Iε× Iε× , (X ∪ Y)Iε× Iε× → YIε× Iε× }. Furthermore, it holds that</p>
        <p>P(XIε× Iε× → (X ∪ Y)Iε× Iε× ) = P(XIε× Iε× → YIε× Iε× ) = P(X → Y) ≥ p,
starting at XIε× Iε× and ending at (X ∪ Y)Iε× Iε× , i.e.,
and the second implication (X ∪ Y)Iε× Iε× → YIε× Iε× certainly holds, i.e., follows from
BK,1. Finally, Lemma 11 states that there is a chain of neighboring intents of K×
ε</p>
        <p>XIε× Iε× = X0Iε× Iε× ≺ X1Iε× Iε× ≺ X2Iε× Iε× ≺ . . . ≺ XnIε× Iε× = (X ∪ Y)Iε× Iε× ,
such that all implications XiI−ε×1Iε× → XiIε× Iε× have a probability ≥ p, and are thus
contained in BK,p. Hence, BK,p entails the implication X → Y.</p>
        <p>BK,ε := BK,1 ∪ { X → Y | X, Y ∈ Int(Kε×) and X ≺ Y and P(X → Y) &gt; 0 }.</p>
        <p>However, it is not possible to show irredundancy or minimality for the base of
probabilistic implications given above in Theorem 12. Consider the probabilistic formal
context K = ({ g1, g2 }, { m1, m2 }, { w1, w2 }, I, { { w1 } 7→ 2 , { w2 } 7→ 21 }) whose
1
incidence relation I is defined as follows:
The only pseudo-intent of K× is ∅, and the concept lattice of K× is shown above.
Hence, we have the following probabilistic implicational base for p = 12 :</p>
        <p>BK, 12 = { ∅ → { m2 }, { m2 } → { m1, m2 } }.</p>
        <p>However, the set B := { ∅ → { m1, m2 } } is also a probabilistic implicational base for
K and 21 with less elements.</p>
        <p>In order to compute a minimal base for the implications holding in a probabilistic
formal context with a probability ≥ p, one can for example determine the above given
probabilistic base, and minimize it by means of constructing the Duquenne-Guigues
base of it. This either requires the transformation of the implication set into a formal
context that has this implication set as an implicational base, or directly compute all
pseudoclosures of the closure operator induced by the (probabilistic) implicational base.</p>
        <p>Recall that the confidence of an implication X → Y in a formal context (G, M, I)
is defined as conf(X → Y) := (X ∪ Y)I / XI , cf. [12]. In general, there is no
correspondence between the probability of an implication in K and its confidence in K×
or Kε×. To prove this we will provide two counterexamples. As first counterexample
we consider the context K above. It is readily verified that P({ m2 } → { m1 }) = 21
and conf({ m2 } → { m1 }) = 34 , i.e., the confidence is greater than the probability.
Furthermore, consider the following modification of K as second counterexample:
Then we have that P({ m2 } → { m1 }) = 12 and conf({ m2 } → { m1 }) = 31 , i.e., the
confidence is smaller than the probability.</p>
        <p>
          The Description Logic EL⊥ and Probabilistic Interpretations
This section gives a brief overview on the light-weight description logic EL⊥ [
          <xref ref-type="bibr" rid="ref1">1</xref>
          ]. First,
assume that (NC, NR) is a signature, i.e., NC is a set of concept names, and NR is a
set of role names, respectively. Then EL⊥-concept descriptions C over (NC, NR) may be
constructed according to the following inductive rule (where A ∈ NC and r ∈ NR):
        </p>
        <p>C ::= ⊥ | &gt; | A | C u C | ∃ r. C.</p>
        <p>We shall denote the set of all EL⊥-concept descriptions over (NC, NR) by EL⊥(NC, NR).
Second, the semantics of EL⊥-concept descriptions is defined by means of
interpretations: An interpretation is a tuple I = (ΔI , ·I ) that consists of a set ΔI , called domain,
and an extension function ·I : NC ∪ NR → 2ΔI ∪ 2ΔI ×ΔI that maps concept names
A ∈ NC to subsets AI ⊆ ΔI and role names r ∈ NR to binary relations rI ⊆ ΔI × ΔI .
The extension function is extended to all EL⊥-concept descriptions as follows:
⊥I := ∅,
&gt;I := ΔI ,
(C u D)I := CI ∩ DI ,
(∃ r. C)I := { d ∈ ΔI | ∃e ∈ ΔI : (d, e) ∈ rI and e ∈ CI }.</p>
      </sec>
      <sec id="sec-3-10">
        <title>A general concept inclusion (GCI) in EL⊥ is of the form C v D where C and D are EL⊥</title>
        <p>concept descriptions. It holds in an interpretation I if CI ⊆ DI is satisfied, and we then
also write I |= C v D, and say that I is a model of C v D. Furthermore, C is subsumed
by D if C v D holds in all interpretations, and we shall denote this by C v D, too. A
TBox is a set of GCIs, and a model of a TBox is a model of all its GCIs. A TBox T entails
a GCI C v D, denoted by T |= C v D, if every model of T is a model of C v D.</p>
        <p>To introduce probability into the description logic EL⊥, we now present the notion
of a probabilistic interpretation from [11]. It is simply a family of interpretations over
the same domain and the same signature, indexed by a set of worlds that is equipped
with a probability measure.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Definition 14 (Probabilistic Interpretation, [11]). Let (NC, NR) be a signature. A prob</title>
      <p>abilistic interpretation I is a tuple (ΔI , (·Iw )w∈W, W, P) consisting of a set ΔI , called
domain, a countable set W of worlds, a probability measure P on W, and an extension
function ·Iw for each world w ∈ W, i.e., (ΔI , ·Iw ) is an interpretation for each w ∈ W.</p>
      <sec id="sec-4-1">
        <title>For a general concept inclusion C v D its probability in I is defined as follows:</title>
        <p>P(C v D) := P{ w ∈ W | CIw ⊆ DIw }.</p>
        <p>Furthermore, for a GCI C v D we define the following properties (as for probabilistic formal
contexts): 1. C v D holds in world w if CIw ⊆ DIw . 2. C v D certainly holds in I if it
holds in all worlds. 3. C v D almost certainly holds in I if it holds in all possible worlds.</p>
      </sec>
      <sec id="sec-4-2">
        <title>4. C v D possibly holds in I if it holds in a possible world. 5. C v D is impossible in</title>
      </sec>
      <sec id="sec-4-3">
        <title>I if it does not hold in any possible world. 6. C v D is refuted by I if it does not hold in</title>
        <p>any world.</p>
        <p>It is readily verified that P(C v D) = P{ w ∈ Wε | CIw ⊆ DIw } = ∑{ P{ w } | w ∈
Wε and CIw ⊆ DIw } for all general concept inclusions C v D.
5</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>Probabilistic Bases of GCIs</title>
      <p>In the following we construct from a probabilistic interpretation I a base of GCIs that
entails all GCIs with a probability greater than a given threshold p w.r.t. I.</p>
      <sec id="sec-5-1">
        <title>Definition 15 (Probabilistic Base). Let I be a probabilistic interpretation, and p ∈ [0, 1]</title>
        <p>a threshold. A probabilistic base of GCIs for I and p is a TBox B that satisfies the following
conditions:</p>
      </sec>
      <sec id="sec-5-2">
        <title>1. B is sound for I and p, i.e., P(C v D) ≥ p for all GCIs C v D ∈ B, and 2. B is complete for I and p, i.e., if P(C v D) ≥ p, then B |= C v D.</title>
      </sec>
      <sec id="sec-5-3">
        <title>A probabilistic base B is irredundant if none of its GCIs follows from the others, and is minimal if it has minimal cardinality among all probabilistic bases for I and p.</title>
        <p>For a probabilistic interpretation I we define its certain scaling as the disjoint union
of all interpretations Iw with w ∈ W, i.e., as the interpretation I × := (ΔI × W, ·I× )
whose extension mapping is given as follows:</p>
        <p>AI× := { (d, w) | d ∈ AIw }
rI× := { ((d, w), (e, w)) | (d, e) ∈ rIw }
(A ∈ NC),
(r ∈ NR).</p>
        <p>Furthermore, the almost certain scaling Iε× of I is the disjoint union of all interpretations
Iw where w ∈ Wε is a possible world. Analogously to Lemma 5, a GCI C v D certainly
holds in I iff it holds in I ×, and almost certainly holds in I iff it holds in Iε×.</p>
        <p>
          In [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ] the so-called model-based most-specific concept descriptions (mmscs) have been
defined w.r.t. greatest fixpoint semantics as follows: Let J be an interpretation, and X ⊆
ΔJ . Then a concept description C is a mmsc of X in J , if X ⊆ CJ is satisfied, and C v
D for all concept descriptions D with X ⊆ DJ . It is easy to see that all mmscs of X are
unique up to equivalence, and hence we denote the mmsc of X in J by XJ . Please note
that there is also a role-depth bounded variant w.r.t. descriptive semantics given in [
          <xref ref-type="bibr" rid="ref3">3</xref>
          ].
Lemma 16. Let I be a probabilistic interpretation. Then the following statements hold:
1. CIw × { w } = CI× ∩ (ΔI × { w }) for all concept descriptions C and worlds w ∈ W.
2. CIw × { w } = CIε× ∩ (ΔI × { w }) for all concept descriptions C and possible worlds
w ∈ Wε.
3. P(C v D) = P(CI×I× v DI×I× ) = P(CIε×Iε× v DIε×Iε× ) for all GCIs C v D.
Proof. 1. We prove the claim by structural induction on C. By definition, the statement
holds for ⊥, &gt;, and all concept names A ∈ NC. Consider a conjunction C u D, then
(C u D)Iw × { w } = (CIw ∩ DIw ) × { w }
= CIw × { w } ∩ DIw × { w }
I=.H.CI× ∩ DI× ∩ (ΔI × { w })
= (C u D)I× ∩ (ΔI × { w }).
For an existential restriction ∃ r. C the following equalities hold:
        </p>
        <p>(∃ r. C)Iw × { w }
= { d ∈ ΔI | ∃e ∈ ΔI : (d, e) ∈ rIw and e ∈ CIw } × { w }
= { (d, w) | ∃(e, w) : ((d, w), (e, w)) ∈ rI× and (e, w) ∈ CIw × { w } }
I=.H.{ (d, w) | ∃(e, w) : ((d, w), (e, w)) ∈ rI× and (e, w) ∈ CI× }
= (∃ r. C)I× ∩ (ΔI × { w }).
2. analogously.
3. Using the first statement we may conclude that the following equalities hold:</p>
        <p>P(C v D)
= P{ w ∈ W | CIw × { w } ⊆ DIw × { w } }
= P{ w ∈ W | CI× ∩ (ΔI × { w }) ⊆ DI× ∩ (ΔI × { w }) }
= P{ w ∈ W | CI×I×I× ∩ (ΔI × { w }) ⊆ DI×I×I× ∩ (ΔI × { w }) }
= P{ w ∈ W | CI×I×Iw × { w } ⊆ DI×I×Iw × { w } }
= P(CI×I× v DI×I× ).</p>
        <p>The second equality follows analogously.</p>
        <p>For a probabilistic interpretation I = (ΔI , ·I , W, P) and a set M of EL⊥-concept
descriptions we define their induced context as the probabilistic formal context KI,M :=
(ΔI , M, W, I, P) where (d, C, w) ∈ I iff d ∈ CIw .</p>
      </sec>
      <sec id="sec-5-4">
        <title>Lemma 17. Let I be a probabilistic interpretation, M a set of EL⊥-concept descriptions, and</title>
        <p>the probability of the GCI d X v d Y in I, i.e., it hol→ds tYhaitnPth(eXin→ducYed)c=ontPex(tdKXI,Mv edquYal)s.</p>
      </sec>
      <sec id="sec-5-5">
        <title>X, Y ⊆ M. Then the probability of the implication X</title>
        <p>Proof. The following equivalences are satisfied for all Z ⊆ M and worlds w ∈ W:
d ∈ ZIw ⇔ ∀C ∈ Z : (d, C, w) ∈ I ⇔ ∀C ∈ Z : d ∈ CIw ⇔ d ∈ (l Z)Iw .
Now consider two subsets X, Y ⊆ M, then it holds that</p>
        <p>P(X → Y) = P{ w ∈ W | XIw ⊆ YIw }</p>
        <p>= P{ w ∈ W | (l X)Iw ⊆ (l Y)Iw } = P(l X v l Y).</p>
        <p>
          Analogously to [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ], the context KI is defined as KI,MI with the following attributes:
M := { ⊥ } ∪ NC ∪ { ∃ r. XIε× | ∅ 6= X ⊆ ΔI × Wε }.
        </p>
        <p>I</p>
        <p>For an implication set B over a set M of EL⊥-concept descriptions we define its
induced TBox by d B := { d X v d Y | X → Y ∈ B }.
tu
tu</p>
        <sec id="sec-5-5-1">
          <title>Corollary 18. If B contains an almost certain implicational base for KI , then d B is complete</title>
          <p>for the almost certain GCIs of I.</p>
          <p>Proof. We know that a GCI almost certainly holds in I if, and only if, it holds in
Iε×. Let B0 ⊆ B be an almost certain implicational base for KI , i.e., an implicational
base for (KI )ε× = K × . Then according to Distel in [5, Theorem 5.12] it follows that</p>
          <p>Iε
the TBox d B0 is a base of GCIs for Iε×, i.e., a base for the almost certain GCIs of I.
Consequently, d B is complete for the almost certain GCIs of I. tu</p>
        </sec>
      </sec>
      <sec id="sec-5-6">
        <title>Theorem 19. Let I be a probabilistic interpretation, and p ∈ [0, 1] a threshold. If B is a</title>
        <p>probabilistic implicational base for KI and p that contains an almost certain implicational base
for KI , then d B is a probabilistic base of GCIs for I and p.</p>
        <p>Proof. Consider a GCI d X v d Y ∈ d B. Then Lemma 17 yields that the implication
X → Y and the GCI d X v d Y have the same probability. Since B is a probabilistic
implicational base for KI and p, we conclude that P(d X v d Y) ≥ p is satisfied.</p>
        <p>Assume that C v D is an arbitrary GCI with probability ≥ p. We have to show
that d B entails C v D. Let J be an arbitrary model of d B. Consider an
implication X → Y ∈ B, then d X v d Y ∈ d B holds, and hence it follows that
(d X)J ⊆ (d Y)J . Consequently, the implication X → Y holds in the induced
context KJ ,M . (We here mean the non-probabilistic formal context that is induced by</p>
        <p>
          I
a non-probabilistic interpretation, cf. [
          <xref ref-type="bibr" rid="ref2 ref3 ref5">2, 3, 5</xref>
          ].)
        </p>
        <p>
          Furthermore, since all model-based most-specific concept descriptions of Iε× are
expressible in terms of M , we have that E ≡ d πM (E) holds for all mmscs E of Iε×,
cf. [
          <xref ref-type="bibr" rid="ref2 ref3 ref5">2, 3, 5</xref>
          ]. Hence, we maIy conclude that I
        </p>
        <p>P(C v D) = P(CIε×Iε× v DIε×Iε× )
= P(l πM (CIε×Iε× ) v</p>
        <p>I
l πM (DIε×Iε× ))</p>
        <p>I
= P(πM (CIε×Iε× ) → πM (DIε×Iε× )).</p>
        <p>I I
Consequently, B entails the implication πM (CIε×Iε× ) → πM (DIε×Iε× ), hence it holds</p>
        <p>I I
in KJ ,MI , and furthermore the GCI CIε×Iε× v DIε×Iε× holds in J . As J is an arbitrary
interpretation, d B entails CIε×Iε× v DIε×Iε× .</p>
        <p>Corollary 18 yields that d B is complete for the almost certain GCIs of I. In
particular, the GCI C v CIε×Iε× almost certainly holds in I, and hence follows from d B.
We conclude that d B |= C v DIε×Iε× . Of course, the GCI DIε×Iε× v D holds in all
interpretations. Finally, we conclude that d B entails C v D. tu</p>
      </sec>
      <sec id="sec-5-7">
        <title>Corollary 20. Let I be a probabilistic interpretation, and p ∈ [0, 1] a threshold. Then</title>
        <p>d BKI ,p is a probabilistic base of GCIs for I and p where BKI ,p is defined as in Theorem 12.
6</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>Conclusion</title>
      <p>We have introduced the notion of a probabilistic formal context as a triadic context
whose third dimension is a set of worlds equipped with a probability measure. Then
the probability of implications in such probabilistic formal contexts was defined, and a
construction of a base of implications whose probability exceeds a given threshold has
been proposed, and its correctness has been verified. Furthermore, the results have been
applied to the light-weight description logic EL⊥ with probabilistic interpretations,
and so we formulated a method for the computation of a base of general concept
inclusions whose probability satisfies a given lower threshold.</p>
      <p>
        For finite input data-sets all of the provided constructions are computable. In
particular, [
        <xref ref-type="bibr" rid="ref3 ref5">3, 5</xref>
        ] provide methods for the computation of model-based most-specific concept
descriptions, and the algorithms in [
        <xref ref-type="bibr" rid="ref6">6, 10</xref>
        ] can be utilized to compute concept lattices
and canonical implicational bases (or bases of GCIs, respectively).
      </p>
      <p>The author thanks Sebastian Rudolph for proof reading and a fruitful discussion,
and the anonymous reviewers for their constructive comments.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>Franz</given-names>
            <surname>Baader</surname>
          </string-name>
          et al., eds.
          <source>The Description Logic Handbook: Theory</source>
          , Implementation, and
          <string-name>
            <surname>Applications</surname>
          </string-name>
          . New York, NY, USA: Cambridge University Press,
          <year>2003</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>Daniel</given-names>
            <surname>Borchmann</surname>
          </string-name>
          . “
          <article-title>Learning Terminological Knowledge with High Confidence from Erroneous Data”</article-title>
          .
          <source>PhD thesis</source>
          . TU Dresden, Germany,
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>Daniel</given-names>
            <surname>Borchmann</surname>
          </string-name>
          , Felix Distel, and
          <string-name>
            <given-names>Francesco</given-names>
            <surname>Kriegel</surname>
          </string-name>
          .
          <source>Axiomatization of General Concept Inclusions from Finite Interpretations. LTCS-Report 15-13</source>
          .
          <article-title>Chair for Automata Theory, Institute for Theoretical Computer Science</article-title>
          , TU Dresden, Germany,
          <year>2015</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <surname>Alexander</surname>
            <given-names>V.</given-names>
          </string-name>
          <string-name>
            <surname>Demin</surname>
          </string-name>
          ,
          <string-name>
            <surname>Denis K. Ponomaryov</surname>
          </string-name>
          , and Evgenii Vityaev. “
          <article-title>Probabilistic Concepts in Formal Contexts”</article-title>
          .
          <source>In: Perspectives of Systems Informatics - 8th International Andrei Ershov Memorial Conference, PSI</source>
          <year>2011</year>
          , Novosibirsk, Russia, June 27-July 1,
          <year>2011</year>
          , Revised Selected Papers. Ed. by
          <string-name>
            <surname>Edmund M. Clarke</surname>
            , Irina Virbitskaite, and
            <given-names>Andrei</given-names>
          </string-name>
          <string-name>
            <surname>Voronkov</surname>
          </string-name>
          . Vol.
          <volume>7162</volume>
          . Lecture Notes in Computer Science. Springer,
          <year>2011</year>
          , pp.
          <fpage>394</fpage>
          -
          <lpage>410</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>Felix</given-names>
            <surname>Distel</surname>
          </string-name>
          .
          <article-title>“Learning Description Logic Knowledge Bases from Data using Methods from Formal Concept Analysis”</article-title>
          .
          <source>PhD thesis</source>
          . TU Dresden, Germany,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>Bernhard</given-names>
            <surname>Ganter</surname>
          </string-name>
          . “
          <article-title>Two Basic Algorithms in Concept Analysis”</article-title>
          .
          <source>In: Formal Concept Analysis, 8th International Conference, ICFCA</source>
          <year>2010</year>
          , Agadir, Morocco, March
          <volume>15</volume>
          -18,
          <year>2010</year>
          . Proceedings. Ed.
          <article-title>by Léonard Kwuida and Baris Sertkaya</article-title>
          . Vol.
          <volume>5986</volume>
          . Lecture Notes in Computer Science. Springer,
          <year>2010</year>
          , pp.
          <fpage>312</fpage>
          -
          <lpage>340</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>Bernhard</given-names>
            <surname>Ganter</surname>
          </string-name>
          and
          <string-name>
            <given-names>Rudolf</given-names>
            <surname>Wille</surname>
          </string-name>
          .
          <source>Formal Concept Analysis - Mathematical Foundations</source>
          . Springer,
          <year>1999</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <surname>Jean-Luc Guigues</surname>
            and
            <given-names>Vincent</given-names>
          </string-name>
          <string-name>
            <surname>Duquenne</surname>
          </string-name>
          . “
          <article-title>Famille minimale d'implications informatives résultant d'un tableau de données binaires”</article-title>
          .
          <source>In: Mathématiques et Sciences Humaines</source>
          <volume>95</volume>
          (
          <year>1986</year>
          ), pp.
          <fpage>5</fpage>
          -
          <lpage>18</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>Francesco</given-names>
            <surname>Kriegel</surname>
          </string-name>
          . “
          <article-title>Axiomatization of General Concept Inclusions in Probabilistic Description Logics”</article-title>
          .
          <source>In: Proceedings of the 38th German Conference on Artificial Intelligence, KI</source>
          <year>2015</year>
          , Dresden, Germany,
          <source>September 21-25</source>
          ,
          <year>2015</year>
          . Vol.
          <volume>9324</volume>
          .
          <source>Lecture Notes in Artificial Intelligence</source>
          . Springer Verlag,
          <year>2015</year>
          . Francesco Kriegel.
          <article-title>NextClosures - Parallel Exploration of Constrained Closure Operators</article-title>
          .
          <source>LTCSReport</source>
          <volume>15</volume>
          -
          <fpage>01</fpage>
          .
          <article-title>Chair for Automata Theory, Institute for Theoretical Computer Science</article-title>
          , TU Dresden, Germany,
          <year>2015</year>
          . Carsten Lutz and
          <string-name>
            <given-names>Lutz</given-names>
            <surname>Schröder</surname>
          </string-name>
          . “
          <article-title>Probabilistic Description Logics for Subjective Uncertainty”</article-title>
          .
          <source>In: Principles of Knowledge Representation and Reasoning: Proceedings of the Twelfth International Conference, KR 2010</source>
          , Toronto, Ontario, Canada, May 9-
          <issue>13</issue>
          ,
          <year>2010</year>
          . Ed. by Fangzhen Lin, Ulrike
          <string-name>
            <surname>Sattler</surname>
            , and
            <given-names>Miroslaw</given-names>
          </string-name>
          <string-name>
            <surname>Truszczynski</surname>
          </string-name>
          . AAAI Press,
          <year>2010</year>
          .
          <string-name>
            <given-names>Michael</given-names>
            <surname>Luxenburger</surname>
          </string-name>
          . “Implikationen,
          <article-title>Abhängigkeiten und Galois Abbildungen”</article-title>
          .
          <source>PhD thesis</source>
          . TH Darmstadt, Germany,
          <year>1993</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>