=Paper= {{Paper |id=Vol-2972/paper3 |storemode=property |title=FCA Went (Multi-)Relational, But Does It Make Any Difference? |pdfUrl=https://ceur-ws.org/Vol-2972/paper3.pdf |volume=Vol-2972 |authors=Mickaël Wajnberg,Petko Valtchev,Mario Lezoche,Alexandre Blondin-Massé,Hervé Panetto |dblpUrl=https://dblp.org/rec/conf/ijcai/WajnbergVLMP21 }} ==FCA Went (Multi-)Relational, But Does It Make Any Difference?== https://ceur-ws.org/Vol-2972/paper3.pdf
     FCA Went (Multi-)Relational, But Does It
             Make Any Difference?

          Mickaël Wajnberg1 , Petko Valtchev1 , Mario Lezoche2 , Alexandre
                      Blondin-Massé1 , and Hervé Panetto2
      1
              Université du Québec À Montréal, Canada {name.firstname}@uqam.ca
          2
               Université de Lorraine, France {firstname.name}@univ-lorraine.fr



          Abstract. Relational Concept Analysis (RCA) was designed as an ex-
          tension of Formal Concept Analysis (FCA) to multi-relational datasets,
          such as the ones drawn from Linked Open Data (LOD) by the type-wise
          grouping of the resource into data tables. RCA has been successfully
          applied to practical problems of AI such as knowledge elicitation, knowl-
          edge discovery from data and knowledge structuring. A crucial question,
          yet to be answered in a rigorous manner, is to what extent RCA is a true
          extension of FCA, i.e. reveals concepts that are beyond the reach of core
          FCA even using a suitable encoding of the original data. We show here
          that the extension is effective: RCA retrieves all concepts found by FCA
          as well as many further ones.

          Keywords: Multi-relational Data · Formal Concept Analysis · RDF
          · Propositionalization.


1    Introduction
FCA provides the mathematical framework for several Knowledge Discovery in
Databases (KDD) tasks whenever the data is purely, or at least predominantly,
of categorical nature. Indeed, FCA-based association discovery and conceptual
clustering have been applied to knowledge base structuring, ontology learning,
anomaly detection, observation classification, etc. Most real datasets, though,
stray from being purely categorical. FCA thus provides a set of scaling operators
to deal with numerical and otherwise ordered scales. In AI, the majority of
interesting data, such as those compatible with the LOD format, have relational
structure. They can be represented either as graphs (for instance, named graphs
in RDF) or as sets of relational tables. Approaches have been designed for the
former, emphasizing the intra-data object links, e.g. logical FCA [7] and pattern
structures [8], for graph datasets. For the latter, the focus is on inter-object links,
e.g. in datasets structured as a unique RDF graph. In this second trend, more
akin to power-context families [15] and Graph-FCA [6], we focus on the particular
approach of relational concept analysis (RCA). It has already been successfully
applied to a wide range of practical problems such as hydroecology [4], industrial
decision making [12] or biology [1,13]. Rather than in a global graph, RCA shapes
the data as a set of ×-tables, complying to the Entity-Relationship framework [2].

Copyright c 2021 for this paper by its authors. Use permitted under Creative
Commons License Attribution 4.0 International (CC BY 4.0).
Part of the tables have the classical objects × properties format (entity types,
FCA contexts) while the remainder represent objects × objects relations.
    A natural question is whether RCA does extend the reach of FCA, knowing
that for single datasets, whatever the level of complexity of the object descrip-
tions (sequences, trees, graphs), the results of an FCA-based processing on those
descriptions can be brought down to FCA on a context made of suitably-chosen
derived attributes. The question is all the more important as prior studies seem
to imply it does not [3] (though, for a reduced version of RCA). We make the
case here for RCA as a true extension of FCA, in the sense that due to its
multi-relational input and fixed point computation, it detects concepts that are
out of reach for FCA while, in turn, retrieving all concepts that FCA is able to
reveal. To that end, we chose some plausible re-encodings of a simple relational
context family (RCF), the hypothesis being that with more complex datasets,
the phenomenon only amplifies.
    The remainder of the paper is as follows: Section 2 provides background on
RCA while Section 3 presents our FCA-vs-RCA comparison. Next, Section 4
discusses the comparison outcome and Section 5 concludes.


2   Background

Formal concept analysis [14] is a mathematical method for eliciting the con-
ceptual structure of “object × attribute” datasets. Data are gathered within a
(formal) context, a triple K = (O, A, I) where O is a set of objects, A is a set of
attributes and I ⊆ O × A is the context incidence relation, where (o, a) ∈ I, also
written oIa, means that the object o bears the attribute a. A context induces
two derivation operators: one mapping objects to attributes, and the reciprocal.
The object derivation 0 maps a subset X of objects to the set of attributes shared
by all members of X, 0 : ℘(O) −→ ℘(A) with 0 : X 7→ {a ∈ A | oIa ∀o ∈ X}.
The dual attribute derivation, also denoted by 0 , works the other way around,
0
  : ℘(A) −→ ℘(O) with 0 : Y 7→ {o ∈ O | oIa ∀a ∈ Y }. Inside a context K, a
(formal) concept is a pair (X, Y ) ⊆ O × A such that X 0 = Y and Y 0 = X. The
sets X and Y are called extent and intent of the concept (X, Y ), respectively.
    FCA extracts conceptual abstractions on objects by factoring out shared at-
tributes. Relational concept analysis [10] extends it by factoring in relational
information, as available in multi-relational datasets [5]. RCA admits multiple
sorts of objects in its input format, each organized as a separate context, plus
a set of binary relations between contexts. The input data structure, called re-
lational context family (RCF), is thus a pair (K, R) where K = {Ki }i=1,...,n
is a set of distinct contexts Ki = (Oi , Ai , Ii ) and R = {rk }k=1,...,m a set
of binary relations rk ⊆ Oi × Oj where the contexts Ki and Kj are the do-
main and range contexts of rk , respectively. Relational tables are also processed
in their own way, as explained below. A cross in the table of relation r for
(domain_objecti , range_objectj ), can be understood as the first order logic
term r(domain_objecti , range_objectj ) being true.
    RCA distills the shared relational information (i.e., inter-object links) using
propositionalization [9]: It integrates new attributes into an extended version
of the initial context, say Kd = (Od , Ad , Id ), to further refine the conceptual
structure of the underlying object set. To increase shareability, rather than the
individual objects from the target (range) context, say Kt , the new attributes
refer to abstractions on them. In its most basic version, RCA exploits the natural
conceptual structure provided by the concepts of each context. Indeed, two links
of relation r : d −→ t departing from o1 and o2 from Od and referring to
two distinct objects ō1 and ō2 from Ot , respectively, are distinct information.
However, replacing ō1 and ō2 with a common abstraction, say {ō1 , ō2 }0 , makes the
new information shareable. Relational scaling follows a well-known schema from
description logics: Given a relation r, for each concept ct from the range context
of r, it produces, for Ad , an attribute q r : cr where q is an operator chosen
before-hand from a set Q. RCA admits, among others, standard description
logics restrictions (Q = {∃, ∀, ∀∃, . . .}), which behold their respective semantics
(see [10] for details and example 1 for illustration).

Example 1. Assume a RCF made of contexts on people and cars, and an owner-
ship relation, or pos(sesses), which are given in Tables 1, 3 and 2, respectively.
The cars lattice is shown in Figure 1. Now, an ∃-scaling of the relation pos using
that lattice will add, for each car concept c, a new attribute ∃pos : c to the
person context, e.g. ∃pos : cars_4 and ∃pos : cars_3 that can be rewritten as
∃pos : (cp) and ∃pos : (el, pw), respectively, using intents as IDs.


                                 pos tw t3 zo f 5
                                 Fa ×
    KP                           La      ×      ×
         Senior
         Adult
         Male
         Female
         I.T.
         Sport




                                 Sh ×           ×
                                 Tr
                               Table 2: Relation pos
   Fa × ×
   La      × ××                  KC el pw cp ch
   Sh      × ×
                                 tw         × ×
   Tr      ×× ××
                                  t3 × ×
 Table 1: Person Context
                                 zo ×       ×
                                 f5      × ×                Fig. 1: Car Lattice
                               Table 3: Car Context

     A scaling step results in the related contexts being extended, which, in turn,
may lead to the emergence of new concepts. Thus, as the set of available ab-
stractions increases, a scaling step with the differential set of concepts would
produce further relational attributes and the whole process would go on cycling.
The resulting iterative context refinement necessarily ends at a fixed point [10],
i.e. a set of lattices whose concepts refer to each other via relational attributes.
3     What can RCA do for AI (that FCA can’t) ?

Below, we examine two encoding strategies that bring a multi-relational dataset
to a mono-relational one, i.e. aggregate several contexts into a single data table,
so that they can be fed to classical FCA.


3.1    Encoding multiple contexts into a single one

Assume a simple RCF made of two contexts and a relation (see Figure 2). We
use this simple case for our reasoning, knowing that in more complex cases, i.e.
three or more contexts and several relations, it can be extended appropriately.
Moreover, while there could be a wide range of concrete encoding disciplines [11],
the principle behind them admits only two basic cases, i.e. entity-centric and
relation-centric. In our FCA/RCA perspective this boils down to which sort of
RCF element, i.e. context or relation, is put center-stage.
                                              A1   A2                 A1 ρ(A2 )
         A1         O2       A2
      O1 I1    O1   r    O2 I2          O./                      O1

        Fig. 2: Fictitious RCF
                                       Fig. 3: Semi-join      Fig. 4: Aggregation


    The first encoding principle we examine below emphasizes the object-to-
object relation as a primary construct and pivotal element of the encoding.
Its member pairs become first-class objects which carry the attributes of both
contexts incident to the relation. Technically speaking, the method is akin to
the (semi-)join operation of relational algebra. The overall encoding schema is
illustrated in Figure 3 whereas Section 3.2 proposes a formal definition thereof.
It also provides a detailed comparison of the results from applying RCA to the
RCF from Figure 2 with those of FCA on the semi-join of the two initial contexts.
    A bit closer to the RCA propositionalization spirit, a second encoding princi-
ple emphasizes the context as a main construct and driver of the encoding: The
domain context of the relation is extended with some additional attributes that
translate the relation while following a technique akin to relational scaling. The
main difference here is that the context is the one-shot context extension. The
procedure whose details are discussed in Section 3.3, is schematically illustrated
in Figure 4. Moreover, the asymmetric encoding of the relation and the one-
shot extension amount to processing the range context as if it were aggregated
into the domain one. Therefore, we termed the overall encoding principle the
aggregation and the resulting context the aggregated one.
    Finally, please notice that in the detailed investigation of each case (see be-
low), our reasoning follows three steps: 1) We pick an arbitrary formal concept
from the FCA output, 2) we show the RCA output comprises a concept with the
same objects, and 3) we establish the link between the intents of both concepts.
3.2   Semi-join in single relation RCF
We consider here the concurrent case where the FCA is applied on a context
encoding the semi-join of this RCF, as presented in Figure 3. This encoding
consists in creating the objects of O./ as the object pairs (o1 , o2 ) where o1 ∈
O1 ∪ {⊥}, o2 ∈ O2 ∪ {⊥}, according to the RCF modeling of O1 , O2 Figure 2.
The ⊥ object is a fictitious empty object with no attributes used to complete
the semi-join. There are three cases to define the elements of O./ :

 – If o1 ∈ O1 , o2 ∈ O2 , then (o1 , o2 ) ∈ O./ if and only if (o1 , o2 ) ∈ r
 – If o1 = ⊥, o2 ∈ O2 , then (o1 , o2 ) ∈ O./ if and only if r−1 (o2 ) = ∅, i.e. if there
   is no x ∈ O1 such that (x, o2 ) ∈ r
 – If o1 ∈ O1 , o2 = ⊥, then (o1 , o2 ) ∈ O./ if and only if r(o1 ) = ∅, i.e. if there
   is no x ∈ O2 such that (o1 , x) ∈ r

Example 2. As an illustration of the above modeling, assume an RCF made of
contexts for people (Table 1) and for cars (Table 3) plus an ownership relation
(possession, Table 2). In the first context, Farley, Lane, Shana, and Trudy are
described by being senior or adult, male or female, working in IT, and practicing
a lot of sports. Cars –Twingo, Tesla 3, Zoe, and Fiat 500– can be electrical,
powerful, compact or (not exclusive) cheap. The corresponding semi-join context
is presented in Table 4.


                           K./                     el pw cp ch
                                 Senior
                                 Adult
                                 Male
                                 Female
                                 I.T.
                                 Sport




                        (Fa,tw) × ×                  × ×
                        (La,t3)    × ××         × ×
                        (La,f 5) × × ×             × ×
                        (Sh,tw) × ×                  × ×
                        (Sh,f 5) × ×               × ×
                         (Tr,⊥)    ×× ××
                         (⊥,zo)                 ×    ×
                   Table 4: Semi-join context of Example 2 RCF


   To avoid ambiguity, we consider the derivations in the K1 and K2 contexts
always denoted x0 , while the derivation in the join context is denoted x∇ (and
the double derivation x∇∇ ).
   We are first interested in describing a formal concept of the joined context.
Let X ⊆ O./ be a set of objects. So, for all (o1 , o2 ) ∈ X we have o1 ∈ O1 ∪ {⊥}
and o2 ∈ O2 ∪ {⊥}. Thus, by definition, C = (X ∇∇ , X ∇ ) is a formal concept.
Let us now take the projections on the first and second elements of the pairs of
X ∇∇ , i.e. π1 = {o1 | ∃o2 , (o1 , o2 ) ∈ X ∇∇ } and π2 = {o2 | ∃o1 , (o1 , o2 ) ∈ X ∇∇ }.
We start by defining X ∇∇ in terms of these projections.
Lemma 1 We have X ∇∇ = (π1 × π2 ) ∩ O./

Proof. Let (u, v) ∈ X ∇∇ . By definition X ∇∇ ⊆ O./ , so (u, v) ∈ O./ . Moreover,
by construction u ∈ π1 and v ∈ π2 so (u, v) ∈ π1 × π2 . Thus, X ∇∇ ⊆ (π1 × π2 ) ∩
O./ .
    Let (u, v) ∈ (π1 × π2 ) ∩ O./ . Since (u, v) ∈ (π1 × π2 ), it exists ũ and ṽ s.t.
(u, ũ) ∈ X ∇∇ and (ṽ, v) ∈ X ∇∇ . But, by construction {(u, ũ), (ṽ, v)}∇ ⊆ u0 ∪ v 0 .
And, since (u, v) ∈ O./ we can write (u, v)∇ = u0 ∪ v 0 . Thus, by derivation
property we have X ∇∇∇ ⊆ {(u, ũ), (ṽ, v)}∇ , by transitivity X ∇∇∇ ⊆ (u, v)∇ .
Thus, by deriving this expression we obtain (u, v)∇∇ ⊆ X ∇∇∇∇ . Finally, as
(u, v) ∈ (u, v)∇∇ and X ∇∇∇∇ = X ∇∇ we have (u, v) ∈ X ∇∇ .                          t
                                                                                     u

   We first study the particular cases containing the object ⊥ by starting with
the case where this element appears in both projections.

Proposition 1 If ⊥ ∈ π1 and ⊥ ∈ π2 then X ∇ = ∅ and X ∇∇ = O./

Proof. Suppose ⊥ ∈ π1 and ⊥ ∈ π2 then by definition of ⊥ we have X ∇ ∩ A1 = ∅
and X ∇ ∩ A2 = ∅ and therefore X ∇ = ∅. By definition of the derivation we have
∅∇ = O./ therefore X ∇∇ = O./ .The second assertion holds by symmetry.        t
                                                                              u

   In the case described by the lemma 1, it is immediate to show that we can
construct (X ∇∇ , X ∇ ). We show that the same is true when only one of the
components π1 or π2 contains ⊥ by first describing X ∇ then X ∇∇ in the lemmas
2 and 3.

Lemma 2 If ⊥ ∈ π1 and ⊥ 6∈ π2 , X ∇ = π20 . If ⊥ ∈ π2 and ⊥ 6∈ π1 , X ∇ = π10 .

Proof. Let us suppose ⊥ ∈ π1 and ⊥ 6∈ π2 . We have a ∈ X ∇ iff X ∇∇ ⊆ a∇ . Yet,
since ⊥ ∈ π1 we have construction X ∇ ∩ A1 = ∅. thus, we have a ∈ A2 .therefore,
we have a ∈ X ∇ iff for all (o1 , o2 ) ∈ X ∇∇ o2 carries the attribute a, i.e. a ∈ π20 .
Since we have a ∈ X ∇ iff a ∈ π20 , we have X ∇ = π20 . We show the second
assertion symmetrically.                                                             t
                                                                                     u

Lemma 3 If ⊥ ∈ π1 and ⊥ 6∈ π2 , X ∇∇ = (O1 ∪ {⊥} × π2 ) ∩ O./ . If ⊥ ∈ π2 and
⊥ 6∈ π1 , X ∇∇ = (π1 × O2 ∪ {⊥}) ∩ O./ .

Proof. Let us suppose ⊥ ∈ π1 and ⊥ 6∈ π2 . Let v ∈ π2 . Any pair (u, v) ∈ O./
verifies π20 ⊆ (u, v)∇ . Since, by the lemma 2, we have X ∇ = π20 , we can write
X ∇ ⊆ (u, v)∇ and thus, by derivation (u, v)∇∇ ⊆ X ∇∇ . Finally, for any u ∈
O1 ∪ {⊥}, we have (u, v) ∈ X ∇∇ donc X ∇∇ = (O1 ∪ {⊥} × π2 ) ∩ O./ . We show
the second assertion symmetrically.                                           t
                                                                              u

   The lemmas 2 and 3 allow us to determine that in cases where only one of
the projections contains ⊥ we can write a formal concept of K./ only with the
other projection. Let us now study a formal concept based on this projection
determined by RCA.
Lemma 4 If ⊥ ∈ π1 and ⊥ 6∈ π2 , there exists a concept C2 = (π2 , π20 ) on K2 .
If ⊥ ∈ π2 and ⊥ 6∈ π1 , there exists a concept C1 = (π1 , π10 ) on K1 .

Proof. Let us suppose ⊥ ∈ π1 and ⊥ 6∈ π2 . Since π2 ⊆ O2 , (π200 , π20 ) is a concept
son K2 . It is therefore sufficient to show that π200 = π2 , or more simply π200 ⊆ π2 .
Let o ∈ π200 . By construction at least one couple (ō, o) ∈ O./ and o0 ⊆ (ō, o)∇ .
Now, we have o ∈ π200 so by derivation, π20 ⊆ o0 . Moreover, by the lemma 2, we
have X ∇ = π20 . Thus, X ∇ ⊆ (ō, o)∇ so by derivation, (ō, o) ∈ X ∇∇ . Finally, by
definition of the projections o ∈ π2 . The second assertion holds by symmetry. t    u

    The following proposition gathers the previous lemmas. It emphasizes that,
in the case where only one of the two projections contains ⊥, any concept of
K./ can be expressed with the other projection. Moreover, there exists a con-
cept generated by RCA, of the same intent and whose extent corresponds to a
projection of the extent of the concept generated by FCA.

Proposition 2 Let C = (X ∇∇ , X ∇ ). If ⊥ ∈ π1 and ⊥ 6∈ π2 , C = ((O1 ∪ {⊥} ×
π2 ) ∩ O./ , π20 ) and there exists a corresponding concept C2 = (π2 , π20 ) on K2 .
If ⊥ ∈ π2 and ⊥ 6∈ π1 , C = ((π1 × O2 ∪ {⊥}) ∩ O./ , π10 ) and there exists a
corresponding concept C1 = (π1 , π10 ) on K1 .

Proof. Follows from the lemmas 2, 3 and 4.                                           t
                                                                                     u

   There remains a specific case, described by the lemma 5, to complete the
exhaustive description of a formal concept on the join table.

Lemma 5 For any X ⊆ O1 × O2 we have X ∇ = π10 ∪ π20 and X ∇∇ = {(o1 , o2 ) |
π10 ⊆ o01 ∧ π20 ⊆ o02 }

Proof. Let us show X ∇ = π10 ∪ π20 by double inclusion.

    (i) X ∇ ⊆ π10 ∪ π20 .
    The RCF modeling assures us that A1 ∩ A2 = ∅. Thus, an attribute a ∈ X ∇
is either in A1 or in A2 . If a ∈ X ∇ ∩ A1 , it must be shared by all the elements
of π1 ; and so a is in π10 . Similarly, if a is in X ∇ ∩ A2 , a ∈ π20 . We deduce that
X ∇ ⊆ π10 ∪ π20 .

    (ii) π10 ∪ π20 ⊆ X ∇ .
    On the other hand, if an attribute a is in π10 , then any pair of X ∇∇ has a first
component that carries the attribute a. Since this property is true for any pair
of X ∇∇ and X ⊆ X ∇∇ , then any pair of X carries the attribute a. Therefore,
we have a ∈ X ∇ . In the same way, we show that if a ∈ π20 , then a ∈ X ∇ . thus,
we have π10 ⊆ X ∇ and π20 ⊆ X ∇ . We can therefore affirm that π10 ∪ π20 ⊆ X ∇ .
    Finally, by (i) and (ii) we have X ∇ = π10 ∪ π20 . As X ∇∇ describes exactly the
set of couples (o1 , o2 ) having the attributes of π10 ∪ π20 , by construction of the
join table we have π10 ⊆ o01 and π20 ⊆ o02 .                                        t
                                                                                    u
    The cases described by the lemmas 1 and 2 allow for the immediate selection
of concepts from the RCA process corresponding in terms of extent to a concept
in the join table. The proposition 3 relies on the lemma 5 to state the main result
of this subsection, dealing with non-degenerate cases (without ⊥ element).
Proposition 3 Let X ⊆ O1 ×O2 . There exists by RCA on K1 a concept (X1 , Y1 )
such that X1 = π1 and π10 ⊆ Y1 and there exists on K2 a concept (X2 , Y2 ) such
that X2 = π2 and π20 ⊆ Y2 .
Proof. As π10 ⊆ A1 and π20 ⊆ A2 , C1 = (π100 , π10 ) and C2 = (π200 , π20 ) are formal
concepts on their respective contexts computed at step 0 of RCA.
    Let us consider the contexts K1 and K2 after graduation by the operator
∃ on the relations r and r−1 . We then have the attributes ∃r : C2 in K1 and
∃r−1 : C1 in K2 . We define the sets of attributes Y1 = π10 ∪ {∃r : C2 } and
Y2 = π20 ∪ {∃r−1 : C1 } as well as the concepts C3 = (Y10 , Y100 ) and C4 = (Y20 , Y200 )
(it is possible that C1 = C3 or C2 = C4 ). We have Y10 = π100 ∩ {∃r : C2 }0 , let us
show that Y10 = π1 by double inclusion.
    Let o ∈ π1 , we have o ∈ π100 . Moreover, by construction, any pair of (o, ō) ∈
X  ∇∇
         verifies (o, ō) ∈ r with ō ∈ π2 and, by hypothesis, ō 6= ⊥. Thus, since
π2 ⊆ π200 , o carries the attribute ∃r : C2 . Thus, we have π1 ⊆ Y10 .
    Let o ∈ Y10 . We have π10 ∪ {∃r : C2 } ⊆ o0 . Since {∃r : C2 } ⊆ o0 , there exists
ō ∈ π200 such that (o, ō) ∈ r and thus (o, ō) ∈ O./ . Moreover, since ō ∈ π200 , we have
π20 ⊆ ō0 . Since π20 ⊆ ō0 , π10 ⊆ o0 and that by the lemma 5 we have X ∇ = π10 ∪ π20 ,
we can affirm X ∇ ⊆ (o, ō)∇ . Finally (o, ō)∇∇ ⊆ X ∇∇ and by definition of π1 ,
on a o ∈ π1 (In a completely analogous way, we show π2 = Y20 ).
    Finally, we have shown the existence of C3 = (Y10 , Y100 ) such that Y10 = π1 and
π1 ⊆ Y100 as well as of C4 = (Y20 , Y200 ) such that Y20 = π2 and π2 ⊆ Y200 .              t
                                                                                           u
    In conclusion, the propositions 1, 2 and 3 show that for any concept C =
(X ∇∇ , X ∇ ) we find on K1 a concept (X1 , Y1 ) such that X1 = π1 and π10 ⊆ Y1
and there exists on K2 a concept (X2 , Y2 ) such that X2 = π2 and π20 ⊆ Y2 . It
is to note that if ⊥ ∈ π1 (respectively π2 ) we have π1 = {x | ∃y, (x, y) ∈ O./ }
(respectively π2 = {x | ∃x, (x, y) ∈ O./ }). Example 3 illustrates these properties.
Example 3. Let us consider the relational family as well as the semi-join context
defined in the Example 2.
    On the joined context, we find the concept C = ({(La, t3), (La, f 5)}, {Adult,
Female, IT, pw}). Here, π1 = {La} and π2 = {t3, f 5}. We check that there exists
on KP a concept (π1 , π10 ), namely the concept C1 = ({La}, {Adult, F emale,
IT }), and on KC a concept (π2 , π20 ), the concept C2 = ({t3, f 5}, {pw}). After an
iteration, RCA extends these concepts’ intents to {Adult, F emale, IT, ∃pos : C2 }
and {pw, ∃pos−1 : C1 }, respectively.

3.3   Aggregation operation in mono-relational case
Assume again the RCF in Figure 2 and let us consider FCA is applied on the
context schematically visualized in Figure 4. Intuitively, this amounts to extend-
ing the domain context of the relation by appending some new attributes. These
are derived from the range context attributes by a technique akin to relational
scaling, i.e. one basically simulating a one-shot RCA-like context refinement.
    Formally speaking, we design the context Kl = (Ol , Al , Il ) where Ol =
O1 , Al = A1 ∪ρ(A2 ), and ρ(A2 ) are the attributes resulting from the application
of the scaling operator ρ to the attribute concept of a ∈ A2 . We will denote such
an attribute ρr : a to avoid confusion with RCA’s own relational attributes. No-
tice that A1 ∩ ρ(A2 ) = ∅ holds. Next, we introduce Constraint(ρ, r, op , (X, Y )),
a predicate verifying whether op and (X, Y ), from the domain and the range of r,
respectively, jointly comply to the semantic of ρ. Thus, Constraint(∀, r, op , (X, Y ))
is true iff r(op ) ⊆ X. The predicate is a compact expression of the incidence Il :
 – if ap ∈ A1 then (op , ap ) ∈ Il iff (op , ap ) ∈ I1 ,
 – if ap ∈ ρ(A2 ) then (op , ap ) ∈ Il iff Constraint(ρ, r, op , (a0p , a00p )) is true.
   Example 4 illustrates the ∀∃ case (reasoning with other operators is similar).
For the O2 perspective, it is enough to swap K1 and K2 and replace r by r−1 .

Example 4. Consider again the RCF in Example 2. We aggregate the family via
∀∃: For an o ∈ OP and a ∈ AC s.t. ∀∃pos : a ∈ ρ(AC ) it holds (o, ∀∃pos : a) ∈ I
iff 1) ∃oC ∈ OC s.t. (o, oC ) ∈ pos and 2) ∀oC ∈ OC , (o, oC ) ∈ pos entails ov ∈ a0
(there is at least one image of o by pos and all such images carry a. Table 5
depicts the resulting aggregated context Kl .


  Kl
       Senior
       Adult
       Male
       Female
       I.T.
       Sport
       ∀∃pos : el
       ∀∃pos : pw
       ∀∃pos : cp
       ∀∃pos : ch




                                                    Kl
                                                         Senior
                                                         Adult
                                                         Male
                                                         Female
                                                         I.T.
                                                         Sport
                                                         ∃pos : el
                                                         ∃pos : pw
                                                         ∃pos : cp
                                                         ∃pos : ch
  Fa × ×                 × ×                         Fa × ×                 × ×
  La   × ××           ×                              La   × ××         × × ×
  Sh   × ×               ×                           Sh   × ×            × × ×
  Tr   ×× ××                                         Tr   ×× ××
   Table 5: Kl for Example 4                          Table 6: Kl for Example 6


   To define a formal concept on the aggregated table, we first identify the
component of the intent on the part ρ(A2 ). Again, we denote the derivations in
K1 and K2 by 0 , and in the aggregated context by ∇ .

Definition 1. The relational deviation of X ⊆ Ol , denoted δ(X), is the set of
its attributes from ρ(A2 ), i.e. δ(X) = X ∇ ∩ ρ(A2 ).

Proposition 4 Given a X ⊆ Ol , δ(X) = ∩o∈X {ρr : a | Constraint(ρ, r, o, (a0 , a00 ))}.

Proof. Let o ∈ X. By construction, for any ρr : a ∈ ρ(A2 ), holds ρr : a ∈ o∇
                       (a0 , a00 )). Thus o∇ ∩ ρ(A2 ) = {ρrT: a | Constraint(ρ, r, o,
iff Constraint(ρ, r, o,T
(a , a ))}. As X = o∈X o0 , we have X 0 ∩ ρ(A2 ) = o∈X o∇ ∩ ρ(A2 ), hence
   0 00         0

δ(X) = ∩o∈X {ρr : a | Constraint(ρ, r, o, (a0 , a00 ))}.                           t
                                                                                   u
    A formal concept on the aggregated context is then characterized by:
Proposition 5 Let X ⊆ Ol , then the concept C = (X ∇∇ , X ∇ ) of the aggre-
gated context satisfies X ∇ = X 0 ∪ δ(X) and X ∇∇ = X 00 ∩ {o | δ(X) ⊆ o∇ }.

Proof. By definition, we have Al = A1 ∪ ρ(A2 ) and A1 ∩ ρ(A2 ) = ∅. Thus we
can write X ∇ = (X ∇ ∩ A1 ) ∪ (X ∇ ∩ ρ(A2 )), that is X ∇ = X 0 ∪ δ(X).
   By deriving X ∇ , we determine that X ∇∇ = {o | o ∈ O1 ∧ X 0 ⊆ o∇ ∧ δ(X) ⊆
o }. Now, as X 0 ⊆ A1 , we have X 0 ⊆ o∇ iff X 0 ⊆ o∇ ∩ A1 , that is X 0 ⊆ o0 .
 ∇

Finally, X ∇∇ = X 00 ∩ {o | δ(X) ⊆ o∇ }.                                    t
                                                                            u

    Now, let us assume we have the result of RCA using the same relational
scaling with ρ along r on the simple RCF in Figure 2. Let X be the extent of
a concept from Kl . The set δ(X) is well-defined, hence we can denote its i-th
member by ρr : aδ,i (where aδ,i ∈ A2 ). As every concept Cδ(X),i = (a0δ,i , a00δ,i ) is
well defined on K2 , in RCA, K1 will be refined with all the attributes ρr : Cδ(X),i
at the first relational scaling step. Let Yδ = X 0 ∪i∈1..|δ(X)| ρr : Cδ(X),i , we claim
that C = (X ∇∇ , X ∇ ) and Cδ = (Yδ0 , Yδ00 ) have the same extent:

Proposition 6 Yδ0 = X ∇∇ .

Proof. Yδ0 ⊆ X ∇∇ : Let o ∈ Yδ0 . First, o carries all the attributes of X 0 , thus
X 0 ⊆ o∇ . Moreover, for each attribute ρr : aδ,i ∈ δ(X) a concept Ci = (a0δ,i , a00δ,i )
exists such that o carries the attribute ρr : Ci (for which Constraint(ρ, r, o, Ci )
is true). Since aδ,i is in the intent of Ci , we can verify that ρr : aδ,i ∈ o∇ . Since for
all i, we have ρr : aδ,i ∈ o∇ , then we have δ(X) ⊆ o∇ . Finally, since δ(X) ⊆ o∇
and X 0 ⊆ o∇ , we have X ∇ ⊆ o∇ . By derivation, we have o∇∇ ⊆ X ∇∇ . Finally,
Yδ0 ⊆ X ∇∇ .                                                                               t
                                                                                           u
     Yδ0 ⊇ X ∇∇ : The 1st relational scaling step will necessarily produce ρr :
(a0δ,i , a00δ,i ) for each ρr : aδ,i ∈ δ(X). Let o ∈ X ∇∇ , then o carries all the attributes
of X 0 . Moreover, after the scaling step, o gets incident to each attribute ρr : aδ,i ∈
δ(X) (Constraint(ρ, r, o, (a0δ,i , a00δ,i )) is necessarily satisfied). Thus, we have Yδ ⊆
o0 and therefore o00 ⊆ Yδ0 . Finally, since o ∈ o00 we conclude that X ∇∇ ⊆ Yδ0 . t        u

    Proposition 6 states that for any concept from the aggregated context, an
RCA concept with the same extent exists. Definition 2 introduces the notion of
relational weakening (illustrated by Example 5) to enable the mapping between
both intents. The latter is given by proposition 7.

Definition 2. Let a concept C be produced by RCA and let Yr be the set of
                         S intent of C. We call relational weakening of C, noted
relational attributes of the
Ω(C), the set Ω(C) = ρr:(U,V )∈Yr {ρr : v | v ∈ V }

Example 5. Assume the contexts of Example 2: Context KC gives rise to the
concepts C1 = ({t3}, {el, pw}), C2 = ({f 5}, {pw, cp}), C3 = ({t3, zo}, {el})
and C4 = ({zo, f 5}, {cp}). After a scaling with ∃, KP yields the concept C =
({La}, {Adult, F emale, I.T., ∃pos : C1 , ∃pos : C2 , ∃pos : C3 , ∃pos : C4 , ∃pos :
>}). Then Ω(C) = {∃pos : el, ∃pos : pu, ∃pos : cp}.
Proposition 7 δ(X) ⊆ Ω(Cδ )

Proof. Let’s denote by Yr the set of relational attributes in the intent of Cδ . Let
ρr : a ∈ δ(X), then by scaling and construction of Cδ , it holds ρr : (a0 , a00 ) ∈ Yr
and as a ∈ a00 , one concludes δ(X) ⊆ Ω(Cδ ).                                        t
                                                                                     u

    While we’ve just shown that the extents of C and Cδ are equal, their intents
might differ: As proposition 7 states, the intent of the aggregate table concept
is a subset of the weakening of the RCA concept with the same extent (see
Example 6 below). In this sense, we see the RCA concept as more informative.

Example 6. Assume the relational family defined by Example 2 with the ∃ op-
erator. The aggregated context is presented in table 6. Now, after one iteration,
RCA discovers the concept ({F a}, {Senior, M ale, ∃pos : (cp, ch)}) whereas FCA
finds ({F a}, {Senior, M ale, ∃pos : (cp), ∃pos : (nd))}). While ∃pos : (cp, ch) im-
plies ∃pos : (cp) and ∃pos : (ch), the reverse does not hold.


4    Discussion

We’ve shown that for any FCA concept from an encoded context, RCA would
reveal a counterpart concept, or a pair of such, conveying the same semantics
(equal extent). Moreover, the syntactic expression of the RCA concept(s) is
clearer than the FCA one, whatever the encoding. With semi-join, since separate
RCA concepts map to the 1st and 2nd projections of a FCA concept, the clarity
gain is immediate. Indeed, no confusion is ever possible as to which attribute of
the semi-join intent is incident to which object. Moreover, redundancy in FCA
concepts, e.g. shared 1st or 2nd projection, is avoided in RCA.
     With aggregation, RCA trivially produces a concept of the same extent, yet
it is more precise: The FCA counterpart is readily obtained by relational weak-
ening. Here, higher-order encoding schemata are conceivable that mimic RCA
iterations by nesting the scaling operators. Yet the maximal depth of these nest-
ings in the resulting (pseudo-)relational attributes must be fixed beforehand.
This is a serous limitation since we know of no simple way to determine the
number of iterations required till the fixed point for a given RCA task, i.e. a
RCF and a vector of scaling operators. This means that, at least in the realistic
cases, RCA will be revealing concepts that FCA –over the aggregated context
with all possible nestings of a depth up till the limit– will miss. A relevant
question here is whether knowing the fixed point contexts in RCA, there is an
equivalent aggregation context that comprises only nested operator attributes
referring exclusively to attribute concepts. This would mean that a static en-
coding, i.e. without the need for explicitly composing RCA concepts popping at
iterations 2+, exists. The cost of constructing such an encoding is, though, a
separate concern.
5    Conclusion

We tackled here the question of whether RCA brings some effective scope ex-
tension to the realm of FCA, given that FCA is at its core. We’ve examined two
complementary principles of encoding a relation into a single augmented context
and compared FCA output on each of the contexts to the output of RCA on the
original RCF. It was shown that in both cases, RCA is able to find counterpart
concepts (same extent) to those found by FCA, while the RCA intents at its 1st
iteration are at least as expressive as the FCA ones.
    A more systematic study should allow us to demonstrate similar results in
the more complex cases of multiple relations in the RCF as well as multiple
relations between the same pair of contexts.


References
 1. M. Alam et al. Lattice based data access: An approach for organizing and accessing
    linked open data in biology. In DMoLD@ECML/PKDD. Springer, 2013.
 2. P. Chen. The Entity-Relationship Model - Toward a Unified View of Data. ACM
    Transactions on Database Systems, 1(1):9–36, 1976.
 3. X. Dolques et al. RCA as a data transforming method: a comparison with propo-
    sitionalisation. In ICFCA, pages 112–127. Springer, 2014.
 4. X. Dolques et al. Performance-friendly rule extraction in large water data-sets with
    aoc posets and relational concept analysis. Intl. J-l of General Syst., 2016.
 5. S. Džeroski. Multi-relational data mining: an introduction. ACM SIGKDD Explo-
    rations Newsletter, 5(1):1–16, 2003.
 6. S. Ferré and P. Cellier. Graph-fca: An extension of formal concept analysis to
    knowledge graphs. Discrete Applied Mathematics, 273:81–102, 2020.
 7. S. Ferré and O. Ridoux. A logical generalization of formal concept analysis. In
    ICCS, pages 371–384. Springer, 2000.
 8. B. Ganter and S. Kuznetsov. Pattern structures and their projections. In ICCS,
    pages 129–142, 2001.
 9. S. Kramer et al. Propositionalization approaches to relational data mining. In
    Relational Data Mining, pages 262–291. Springer, 2001.
10. M. Rouane-Hacene et al. Relational concept analysis: mining concept lattices from
    multi-relational data. AMAI, 67(1):81–108, 2013.
11. E. Spyropoulou and T. De Bie. Interesting multi-relational patterns. In ICDM,
    pages 675–684. IEEE, 2011.
12. M. Wajnberg et al. Mining process factor causality links with multi-relational
    associations. In K-Cap, pages 263–266, 2019.
13. M. Wajnberg et al. Mining heterogeneous associations from pediatric cancer data
    by relational concept analysis. In MSDM@ICDM. IEEE, 2020.
14. R. Wille. Restructuring Lattice Theory: An Approach Based on Hierarchies of
    Concepts, pages 445–470. Springer Netherlands, Dordrecht, 1982.
15. R. Wille. Conceptual graphs and formal concept analysis. In ICCS, pages 290–303.
    Springer, 1997.