<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Optimal Decompositions of Matrices with Grades into Binary and Graded Matrices?</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Eduard Bartl</string-name>
          <email>ebartl1@binghamton.edu</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Radim Belohlavek</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Jan Konecny</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Dept. Computer Science, Palacky University</institution>
          ,
          <addr-line>Olomouc Tomkova 40, CZ-779 00 Olomouc</addr-line>
          ,
          <country country="CZ">Czech Republic</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Dept. Systems Science and Industrial Engineering T. J. Watson School of Engineering and Applied Science Binghamton University-SUNY</institution>
          ,
          <addr-line>PO Box 6000, Binghamton, NY 13902-6000</addr-line>
          ,
          <country country="US">USA</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2008</year>
      </pub-date>
      <fpage>59</fpage>
      <lpage>70</lpage>
      <abstract>
        <p>The paper contributes to factor analysis of relational data. We study the problem of decomposition of object-attribute matrices with grades, i.e. matrices whose entries contain degrees to which objects have attributes. The degrees are taken from a bounded partially ordered scale. Examples of such matrices are binary matrices, matrices with entries from a finite chain, or matrices with entries from the unit interval [0, 1]. We study the problem of decomposition of a given object-attribute matrix I with grades into an object-factor matrix A and a binary factorattribute matrix B, with the number of factors as small as possible. We present a theorem describing optimal decompositions. The theorem shows that decompositions which use as factors particular formal concepts associated to I are optimal in that the number of factors involved is the smallest possible. Furthermore, we present an approximation algorithm for finding those decompositions and illustrative examples.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>
        Problem description in brief This paper presents results on optimal
decompositions of matrices with grades. Examples of such matrices are binary (or Boolean)
matrices, i.e. matrices which entries are 0 or 1. Other examples are matrices
which contain numbers from the unit interval [
        <xref ref-type="bibr" rid="ref1">0, 1</xref>
        ] as their entries. In general we
consider non-numerical matrices with entries from particular complete lattices
L (binary matrices and matrices with entries from [
        <xref ref-type="bibr" rid="ref1">0, 1</xref>
        ] are particular examples
with L = {0, 1} and L = [
        <xref ref-type="bibr" rid="ref1">0, 1</xref>
        ], respectively).
      </p>
      <p>We consider the following problem. Let L be a partially ordered scale bounded
from below and above by 0 and 1 (details specified later). Given an n × m matrix
I with entries from L (i.e. Iij ∈ L), we want to decompose I into a product</p>
      <p>I = A ◦ B
of an n × k matrix A with entries from L (i.e. Ail ∈ L) and a k × m binary matrix
B (i.e. Blj ∈ {0, 1}) with k as small as possible. The composition operation ◦
which we consider is defined by
k
(A ◦ B)ij = _ Ail ⊗ Blj , (1)</p>
      <p>
        l=1
where ⊗ is defined by a ⊗ 1 = a and a ⊗ 0 = 0. Note that if L = {0, 1} then A ◦ B
is the well-known Boolean product of binary matrices. Note also that if we allow
Ail ∈ L and Blj ∈ L and if ⊗ is a t-norm then ◦ is the product of graded matrices
well-known in fuzzy set theory, see e.g. [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ], and that such decompositions were
considered in [
        <xref ref-type="bibr" rid="ref4 ref7">4, 7</xref>
        ].
      </p>
      <p>
        Factor analysis model For a decomposition I = A ◦ B given by (1), Iij can be
interpreted as a degree to which there is a factor l such that l applies to object i
and l is associated to attribute j (j is a particular manifestation of l). This way,
a decomposition I = A ◦ B provides us with a factor analysis model (see [
        <xref ref-type="bibr" rid="ref1 ref13 ref16">1, 13,
16</xref>
        ] for references on factor analysis): A relationship between objects and original
attributes given by I is described using a relationship between the objects and
new variables, called factors, which is given by A, and a relationship between
factors and the original attributes, which is given by B. Note that we assume
that B is binary, i.e. that the relationship between factors and attributes is a
yes-or-no relationship. This feature distinguishes our approach from those which
we considered earlier.
      </p>
      <p>Needless to say, one can consider decompositions I = A ◦ B given by (1),
in which A is binary and B arbitrary. Obviously, using IT = BT ◦ AT , one
can reduce this type of decomposition to the first type (A arbitrary, B binary).
Therefore, we do not consider such case.</p>
      <p>Contribution of the paper We present a theorem regarding optimal
decompositions of a given matrix I which shows that decompositions which use as factors
particular formal concepts, called crisply generated concepts, are optimal in that
they involve the least number of factors among all decompositions of I.
Furthermore, we present an approximation algorithm for finding those decompositions
and provide illustrative examples.</p>
      <p>
        Related and previous work The paper is a continuation of our previous work [
        <xref ref-type="bibr" rid="ref4 ref6 ref7">4,
6, 7</xref>
        ]. In particular, in [
        <xref ref-type="bibr" rid="ref4 ref7">4, 7</xref>
        ] we considered decompositions I = A ◦ B given by
(1), in which both A and B were arbitrary, i.e. none of them was required to be
binary.
      </p>
      <p>
        Preliminaries from fuzzy logic We use standard notions of fuzzy logic and fuzzy
sets, see e.g. [
        <xref ref-type="bibr" rid="ref12 ref15 ref2">2, 12, 15</xref>
        ]. In particular, we use complete residuated lattices as
structures of truth degrees. Recall that a complete residuated lattice is an algebra
L = hL, ∧, ∨, ⊗, →, 0, 1i such that hL, ∧, ∨, 0, 1i is a complete lattice, hL, ⊗, 1i
is a commutative monoid, and ⊗ and → satisfy so-called adjointness condition,
i.e. a ⊗ b ≤ c if and only if a ≤ b → c. We assume familiarity with examples and
basic properties of residuated lattices. As an example, for L = [
        <xref ref-type="bibr" rid="ref1">0, 1</xref>
        ], a ⊗ b =
max(0, a+b−1), a → b = min(1, 1−a+b), the algebra L = h[
        <xref ref-type="bibr" rid="ref1">0, 1</xref>
        ], ∧, ∨, ⊗, →, 0, 1i
is a complete residuated lattice (so-called standard Lukasiewicz algebra). An
Lset in a universe set U is a mapping A : U → L.
2
2.1
      </p>
      <p>Optimal Decompositions</p>
      <p>Composition as W-superposition of matrices
We first observe that I = A ◦ B for n × k and k × m matrices A (graded) and B
(binary) means that I is a W-superposition of particular rectangular matrices.
Definition 1. Let K1, K2 ⊆ L. An n × m matrix J with entries from L is called
(K1, K2)-rectangular iff there exist L-sets C in {1, . . . , n} and D in {1, . . . , m}
with C(i) ∈ K1 and D(j) ∈ K2 such that J = C ⊗ D, i.e.</p>
      <p>Jij = C(i) ⊗ D(j)
(2)
for 1 ≤ i ≤ n, 1 ≤ j ≤ m.</p>
      <p>
        In particular, we need (L, {0, 1})-rectangular matrices and call these just
“rectangular”. The term “rectangular” is inspired by the “shape” of such
matrices. The following matrices are examples of ({0, 1}, {0, 1})-rectangular (J1) and
([
        <xref ref-type="bibr" rid="ref1">0, 1</xref>
        ], {0, 1})-rectangular (J2) matrices:
Theorem 1. I = A ◦ B for n × k and k × m matrices A and B with Ail ∈ L
and Blj ∈ {0, 1} iff I is a W-superposition of k (L, {0, 1})-rectangular matrices
J1, . . . , Jk, i.e. iff
      </p>
      <p>I = J1 ∨ J2 ∨ · · · ∨ Jk.</p>
      <p>Proof. Denote by Jl the ◦-product A l ◦ Bl of the l-th column A l of A and the
l-th row Bl of B, i.e. (Jl)ij = Ail ⊗ Blj . I = A ◦ B means Iij = (A ◦ B)ij , i.e.
Iij = Wk</p>
      <p>
        l=1(Ail ⊗ Blj ). Therefore, I = J1 ∨ J2 ∨ · · · ∨ Jk. Since B is a binary
matrix, Jl are (L, {0, 1})-rectangular matrices.
Example 1. To illustrate the content of Theorem 1, consider the following
decomposition I = A ◦ B:
Theorem 1 says that in order to find a decomposition I = A ◦ B, we need to find
a suitable set of (L, {0, 1})-rectangular matrices Jl whose W-superposition gives
I. We now describe decompositions of I which are optimal among all possible
decompositions in that the number k of factors is the smallest possible one. The
decompositions use so-called crisply generated formal concepts of I [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ].
Preliminaries on crisply generated formal concepts This section presents
preliminaries on formal concepts of data with fuzzy attributes, particularly on crisply
generated formal concepts. The reader is referred, e.g., to [
        <xref ref-type="bibr" rid="ref3 ref5">3, 5</xref>
        ] for details.
      </p>
      <p>Let X = {1, . . . , n} and Y = {1, . . . , m} be sets (of objects and attributes,
respectively), I be an n × m matrix with entries from a support set L of a
complete residuated lattice L. The degree Ixy ∈ L is interpreted as a degree
to which object x has attribute y. Consider the operators ↑ : LX → LY and
↓ : LY → LX defined by</p>
      <p>C↑(y) = Vx∈X (C(x) → Ixy),</p>
      <p>
        D↓(x) = Vy∈Y (D(y) → Ixy),
where → is the residuum of the complete residuated lattice L. That is, ↑ assigns
an L-set C↑ in Y to a given L-set C in X, and ↓ assigns an L-set D↓ in X
to a given L-set D in Y . C↑(y) can verbally be described as a degree to which
“for each object x ∈ X: if x is from C then x has attribute y” (note that
C↑(y) is just the degree of the last statement “for each · · · ” according to basic
principles of first-order fuzzy logic, see [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ]). Likewise, D↓(x) is the degree to
which “for each attribute y ∈ Y : if y is from D then x has attribute y” is true.
If L = {0, 1}, ↑ : LX → LY and ↓ : LY → LX coincide with the well-known
concept-derivation operators of the basic setting of formal concept analysis [
        <xref ref-type="bibr" rid="ref11 ref8">8,
11</xref>
        ]. ↑ and ↓ form a fuzzy Galois connection [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] and the compound operators ↑↓
and ↓↑ form particular closure operators in X and Y [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. A pair hC, Di consisting
of an L-set C in X and an L-set D in Y is called a formal concept of I if C↑ = D
and D↓ = C. C and D are called the extent and intent of hC, Di, respectively.
The set of all formal concepts of I is denoted by B(X, Y, I). With a partial order
≤ defined by
      </p>
      <p>
        hC1, D1i ≤ hC2, D2i iff C1 ⊆ C2 (iff D2 ⊆ D1)
for hC1, D1i, hC2, D2i ∈ B(X, Y, I), B(X, Y, I) happens to be a complete lattice,
so-called concept lattice associated to I [
        <xref ref-type="bibr" rid="ref2 ref3">2, 3</xref>
        ]. Note that C1 ⊆ C2 means that C1
is contained in C2, i.e. for each x ∈ X, C1(x) ≤ C2(x). For L = {0, 1}, B(X, Y, I)
coincides with the ordinary concept lattice [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ]. In [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ], the following notion was
introduced. A formal concept hC, Di ∈ B(X, Y, I) is called crisply generated if
there is a crisp L-set Dc ∈ {0, 1}Y , i.e. for each y ∈ Y : Dc(y) = 0 or Dc(y) = 1,
such that C = Dc↓ (and thus D = Dc↓↑). Let Bc(X, Y, I) denote the collection of
all crisply generated formal concepts of I, i.e.
      </p>
      <p>Bc(X, Y, I) = {hC, Di ∈ B(X, Y, I) | there is Dc ∈ {0, 1}Y : C = Dc↓}.
We need the following characterization of crisply generated formal concepts. For
L-sets C1, C2 ∈ LX and D1, D2 ∈ LY , we put hC1, D1i E hC2, D2i if for each
x ∈ X, y ∈ Y we have C1(x) ≤ C2(x) and D1(y) ≤ D2(y).</p>
      <p>
        Lemma 1 ([
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]). hC, Di is a crisply generated formal concept iff hC, Di is
maximal (w.r.t. E) such that (1) the rectangular matrix J defined by Jxy = C(x) ⊗
D(y) is contained in I (i.e. Jxy ≤ Ixy for all x, y) and (2) C(x) = VD(y)=1 Ixy.
Remark 1. Note that condition (2) of Lemma 1 means that for the crisp L-set
Dc ∈ {0, 1}Y corresponding to the 1-cut of D, which is defined by
Dc(y) =
1 if D(y) = 1,
0 if D(y) &lt; 1,
(3)
we have C = Dc↓.
      </p>
      <p>Matrices AF and BF For convenience, we identify 1 × p vectors with entries
from L with L-sets in {1, . . . , p} (the l-th coordinate of the vector = the degree
to which l belongs to the L-set). Given a set</p>
      <p>F = {hC1, D1i, . . . , hCk, Dki}
of L-sets Cl and Dl in {1, . . . , n} and {1, . . . , m}, respectively, with values from
L, define n × k and k × m matrices AF and BF by</p>
      <p>(AF )il = (Cl)(i) and (BF )lj = (Dl)(j).</p>
      <p>That is, the l-th column of AF is the transpose of the vector corresponding to
Cl and the l-th row of BF is the vector corresponding to Dl.
For F ⊆ B(X, Y, I), denote</p>
      <p>Fc = {hC, Dci | hC, Di ∈ F }.</p>
      <p>Note that Dc is defined by (3). We will show that sets Fc corresponding to sets
F of crisply generated formal concepts are fundamental for decompositions we
are looking for.</p>
      <p>The first theorem says that for every I, there is a decomposition AFc ◦ BFc
for some F ⊆ Bc(X, Y, I).</p>
      <p>Theorem 2 (universality). For every I with entries from L there is F ⊆
Bc(X, Y, I) such that I = AFc ◦ BFc , i.e. I is a product of A with entries from
L and B with entries from {0, 1}.</p>
      <p>Proof. Denote for l ∈ {1, . . . , m}, hCl, Dli = h{1/l}↓, {1/l}↓↑i. Here, {1/l} is a
singleton in {1, . . . , m}, i.e. and L-set defined by {1/l}(l) = 1 and {1/l}(j) =
0 for j 6= l. hCl, Dli are particular crisply generated formal concepts from
B(X, Y, I) and we have</p>
      <p>Iij = Wm</p>
      <p>
        l=1 Cl(i) ⊗ Dl(j),
see [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. Putting thus F = {hCl, Dli | l = 1, . . . , m}, we get I = AFc ◦ BFc .
      </p>
      <p>However, Theorem 2 and its proof yield only |F | = m, i.e. the number k = |F |
of factors equals the number m of attributes. In general, better decompositions
may exist, i.e. those with k &lt; m. The next theorem shows that the
decompositions which use crisply generated formal concepts of I as factors are optimal
among all decompositions of I.</p>
      <p>Theorem 3 (optimality). Let I = A ◦ B for n × k and k × m matrices A and
B with Ail ∈ L, Blj ∈ {0, 1}. Then there exists a set F ⊆ Bc(X, Y, I) of crisply
generated formal concepts of I such that for Fc we have</p>
      <p>|Fc| ≤ k
and for the n × |Fc| and |Fc| × m matrices AFc with entries from L and BFc
with entries from {0, 1} we have</p>
      <p>I = AFc ◦ BFc .</p>
      <p>Proof. Sketch: Let I = A◦B for an n×k matrix A with entries from L and a k×m
binary matrix B. Consider the corresponding rectangular matrices J1, . . . , Jk of
which I is a W-superposition according to Theorem 1. Denoting now the
Lsets in {1, . . . , n} and {1, . . . , m} corresponding to the l-th column of A and
the l-th row of B by Gl and Hl, respectively, we have Jl = Gl ⊗ Hl. We have
Gl ⊗ Hl ⊆ I and one can check that also H↓
l ⊗ Hl ⊆ I. The pair hHl↓, Hli satisfies
condition (2) of Lemma 1 (see also Remark 1). Therefore, hH↓, Hli is contained
l
in a maximal (w.r.t. E defined in the paragraph preceding Lemma 1) hCl, Dli
which is then, according to Lemma 1, a crisply generated formal concept of I.
As a result, Cl ⊗ Dl ⊆ I. Therefore, for F = {hC1, D1i, . . . , hCk, Dki} we have
|F | ≤ k. Because (Hl)j ∈ {0, 1} and because we may assume Hl ⊆ Dl, we get</p>
      <p>Note that using the notation from the proof of Theorem 3, two distinct
hGl, Hli’s may be contained in a single hCl, Dli, i.e. for hGl1 , Hl1 i 6= hGl2 , Hl2 i
we can have hCl1 , Dl1 i = hCl2 , Dl2 i. As a consequence, we may have |F | &lt; k.
3</p>
      <p>Algorithm
In this section, we present an approximation algorithm for computing a
decomposition I = A ◦ B of an n × m matrix I with entries from L into an n × k matrix
A with entries from L and a k × m binary matrix B with k as small as possible.
Note that we do not provide the approximation factor for this algorithm.</p>
      <p>
        Recall that for L = {0, 1} (i.e. the set of grades contains just 0 and 1), our
problem becomes a problem of decomposition of binary matrices. In particular, if
L = {0, 1}, we are given a binary matrix I and our aim is to find a decomposition
I = A ◦ B into an n × k binary matrix A and a k × m binary matrix B with
k as small as possible. This problem is NP-hard and its decision version is
NPcomplete, see e.g. [
        <xref ref-type="bibr" rid="ref17 ref18 ref19">17–19</xref>
        ], and also [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ].
      </p>
      <p>
        Due to NP-hardness of a problem of decomposition of binary matrices which
is a particular instance of our problem, we need to look for suitable
approximation algorithms. In the following, we propose a greedy approximation
algorithm inspired by the algorithms presented in [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] and [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]. Briefly, starting with
empty Fc, the algorithm selects a crisply generated concept hC, Di of I that
covers a large part of I which is still uncovered. For each such selected hC, Di,
the corresponding hC, Dci, see (3), is added to Fc. For determining hC, Di,
we use |D ⊕ j| which denotes the number of pairs hi, j0i of indices, for which
Iij0 = IFc ∨ (D ∪ { 1/j})↓ ⊗ (D ∪ { 1/j})↓↑ ij0 . We refer to this approach as to
Method 1. We also used Method 2 for which |D ⊕ j| takes into account also
entries IFc ∨ (D ∪ { 1/j})↓ ⊗ (D ∪ { 1/j})↓↑ ij0 which are close to Iij0 but not
necessarily equal (details will appear in a full version of this paper).
      </p>
      <p>
        Note that if L = {0, 1}, our algorithm works the same way as the one from
[
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. We performed several experiments with our algorithm. Due to limited scope,
we present the following one. We generated 1,000 matrices I of dimension 15×15
over 5-element chain L with Lukasiewicz operations. Each matrix was generated
as a product of a 15 × k matrix A and a k × 15 binary matrix B, so we knew
the number of factors (its upper bound, in fact). Table 1 shows the numbers of
factors (average value ± standard deviation) for decompositions of I obtained
by our algorithm (both for Methods 1 and 2).
      </p>
    </sec>
    <sec id="sec-2">
      <title>Algorithm 1 Find Factors</title>
      <p>Input: I (matrix with entries from L)
Output: Fc (set Fc for which I = AFc ◦ BFc )
set IFc to empty matrix ((IFc )ij = 0)
while I 6= IFc do
set D to ∅
set V to 0
while there is j such that D(j) &lt; 1 and |D ⊕ j| &gt; V do
select j such that D(j) &lt; 1 which maximizes |D ⊕ j|
set D to (D ∪ { 1/j})↓↑
set V to |D ⊕ j|
end while
set C to D↓
add hC, Dci to Fc
set IFc to IFc ∨ C ⊗ Dc
end while
In this section, we present an illustrative example regarding decompositions of
a matrix with grades into a matrix with grades and a binary matrix.</p>
      <p>In our example, we consider n users, m permissions, and a user-to-permission
assignment. The assignment can be represented by an n × m matrix I with
entries from a scale L = {0, r, w, 1}, with 0 representing “no permission”, r and
w representing “permission to read” and “permission to write”, respectively, and
1 representing “full permission”. We define a partial order on L such that 0 is
the least element, 1 is the greatest one, and elements r and w are incomparable,
see Fig. 1.</p>
      <p>Furthermore, we need to define operations of multiplication ⊗. We put x⊗y =
x ∧ y, for all x, y ∈ L. The residuum is then determined by ⊗ (due to the
requirement of adjointness, see Section 1) and is defined by x → y = 1 for x ≤ y,
x → y = y for all x &gt; y, and r → w = w, w → r = r.</p>
      <p>We want to decompose I into a product of n × k matrix A and k × m matrix
B where A and B represent a user-to-role and a role-to-permission relationship,
respectively. Therefore, the factors we want to discover are to be interpreted as
roles, such as “system administrator”, “standard user” or the like. Naturally, we
expect A to be a binary matrix (i.e. Ail ∈ {0, 1}), assigning roles to users (a
user has a given role or not), whereas B is graded matrix (i.e. Blj ∈ L). In order
to be consistent with previous chapters, A should be graded and B should be
binary matrix. Therefore, we use well-known fact that I = A ◦ B is equivalent
to I−1 = B−1 ◦ A−1. That is, instead of I we decompose I−1.</p>
      <p>As a particular example, we consider 9 users (or employees) and 5 file-types
in some computer system (for instance, “documents”, “archive files” or “system
files” could be some of these types). The user-to-permission relationship is
described in the table thereunder. The data can be visualized using a rectangular
grid, where , , , and represent permissions 0, r, w, and 1, respectively:</p>
    </sec>
    <sec id="sec-3">
      <title>Alice</title>
      <p>Bob
Charles
David
Eve
Frank
George
Henry
Isaac
i.e.,
 1 1 0 </p>
      <p>0 1 0
 1 1 0 
I = A ◦ B =  0111 1111 0100  ◦
 0 1 0 </p>
      <p>0 1 0
This decomposition can be displayed as:</p>
      <p>As we obtained a 9 × 3 binary matrix A describing a user-to-role assignment
and 3 × 5 matrix B describing a role-to-permission assignment. Therefore, we
obtained 3 factors: role1, role2, role3. The first role (corresponding to the first
row of matrix B) might be interpreted as “standard user”, the second one (the
middle row of B) as “anonymous user” (“guest”), and the third one (the last
row of B) as “system administrator”.</p>
      <p>According to matrix A, we assign roles to users by:
=</p>
      <p>◦
Alice - role1, role2,
Bob - role2,
Charles - role1, role2,
David - role2,
Eve - all roles,
Frank - role1, role2,
George - role1, role2,
Henry - role2,</p>
      <p>Isaac - role2.</p>
      <p>Next, we compute an approximate decomposition of I ≈ A ◦ B. By this we
mean that we want the entries of I to by similar to the corresponding entries of
A◦B to a degree which exceeds a given similarity threshold f . In our example we
set f = 0.9. Details regarding such similarity will be presented in a full version of
this paper. Let us just note that the similarity is based on the number of matrix
entries which have equal values in I and A ◦ B. A graphical representation of an
approximation decomposition computed by our algorithm depicted below.</p>
      <p>We can see that the approximate decomposition involves the two factors
corresponding to “standard user” and “anonymous user”, which were involved
also in the exact decomposition. However, the factor corresponding to “system
administrator” is no longer involved in the approximate decomposition. This
can be seen as the result of our attempt, due to performing an approximate
decomposition, to discover only a small number of factors (roles) which account
for most of the data and, hence, are common. The role of “system administrator”
is not common since the only user with this role is Eve.
5</p>
      <p>Conclusions and Future Research
We presented a theorem regarding optimal decomposition of a matrix with grades
into a matrix with grades and a binary matrix. Furthermore, we proposed a
greedy approximation algorithm for computing such decompositions and
examples illustrating such decompositions.</p>
      <p>Further issues and future research include the following items:
– Independence of ⊗ and →. It can be shown that the decompositions of a
graded matrix into a graded and a binary matrix do not depend, in a certain
sense, on the operations ⊗ and → on the scale L of grades. We sticked to
the framework which involves ⊗ and → to show how the problem addressed
in this paper fits into the results developed earlier. Details will be presented
in the full version of this paper.
– Decompositions of matrices with grades into matrices with further
constraints, different from the requirement of binarity of B.
– Approximation algorithms for approximate and exact decompositions of
matrices with grades.
– Applications of the underlying factor analysis model and comparison to other
models of factor analysis.
– Role of decompositions in machine learning and data mining (esp.
dimensionality reduction).</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Bartholomew</surname>
            ,
            <given-names>D. J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Knott</surname>
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Latent Variable Models</article-title>
          and
          <string-name>
            <given-names>Factor</given-names>
            <surname>Analysis</surname>
          </string-name>
          , 2nd Ed., London, Arnold,
          <year>1999</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Belohlavek</surname>
          </string-name>
          , R.:
          <source>Fuzzy Relational Systems: Foundations and Principles</source>
          . Kluwer, Academic/Plenum Publishers, New York,
          <year>2002</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Belohlavek</surname>
          </string-name>
          , R.:
          <article-title>Concept lattices and order in fuzzy logic</article-title>
          .
          <source>Annals of Pure and Applied Logic</source>
          <volume>128</volume>
          (
          <issue>1-3</issue>
          )(
          <year>2004</year>
          ),
          <fpage>277</fpage>
          -
          <lpage>298</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Belohlavek</surname>
          </string-name>
          , R.:
          <article-title>Optimal decompositions of matrices with grades</article-title>
          .
          <source>IEEE Intelligent Systems</source>
          <year>2008</year>
          (to appear).
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Belohlavek</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sklenar</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zacpal</surname>
          </string-name>
          , J.:
          <article-title>Crisply generated fuzzy concepts</article-title>
          .
          <source>In: B. Ganter and R</source>
          . Godin (Eds.):
          <source>ICFCA 2005, Lecture Notes in Artificial Intelligence 3403</source>
          , pp.
          <fpage>268</fpage>
          -
          <lpage>283</lpage>
          , Springer-Verlag, Berlin/Heidelberg,
          <year>2005</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Belohlavek</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Vychodil</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          :
          <article-title>Discovery of optimal factors in binary data via a novel method of matrix decomposition (to appear). Preliminary version appeared as: On Boolean factor analysis with formal concepts as factors</article-title>
          .
          <source>SCIS &amp; ISIS</source>
          <year>2006</year>
          ,
          <article-title>Int</article-title>
          .
          <source>Conf. Soft Computing and Intelligent Systems &amp; Int. Symposium on Intelligent Systems, Sep 20-24</source>
          ,
          <year>2006</year>
          , Tokyo, Japan, pp.
          <fpage>1054</fpage>
          -
          <lpage>1059</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Belohlavek</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Vychodil</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          :
          <article-title>Optimal decompositions of matrices with ordinal data (submitted).</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>Carpineto</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Romano</surname>
          </string-name>
          , G.:
          <article-title>Concept Data Analysis</article-title>
          .
          <article-title>Theory and Applications</article-title>
          . J. Wiley,
          <year>2004</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Cormen</surname>
            ,
            <given-names>T. H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Leiserson</surname>
            ,
            <given-names>C. E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rivest</surname>
            ,
            <given-names>R. L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Stein</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          : Introduction to Algorithms, 2nd Ed. MIT Press,
          <year>2001</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Frolov</surname>
            ,
            <given-names>A. A.</given-names>
          </string-name>
          , Hu´sek,
          <string-name>
            <given-names>D.</given-names>
            ,
            <surname>Muraviev</surname>
          </string-name>
          ,
          <string-name>
            <given-names>I. P.</given-names>
            ,
            <surname>Polyakov</surname>
          </string-name>
          ,
          <string-name>
            <surname>P. A.</surname>
          </string-name>
          :
          <article-title>Boolean factor analysis by Hopfield-like autoassociative memory</article-title>
          .
          <source>IEEE Transactions on Neural Networks</source>
          Vol.
          <volume>18</volume>
          , No. 3, May
          <year>2007</year>
          , pp.
          <fpage>698</fpage>
          -
          <lpage>707</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>Ganter</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wille</surname>
          </string-name>
          , R.:
          <article-title>Formal Concept Analysis</article-title>
          .
          <source>Mathematical Foundations</source>
          . Springer, Berlin,
          <year>1999</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12. H´ajek, P.: Metamathematics of Fuzzy Logic. Kluwer, Dordrecht,
          <year>1998</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13.
          <string-name>
            <surname>Harman</surname>
            ,
            <given-names>H. H.</given-names>
          </string-name>
          :
          <article-title>Modern Factor Analysis</article-title>
          , 2nd Ed. The Univ. Chicago Press, Chicago,
          <year>1970</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          14.
          <string-name>
            <surname>Keprt</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          , Sn´aˇsel, V.:
          <article-title>Binary factor analysis with help of formal concepts</article-title>
          .
          <source>In Proc. CLA</source>
          <year>2004</year>
          , Ostrava, Czech Republic,
          <year>2004</year>
          , pp.
          <fpage>90</fpage>
          -
          <lpage>101</lpage>
          , ISBN 80-248-0597-9.
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          15.
          <string-name>
            <surname>Klir</surname>
            ,
            <given-names>G. J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yuan</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          :
          <article-title>Fuzzy Sets and Fuzzy Logic</article-title>
          .
          <source>Theory and Applications</source>
          . Prentice-Hall,
          <year>1995</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          16.
          <string-name>
            <surname>McDonald</surname>
            ,
            <given-names>R. P.</given-names>
          </string-name>
          :
          <article-title>Factor Analysis and Related Methods</article-title>
          . Lawrence Erlbaum Associates, Inc.,
          <year>1985</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          17.
          <string-name>
            <surname>Nau</surname>
            ,
            <given-names>D. S.:</given-names>
          </string-name>
          <article-title>Specificity covering: immunological and other applications, computational complexity and other mathematical properties, and a computer program</article-title>
          . A.
          <string-name>
            <surname>M. Thesis</surname>
          </string-name>
          ,
          <source>Technical Report CS-1976-7</source>
          ,
          <string-name>
            <given-names>Computer</given-names>
            <surname>Sci</surname>
          </string-name>
          .Dept., Duke Univ.,
          <string-name>
            <surname>Durham</surname>
            ,
            <given-names>N. C.</given-names>
          </string-name>
          ,
          <year>1976</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          18.
          <string-name>
            <surname>Nau</surname>
            ,
            <given-names>D. S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Markowsky</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Woodbury</surname>
            <given-names>M. A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Amos D</surname>
          </string-name>
          . B.:
          <source>A Mathematical Analysis of Human Leukocyte Antigen Serology. Math. Biosciences</source>
          <volume>40</volume>
          (
          <year>1978</year>
          ),
          <fpage>243</fpage>
          -
          <lpage>270</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          19.
          <string-name>
            <surname>Stockmeyer</surname>
            ,
            <given-names>L. J.:</given-names>
          </string-name>
          <article-title>The set basis problem is NP-complete</article-title>
          .
          <source>IBM Research Report RC5431</source>
          , Yorktown Heights, NY,
          <year>1975</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          20.
          <string-name>
            <surname>Wille</surname>
          </string-name>
          , R.:
          <article-title>Restructuring lattice theory: an approach based on hierarchies of concepts</article-title>
          .
          <source>In: I. Rival (Ed.): Ordered Sets</source>
          ,
          <fpage>445</fpage>
          -
          <lpage>470</lpage>
          , Reidel, Dordrecht-Boston,
          <year>1982</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>