<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>An FCA-based Boolean Matrix Factorisation for Collaborative Filtering</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Elena Nenova</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Dmitry I. Ignatov</string-name>
          <email>dignatov@hse.ru</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Andrey V. Konstantinov</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Imhonet</institution>
          ,
          <addr-line>Moscow</addr-line>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>National Research University Higher School of Economics</institution>
          ,
          <addr-line>Moscow</addr-line>
        </aff>
      </contrib-group>
      <fpage>57</fpage>
      <lpage>73</lpage>
      <abstract>
        <p>We propose a new approach for Collaborative ltering which is based on Boolean Matrix Factorisation (BMF) and Formal Concept Analysis. In a series of experiments on real data (Movielens dataset) we compare the approach with the SVD- and NMF-based algorithms in terms of Mean Average Error (MAE). One of the experimental consequences is that it is enough to have a binary-scaled rating data to obtain almost the same quality in terms of MAE by BMF than for the SVD-based algorithm in case of non-scaled data.</p>
      </abstract>
      <kwd-group>
        <kwd>Boolean Matrix Factorisation</kwd>
        <kwd>Formal Concept Analysis</kwd>
        <kwd>Singular Value Decomposition</kwd>
        <kwd>Recommender Algorithms</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Introduction</title>
      <p>Recently Recommender Systems is one of the most popular subareas of Machine
Learning. In fact, the recommender algorithms based on matrix factorisation
techniques (MF) has become industry standard.</p>
      <p>
        Among the most frequently used types of Matrix Factorisation we de nitely
should mention Singular Value Decomposition (SVD) [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] and its various
modi cations like Probabilistic Latent Semantic Analysis (PLSA) [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ]. However,
the existing similar techniques, for example, non-negative matrix factorisation
(NMF) [
        <xref ref-type="bibr" rid="ref13 ref16 ref9">16,13,9</xref>
        ] and Boolean matrix factorisation (BMF) [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ], seem to be less
studied. An approach similar to matrix factorization is biclustering which was
also successfully applied in recommender system domain [
        <xref ref-type="bibr" rid="ref11 ref18">18,11</xref>
        ]. For example,
Formal Concept Analysis [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] can also be used as a biclustering technique and
there are some of its applications in recommenders' algorithms [
        <xref ref-type="bibr" rid="ref10 ref6">6,10</xref>
        ].
      </p>
      <p>The aim of this paper is to compare recommendation quality of some of the
aforementioned techniques on real datasets and try to investigate the methods'
interrelationship. It is especially interesting to conduct experiments on
comparison of recommendations quality in case of an input matrix with numeric
values and in case of a Boolean matrix in terms of Precision and Recall as well
as MAE. Moreover, one of the useful properties of matrix factorisation is its
ability to keep reliable recommendation quality even in case of dropping some
insu cient factors. For BMF this issue is experimentally investigated in section
4.</p>
      <p>
        The novelty of the paper is de ned by the fact that it is a rst time when
BMF based on Formal Concept Analysis [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] is investigated in the context of
Recommender Systems.
      </p>
      <p>The practical signi cance of the paper is determined by demands of the
recommender systems' industry, that is to gain reliable quality in terms of Mean
Average Error (MAE), Precision and Recall as well as time performance of the
investigated method.</p>
      <p>The rest of the paper consists of ve sections. The second section is an
introductory review of the existing MF-based recommender approaches. In the third
section we describe our recommender algorithm which is based on Boolean
matrix factorisation using closed sets of users and items (that is FCA). Section 4
contains methodology of our experiments and results of experimental
comparison of di erent MF-based recommender algorithms by means of cross-validation
in terms of MAE, Precision and Recall. The last section concludes the paper.
2</p>
    </sec>
    <sec id="sec-2">
      <title>Introductory review of some matrix factorisation approaches</title>
      <p>In this section we brie y describe di erent approaches to the decomposition of
both real-valued and Boolean matrices. Among the methods of the SVD group
we describe only SVD. We also discuss nonnegative matrix factorization (NMF)
and Boolean matrix factorization (BMF).
2.1</p>
      <sec id="sec-2-1">
        <title>Singular Value Decomposition (SVD)</title>
        <p>Singular Value Decomposition (SVD) is a decomposition of a rectangular matrix
A 2 Rm n(m &gt; n) into the product of three matrices</p>
        <p>A = U
0</p>
        <p>
          V T ;
(1)
where U 2 Rm m and V 2 Rn n are orthogonal matrices, and 2 Rn n is
a diagonal matrix such that = diag( 1; : : : ; n) and 1 2 : : : n 0.
The columns of the matrix U and V are called singular vectors, and the numbers
i are singular values [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ].
        </p>
        <p>In the context of recommendation systems rows of U and V can be interpreted
as vectors of the user's and items's loyalty (attitude) to a certain topic (factor),
and the corresponding singular values as the importance of the topic among the
others. The main disadvantage is in the fact that the matrix may contain both
positive and negative numbers; the last ones are di cult to interpret.</p>
        <p>The advantage of SVD for recommendation systems is that this method
allows to obtain the vector of its loyalty to certain topics for a new user without
SVD decomposition of the whole matrix.</p>
        <p>
          The evaluation of computational complexity of SVD according to [
          <xref ref-type="bibr" rid="ref15">15</xref>
          ] is
O(mn2) oating-point operations if m n or more precisely 2mn2 + 2n3.
        </p>
        <p>Consider as an example the following table of movie ratings:
A = BBBB 00 00 00 45 44 05 03 CCCC :
U = BBBB 00::2597</p>
        <p>It can be seen that the greatest weight have the rst three singular values,
which is con rmed by the calculations:
Non-negative Matrix Factorization (NMF) is a decomposition of non-negative
matrix V 2 Rn m for a given number k into the product of two non-negative
matrices W 2 Rn k and H 2 Rk m such that</p>
        <p>V</p>
        <p>W H:
(2)</p>
        <p>
          NMF is widely used in such areas as nding the basis vectors for images,
discovering molecular structures, etc. [
          <xref ref-type="bibr" rid="ref16">16</xref>
          ].
        </p>
        <p>Consider the following matrix of ratings:</p>
        <p>Its decomposition into the product of two non-negative matrices for k = 3
can be, for example, like this:
Basic FCA de nitions. Formal Concept Analysis (FCA) is a branch of
applied mathematics and it studies (formal) concepts and their hierarchy. The
adjective \formal" indicates a strict mathematical de nition of a pair of sets,
called, the extent and intent. This formalisation is possible because the use of
the algebraic lattice theory.</p>
        <p>Definition 1. Formal context K is a triple (G; M; I), where G is the set of
objects, M is the set of attributes , I G M is a binary relation.</p>
        <p>The binary relation I is interpreted as follows: for g 2 G, m 2 M we write
gIm if the object g has the attribute m.</p>
        <p>For a formal context K = (G; M; I) and any A G and B M a pair of
mappings is de ned:</p>
        <p>A0 = fm 2 M j gIm for all g 2 Ag;</p>
        <p>
          B0 = fg 2 G j gIm for all m 2 Bg;
these mappings de ne Galois connection between partially ordered sets (2G; )
and (2M ; ) on disjunctive union of G and M . The set A is called closed set, if
A00 = A [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ].
        </p>
        <p>Definition 2. A formal concept of the formal context K = (G; M; I) is a
pair (A; B), where A G, B M , A0 = B and B0 = A. Set A is called the
extent, and B is the intent of the formal concept (A; B).</p>
        <p>It is evident that extent and intent of any formal concept are closed sets.</p>
        <p>The set of formal concepts of a context K is denoted by B(G; M; I).
Description of FCA-based BMF Boolean matrix factorization (BMF) is
a decomposition of the original matrix I 2 f0; 1gn m, where Iij 2 f0; 1g;
into a Boolean matrix product P Q of binary matrices P 2 f0; 1gn k and
Q 2 f0; 1gk m for the smallest possible number of k: We de ne boolean matrix
product as follows:
(P</p>
        <p>Q)ij =
k
_ Pil Qlj ;
l=1
where W denotes disjunction, and conjunction.</p>
        <p>Matrix I can be considered as a matrix of binary relations between set X
objects (users), and the set Y attributes (items that users have evaluated). We
assume that xIy i user x estimated object y. The triple (X; Y; I) is clearly
composes a formal context.</p>
        <p>
          Consider the set F B(X; Y; I), a subset of all formal concepts of context
(X; Y; I), and introduce the matrices PF and QF :
(PF )il =
1; i 2 Al;
0; i 2= Al;
(QF )lj =
1; j 2 Bl;
0; j 2= Bl:
We can consider the decomposition of the matrix I into binary matrix product
PF and QF as described above. The following theorems are proved in [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ]:
Theorem 1. (Universality of formal concepts as factors). For every I there is
        </p>
        <p>
          F B(X; Y; I), such that I = PF QF :
Theorem 2. (Optimality of formal concepts as factors). Let I = P Q for n k
and k m binary matrices P and Q. Then there exists a F B(X; Y; I)
of formal concepts of I such that jF j k and for the n jF j and jF j m
binary matrices PF and QF we have I = PF QF :
There are several algorithms for nding PF and QF by calculating formal
concepts based on these theorems [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ].
        </p>
        <p>
          The algorithm we use (Algoritm 2 from [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ]) avoid the computation of all the
possible formal concepts and therefore works much faster [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ]. Time estimation
of the calculation algorithm in the worst case yields O(kjGjjM j3), where k is the
number of found factors, jGj is the number of objects, jM j this the number of
attributes.
        </p>
        <p>Transform the matrix of ratings described above, to a boolean matrix, as
follows:
0 4 4 5 0 0 0 0 1 0 1 1 1 0 0 0 0 1
B 5 5 3 4 3 0 0 C B 1 1 1 1 1 0 0 C
BB 0 0 0 4 4 0 0 CC
BB 0 0 0 5 4 5 3 CC ) BBBB 00 00 00 11 11 10 10 CCCC = I:
The decomposition of the matrix I into the Boolean product of I = AF
the following:
BF is
0 1 1 1 0 0 0 0 1
B 1 1 1 1 1 0 0 C
BBBB 00 00 00 11 11 01 01 CCCC = BBBB 00 11 01 CCCC
Once the matrix of rates is factorized we need to learn how to compute
recommendations for users and to evaluate whether a particular method handles well
with this task.</p>
        <p>For factorized matrices the already well-known algorithm based on the
similarity of users can be applied, where for nding K nearest neighbors we use not
the original matrix of ratings A 2 Rm n, but the matrix U 2 Rm f , where m
is a number of users, and f is a number of factors. After selection of K users,
which are the most similar to a given user, based on the factors that are peculiar
to them, it is possible, based on collaborative ltering formulas to calculate the
projected rates for a given user.</p>
        <p>After formation of the recommendations the performance of the
recommendation system can be estimated by measures such as Mean Absolute Error (MAE),
Precision and Recall.</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>A recommender algorithm using FCA-based BMF</title>
      <p>kNN-based algorithm
Collaborative recommender systems try to predict the utility of items for a
particular user based on the items previously rated by other users.</p>
      <p>Denote u(c; s) the utility of item s for user c. u(c; s) is estimated based on the
utilities u(ci; s) assigned to item s by those users ci 2 C who are \similar" to user
c. For example, in a movie recommendation application, in order to recommend
movies to user c, the collaborative recommender system nds the users that have
similar tastes in movies with c (rate the same movies similarly). Then, only the
movies that are most liked by those similar users would be recommended.</p>
      <p>Memory-based recommendation system, which are based on the previous
history of the ratings, are one of the key classes of collaborative recommendation
systems.</p>
      <p>
        Memory-based algorithms make rating predictions based on the entire
collection of previously rated items by the users. That is, the value of the unknown
rating rc;s for user c and item s is usually computed as an aggregate of the
ratings of some other (usually, the K most similar) users for the same item s:
rc;s = aggrc02Cbrc;s;
where Cb denotes the set of K users that are the most similar to user c , who
have rated item s. For example, function aggr may has the following form [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]
rc;s = k X sim(c0; c)
      </p>
      <p>rc0;s;
c02Cb
where k serves as a normalizing factor and selected as k = 1= P sim(c; c0).
c02Cb</p>
      <p>The similarity measure between users c and c0, sim(c; c0), is essentially a
distance measure and is used as a weight, i.e., the more similar users c and c0
are, the more weight rating rc0;s will carry in the prediction of rc;s:</p>
      <p>The similarity between two users is based on their ratings of items that
both users have rated. The two most popular approaches are correlation and
cosine-based. One common strategy is to calculate all user similarities sim(x; y)
in advance and recalculate them only once in a while (since the network of
peers usually does not change dramatically in a short time). Then, whenever the
user asks for a recommendation, the ratings can be calculated on demand using
precomputed similarities.</p>
      <p>To apply this approach in case of FCA-based BMF recommender algorithm
we simply consider as an input the user-factor matrices obtained after
factorisation of the initial data.
3.2</p>
      <sec id="sec-3-1">
        <title>Scaling</title>
        <p>In order to move from a matrix of ratings to a Boolean matrix, and use the
results of Boolean matrix factorization, scaling is required. It is well known that
scaling is a matter of expert interpretation of original data. In this paper, we
use several variants of scaling and compare the results in terms of MAE.
1. Iij = 1 if Rij &gt; 0, else Iij = 0 (user i rates item j).
2. Iij = 1 if Rij &gt; 1; else Iij = 0:
3. Iij = 1 if Rij &gt; 2; else Iij = 0.
4. Iij = 1 if Rij &gt; 3, else Iij = 0:
4</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Experiments</title>
      <p>To test our hypotheses and study the behavior of recommendations based on the
factorization of a ratings matrix by di erent methods we used MovieLens data.
We used the part of data, containing 100,000 ratings, while considered only users
who have given over 20 ratings.</p>
      <p>User ratings are split into two sets, a training set consisting of 80 000 ratings,
and test set consisting of 20 000 ratings. Original data matrix is 943 1682, where
the number of rows is the number of users and the number of columns is the
number of rated movies (each lm has at least one vote).
4.1</p>
      <p>The number of factors that cover p% of evaluations in an input
data for SVD and BMF
The main purpose of matrix factorization is a reduction of matrices
dimensionality. Therefore we examine how the number of factors varies depending on the
method of factorization, and depending on p % of the data that is covered by
factorization. For BMF the coverage of a matrix is calculated as the ratio of
the number of ratings covered by Boolean factorization to the total number of
ratings.</p>
      <p>MAE-based recommender quality comparison of SVD and
BMF for various levels of evaluations coverage
The main purpose of matrix factorisation is a reduction of matrices
dimensionality. As a result some part of the original data remains uncovered, so it was
interesting to explore how the quality of recommendations changes based on
different factorisations, depending on the proportion of the data covered by factors.</p>
      <p>Two methods of matrix factorisation were considered: BMF and SVD. The
fraction of data covered by factors for SVD was calculated as
and for BMF as</p>
      <sec id="sec-4-1">
        <title>To quality assessment we chose M AE.</title>
        <p>PK i2
p% = iP=1 i2</p>
        <p>100%;
p% = jcovered ratingsj 100%:</p>
        <p>jall ratingsj</p>
        <p>Fig. 1 shows that M AESV D60, calculated for the model based on 60% of
factors, is not very di erent from M AESV D80, calculated for the model built for
80% factors. At the same time, for the recommendations based on a Boolean
factorization covering 60% and 80% of the data respectively, it is clear that
increasing the number of factors improves MAE, as shown in Fig. 2.</p>
        <p>Table 3 shows that the MAE for recommendations built on a Boolean
factorisation covering 80 % of the data for the number of neighbors less than 50 is
better than the MAE for recommendations built on SVD factorization. It is also
easy to see that M AESV D80 and M AEBMF 80 are di erent from M AEall in no
more than 1 7%.</p>
        <p>Comparison of kNN-based approach and BMF by Precision and
Recall
Besides comparison of algorithms using MAE other evaluation metrics can also
be exploited, for example</p>
        <p>Recall = jobjects in recommendation \ objects in testj
;
jobjects in testj
P recision = jobjects in recommendation \ objects in testj</p>
        <p>jobjects in recommendationj
and</p>
        <p>It is widely spread belief that the larger Recall, Precision and F1 are, the
better is recommendation algorithm.</p>
        <p>Figures 3, 4 and 5 show the dependence of relevant evaluation metrics on
the percentage of the data covered by BMF-decomposition, and the number of
nearest neighbors. The number of objects to recommend was chosen to be 20.
The gures show that the recommendation based on the Boolean decomposition,
is worse than recommendations built on the full matrix of ratings.</p>
        <p>Scaling in uence on the recommendations quality for BMF in
terms of MAE
Another thing that was interesting to examine was the impact of scaling
described in 3.2 on the quality of recommendations. Four options of scaling were
considered:
1. I0;ij = 1 if Aij &gt; 0, else Iij = 0 (user rates an item).
2. I1;ij = 1 if Aij &gt; 1; else Iij = 0:
3. I2;ij = 1 if Aij &gt; 2; else Iij = 0.
4. I3;ij = 1 if Aij &gt; 3, else Iij = 0:</p>
      </sec>
      <sec id="sec-4-2">
        <title>The distribution of ratings in data is on Figure 6</title>
        <p>For each of the boolean matrices we calculate its Boolean factorisation,
covering 60 % and 80 % of the data. Then recommendations are calculated just like
in 4.2. It can be seen that for both types of data coverage M AE1 is almost the
same as M AE0, and M AE2;3 is better than M AE0.</p>
        <p>ltering on MAE for BMF kNN-based
Besides the ability to search for K nearest neighbors not in the full matrix of
ratings A 2 Rn m, but in the matrix U 2 Rm f , where m is a number of users,
and f is a number of factors, Boolean matrix factorization can be used to data
ltering. Because the algorithm returns as an output not only matrices
usersfactors and factors-objects, but also the ratings that were not used for factoring,
we can try to search for users, similar to the user, on the matrix consisting only
of ratings used for the factorization.</p>
        <p>Just as before to nd the nearest neighbors cosine measure is used, and the
predicted ratings are calculated as the weighted sum of the ratings of nearest
users. Figure 9 shows that the smaller the data we use for ltering the bigger is
MAE. Figure 10 shows that the recommendations built on user-factor matrix,
are better then recommendations, constructed on matrix of ratings ltered with
boolean factorization.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>Conclusion</title>
      <p>In the paper we considered main methods of Matrix Factorisation which are
suitable for Recommender Systems. Some of these methods were compared on real
datasets. We investigated BMF behaviour as part of recommender algorithm. We
also conducted several experiments on recommender quality comparison with
numeric matrices, user-factor and factor-item matrices in terms of Recall, Precision
and MAE. We showed that MAE of our BMF-based approach is not greater than
MAE of SVD-based approach for the same number of factors on the real data.
For methods that require the number of factors as an initial parameter in the
user or item pro le (e.g., NMF), we proposed the way of nding this number
with FCA-based BMF. We also have investigated how data ltering, namely
scaling, in uences on recommendations' quality.</p>
      <p>
        As a further research direction we would like to investigate the proposed
approaches in case of graded and triadic data [
        <xref ref-type="bibr" rid="ref3 ref4">3,4</xref>
        ] and reveal whether there are
some bene ts for the algorithm's quality in usage least-squares data imputation
techniques [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ]. In the context of matrix factorisation we also would like to test
our approach on the quality assessment of recommender algorithms that we
performed on some basic algorithms (see bimodal cross-validation in [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ]).
Acknowledgments. We would like to thank Radim Belohlavek, Vilem
Vychodil and Sergei Kuznetsov for their comments, remarks and explicit and
implicit help during the paper preparations. We also express our gratitude to
Gulnaz Bagautdinova; she did her bachelor studies under the second author
supervision on a similar topic and therefore contributed somehow to this paper.
      </p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Adomavicius</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tuzhilin</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Toward the Next Generation of Recommender Systems: A Survey of the State-of-the-Art and Possible Extensions</article-title>
          .
          <source>IEEE Transactions on Knowledge and data engineering</source>
          , vol.
          <volume>17</volume>
          (
          <issue>6</issue>
          ) (
          <year>2005</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Belohlavek</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Vychodil</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          :
          <article-title>Discovery of optimal factors in binary data via a novel method of matrix decomposition</article-title>
          ,
          <source>Journal of Computer and System Scinces</source>
          ,
          <volume>76</volume>
          (
          <year>2010</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Belohlavek</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Osicka</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          :
          <article-title>Triadic concept lattices of data with graded attributes</article-title>
          .
          <source>Int. J. General Systems</source>
          <volume>41</volume>
          (
          <issue>2</issue>
          ),
          <fpage>93</fpage>
          -
          <lpage>108</lpage>
          (
          <year>2012</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Belohlavek</surname>
          </string-name>
          , R.:
          <article-title>Optimal decompositions of matrices with entries from residuated lattices</article-title>
          .
          <source>J. Log. Comput</source>
          .
          <volume>22</volume>
          (
          <issue>6</issue>
          ),
          <fpage>1405</fpage>
          -
          <lpage>1425</lpage>
          (
          <year>2012</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Birkho</surname>
          </string-name>
          , G.:
          <article-title>Lattice Theory, eleventh printing</article-title>
          , Harvard University, Cambridge, MA (
          <year>2011</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>du</surname>
            Boucher-Ryan,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bridge</surname>
            ,
            <given-names>D.G.</given-names>
          </string-name>
          :
          <article-title>Collaborative Recommending using Formal Concept Analysis</article-title>
          .
          <source>Knowl.-Based Syst</source>
          .
          <volume>19</volume>
          (
          <issue>5</issue>
          ),
          <fpage>309</fpage>
          -
          <lpage>315</lpage>
          (
          <year>2006</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Elden</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          :
          <article-title>Matrix Methods in Data Mining and Pattern Recognition</article-title>
          ,
          <source>Society for Industrial and Applied Mathematics</source>
          (
          <year>2007</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>Ganter</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Wille</surname>
          </string-name>
          , R.:
          <source>Formal Concept Analysis: Mathematical Foundations</source>
          , Springer (
          <year>1999</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Gaussier</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Goutte</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          :
          <article-title>Relation between PLSA and NMF and Implications</article-title>
          .
          <source>In: SIGIR'05: Proceedings of the 28th annual international ACM SIGIR conference on Research and development in information retrieval</source>
          . New York, NY, USA. ACM, pp.
          <fpage>601</fpage>
          -
          <lpage>602</lpage>
          (
          <year>2005</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Ignatov</surname>
            ,
            <given-names>D.I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kuznetsov</surname>
            ,
            <given-names>S.O.</given-names>
          </string-name>
          :
          <article-title>Concept-based Recommendations for Internet Advertisement</article-title>
          .
          <source>In.: Proc. of The Sixth International Conference Concept Lattices and Their Applications (CLA'08)</source>
          , Radim Belohlavek, Sergei O. Kuznetsov (Eds.):
          <source>CLA</source>
          <year>2008</year>
          , Palacky University, Olomouc, pp.
          <volume>157166</volume>
          (
          <year>2008</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>Ignatov</surname>
            ,
            <given-names>D.I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kuznetsov</surname>
            ,
            <given-names>S.O.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Poelmans</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          :
          <article-title>Concept-Based Biclustering for Internet Advertisement</article-title>
          .
          <source>ICDM Workshops</source>
          <year>2013</year>
          , pp.
          <fpage>123</fpage>
          -
          <lpage>130</lpage>
          (
          <year>2013</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <string-name>
            <surname>Ignatov</surname>
            ,
            <given-names>D.I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Poelmans</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Dedene</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Viaene</surname>
            ,
            <given-names>S.:</given-names>
          </string-name>
          <article-title>A New Cross-Validation Technique to Evaluate Quality of Recommender Systems</article-title>
          .
          <source>PerMIn</source>
          <year>2012</year>
          , pp.
          <fpage>195</fpage>
          -
          <lpage>202</lpage>
          . (
          <year>2012</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13.
          <string-name>
            <surname>Lee</surname>
            ,
            <given-names>D.D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Seung</surname>
            ,
            <given-names>H.S.</given-names>
          </string-name>
          :
          <article-title>Algorithms for Non-Negative matrix factorization</article-title>
          ,
          <source>advances in Neural Information Processing Systems 13: Proceedings of the 2000 Conference</source>
          . MIT Press. pp.
          <fpage>556562</fpage>
          . (
          <year>2000</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          14.
          <string-name>
            <surname>Leksin</surname>
            ,
            <given-names>A.V.</given-names>
          </string-name>
          :
          <article-title>Probabilistic models in client environment analysis</article-title>
          ,
          <source>PhD Thesis</source>
          (
          <year>2011</year>
          )
          <article-title>(In Russian)</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          15.
          <string-name>
            <surname>Lloyd</surname>
            <given-names>N.</given-names>
          </string-name>
          <string-name>
            <surname>Trefethen</surname>
            and
            <given-names>David</given-names>
          </string-name>
          <string-name>
            <surname>Bau</surname>
          </string-name>
          .
          <article-title>Numerical Linear Algebra, 3rd edition</article-title>
          ,
          <source>SIAM</source>
          (
          <year>1997</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          16.
          <string-name>
            <surname>Lin</surname>
          </string-name>
          ,
          <string-name>
            <surname>Ch</surname>
          </string-name>
          .-J:
          <article-title>Projected Gradient Methods for Non-negative Matrix Factorization</article-title>
          , Neural computation
          <volume>19</volume>
          (
          <issue>10</issue>
          ), pp.
          <volume>2756</volume>
          {
          <issue>2779</issue>
          (
          <year>2007</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          17.
          <string-name>
            <surname>Mirkin</surname>
            ,
            <given-names>B.G.</given-names>
          </string-name>
          :
          <article-title>Core Concepts in Data Analysis: Summarization, Correlation</article-title>
          , Visualization (
          <year>2010</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          18.
          <string-name>
            <surname>Symeonidis</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Nanopoulos</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Papadopoulos</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Manolopoulos</surname>
          </string-name>
          , Ya.:
          <article-title>Nearestbiclusters collaborative ltering based on constant and coherent values</article-title>
          .
          <source>Inf. Retr</source>
          .
          <volume>11</volume>
          (
          <issue>1</issue>
          ):
          <fpage>51</fpage>
          -
          <lpage>75</lpage>
          (
          <year>2008</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          19.
          <string-name>
            <surname>Wasito</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mirkin</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          :
          <article-title>Nearest neighbours in least-squares data imputation algorithms with di erent missing patterns</article-title>
          .
          <source>Computational Statistics &amp; Data Analysis</source>
          <volume>50</volume>
          (
          <issue>4</issue>
          ),
          <fpage>926</fpage>
          -
          <lpage>949</lpage>
          (
          <year>2006</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          20.
          <string-name>
            <surname>Zhou</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cichocki</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Xie</surname>
          </string-name>
          , Sh.:
          <article-title>Fast Nonnegative Matrix/Tensor Factorization Based on Low-Rank Approximation</article-title>
          .
          <source>IEEE Transactions on Signal Processing</source>
          <volume>60</volume>
          (
          <issue>6</issue>
          ), pp.
          <fpage>2928</fpage>
          -
          <lpage>2940</lpage>
          (
          <year>2012</year>
          )
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>