<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Concept Lattice Reduction by Singular Value Decomposition</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Martin Polovincak</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Hussam M. Dahwa</string-name>
          <email>hussamdahwag@vsb.cz</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Proceedings of the Spring Young Researcher's Colloquium on Database and Information Systems</institution>
          ,
          <addr-line>Moscow, Russia, 2007</addr-line>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>VSB Technical University Ostrava</institution>
          ,
          <country country="CZ">Czech Republic</country>
        </aff>
      </contrib-group>
      <fpage>2</fpage>
      <lpage>4</lpage>
      <abstract>
        <p>High complexity of lattice construction algorithms and uneasy way of visualising lattices are two important problems connected with the formal concept analysis. Algorithm complexity plays significant role when computing all concepts from a huge incidence matrix. In this paper we try to modify an incidence matrix using matrix decomposition, creating a new matrix with fewer dimensions as an input for some known algorithms for lattice construction. Results are presented by visualising neural network. Neural network is responsive for reducing result dimension to two dimensional space and we are able to present result as a picture that we are able to analyse.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>
        We are dealing with uses of matrix decompositions for
reduction of concept lattice. These methods are well
known in the area of information retrieval under the
name Latent Semantic Indexing (LSI) or Latent
Semantic Analysis (LSA). LSI and LSA have been used for
discovery of latent dependencies between terms (or
documents) [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. We would like to apply this approach
in the area of formal concept analysis (FCA). In this
paper, we want to present decomposition’s results with
neural network [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. First, we will introduce basic terms of
Formal Concept Analysis (FCA) [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] and then we will
describe the rank-k singe value decomposition (SVD)
[
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]. Singular Value Decomposition is one of the various
matrix decomposition techniques arising from
numerical linear algebra. SVD reduces both the column space
and the row space of the term-document matrix to lower
dimensional spaces [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. After using SVD on input
matrix we can build lattices and visualise them using neural
networks[
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] and u-matrix [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]. Numerous studies suggest
that graphical representation and display of results can
improve information retrieval performance [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ]. Once
having visualisation we can analyze result of matrix
reduction.
In the following paragraphs we would like to introduce
important basic terms of formal concept analysis,
singular value decomposition, self organized maps and unified
distrance matrix.
2.1
      </p>
    </sec>
    <sec id="sec-2">
      <title>Formal concept analysis</title>
      <p>
        Formal Concept Analysis was first introduced by Rudolf
Wille in 1980. FCA is based on the philosophical
understanding of the world in terms of objects and attributes.
It is assumed that a relation exists to connect objects to
the attributes they possess. Formal context and formal
concept are the fundamental notions of Formal Concept
Analysis [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ].
      </p>
      <p>A formal context C = (G; M; I) consists of two sets;
G and M , with I in relation to G and M . The elements
of G are defined as objects and the elements of M are
defined as attributes of the context. In order to express
that an object g 2 G is related to I with the attribute
m 2 M , we record it as gIm or (g; m) 2 I and read it
as the object g has the attribute m. I is also defined as the
context incidence relation.</p>
      <p>For a set A G of object we define</p>
      <p>A0 = fm 2 M j gIm f or all g 2 Ag
(the set of attributes common to the objects in A).
Correspondingly, for a set B of attributes we define</p>
      <p>B0 = fg 2 G j gIm f or all m 2 Bg
(the set of objects which have all attributes in B).</p>
      <p>A formal concept of the context (G; M; I) is a pair
(A; B) with A G, B M , A0 = B and B0 = A.
We call A the extent and B the intent of the concept
(A; B). B(G; M; I) denotes the set of all concepts of
context (G; M; I).</p>
      <p>The concept lattice B(G; M; I) is a complete lattice
in which infimum and supremum are given by:
^ (At; Bt) =
t2T
t2T
_ (At; Bt) =
\ At;
t2T
[ At
t2T</p>
      <p>
        00
[ Bt
t2T
t2T
We refer to [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ].
2.2
      </p>
    </sec>
    <sec id="sec-3">
      <title>Singular Value Decomposition</title>
      <p>
        Singular value decomposition (SVD) is well known
because of its application in information retrieval as LSI.
SVD is especially suitable in its variant for sparse
matrices [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ].
      </p>
    </sec>
    <sec id="sec-4">
      <title>Self-organizing maps</title>
      <p>Vma=l, a(nvd1;a: :d:i;avgro)n,awlmhoastreixcolu=mndivaegc(to r01s; :a:re:;0 orrth).o nT.k ohre.- . 2y . 4Tk d
decomposition A = U V T is called singular value
decomposition of matrix A and numbers 1; : : : ; r are
singular values of the matrix A. Columns of U (or V ) are
called left (or right) singular vectors of matrix A.</p>
      <p>Now we have a decomposition of the original matrix
A. It is not needed to say, that the left and right singular
vectors are not sparse. We have at most r nonzero
singular numbers, where rank r is the smaller of the two
matrix dimensions. However, we would not conserve
much memory by storing the term-by-document matrix
this way. Luckily, because the singular values usually
fall quickly, we can take only k greatest singular values
and corresponding singular vector co-ordinates and
create a k-reduced singular decomposition of A.</p>
      <p>Let us have k; 0 &lt; k &lt; r and singular value
decomposition of A</p>
      <p>A = U V T = (UkU0)
k
0
0
0</p>
      <p>VkT
V0T
We call Ak = Uk kVkT a k-reduced singular value
decomposition (rank-k SVD).</p>
      <p>In information retrieval, if every document is relevant
to only one topic, we obtain a latent semantics -
semantically related words and documents will have similar
vectors in the reduced space. For an illustration of rank-k
SVD see figure 1, the grey areas determine first k
coordinates from singular vectors, which are being used.</p>
      <p>Theorem 2: (Eckart-Young) Among all m n
matrices C of rank at most k Ak is the one, that minimises</p>
      <p>Cw;j )2.
jjAk Ajj2F = P(Ai;j</p>
      <p>i;j</p>
      <p>Because rank-k SVD is the best rank-k
approximation of original matrix A, any other decomposition will
increase the sum of squares of matrix A Ak.</p>
      <p>The SVD is hard to compute and once computed, it
reflects only the decomposition of the original matrix.
The recalculation of SVD is expensive, so it is
impossible to recalculate SVD every time new rows or columns
are inserted. The SVD-Updating is a partial solution, but
since the error increases slightly with inserted rows and
columns when updates occur frequently, the
recalculation of SVD may be needed.</p>
      <p>Note: From now on, we will assume rank-k singular
value decomposition when speaking about SVD.</p>
    </sec>
    <sec id="sec-5">
      <title>Unified distance matrix</title>
      <p>
        The unified distance matrix (U-matrix) makes the 2D
visualization of multi-variate data possible using SOM’s
codevectors as data source [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]. This is achieved by using
topological relations property among neurons after the
learning process. U-matrix contains the distances from
each unit center to all of its neighbours. By U-matrix we
can detect topological relations among neurons and
infer about the input data structure. High values in the
Umatrix represent a frontier region between clusters, and
low values represent a high degree of similarities among
neurons on that region, clusters. This can be a visual task
when we use some color schema.
3
      </p>
      <sec id="sec-5-1">
        <title>Our Experiment</title>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>3.1 Input data</title>
      <p>Data from the Ostrava Mass Transit Authority was used
in this experiment. We can describe data in a graph.
Rows represent tram stops, columns represent tram lines.
Let 1 represent a tram stopping at a tram stop in the
appropriate row and column crossing, otherwise zero. This
data was chosen because we can easily see if the SVD
is working. Many tram lines share the same tram stops.
These parts of tram lines should be minimized with SVD
and it should also be visible after using SVD.
3.2</p>
    </sec>
    <sec id="sec-7">
      <title>Steps of our Experiment</title>
      <p>Our experiment included several steps. The first step was
to read and transform data into adjacency matrix. After
transforming and computing (using SVD) three new
matrixes were obtained as results from SVD A = U V T .
After choosing rank-k, a new matrix was computed.
Lattices were build from both original matrix and matrix
computed by SVD. Concepts of lattices acted as the
source for SOMs learning phase. Having learned SOMs,
visualisation as done using U-matrix technique.
3.3</p>
    </sec>
    <sec id="sec-8">
      <title>Reading and transforming data</title>
      <p>Data was obtained from tram timetables. The first ten
tram lines and their tram stops were chosen. The
transformed matrix had 92 rows and 10 columns.
Transformation was simple. If a tram stops at a particular tram
stop, the number 1 appeared in the appropriate row and
column crossing, otherwise zero. The matrix is also an
incidence matrix and was used in SVD, NMF and FCA
computations.
3.4</p>
    </sec>
    <sec id="sec-9">
      <title>Using SVD</title>
      <p>SVD computations were made by SVDLIBC-fix
software. This software is free and is written in C language.
An incidence matrix acts as the input. Software can
compute all three matrixes - U and and V T . k-rank was
chosen k = 4. Matrixes were U (92x10), (10x10) and
V T (10x10).
3.5</p>
    </sec>
    <sec id="sec-10">
      <title>Visualising result</title>
      <p>Data from the Ostrava Mass Transit Authority was used
in this experiment. We can describe data in a graph.
Rows represent tram stops, columns represent tram lines.
Let ones represent a tram stopping at the tram stop in the
appropriate row and column crossing, otherwise zero.</p>
      <p>First SOM was learned by giving these initial data.
Second SOM was learned by giving data after applying
SVD on initial data. There were two SOM networks
according to type of input data.</p>
      <p>Next part of research was about visualising SOM
network via unified distance matrix (U-matrix) algorithm.
Euclidean metric was used for computing distances
between nodes. We got two greyscale pictures (figure 2,
figure 3) that we are able to analyze, especialy number
and form of visible clusters.</p>
      <sec id="sec-10-1">
        <title>Conclusion</title>
        <p>There are two u-matrix visualisations. Figure 2 matches
concepts of lattice built from the original matrix, figure 3
matches concepts of lattice built from the original matrix
changed with SVD. The number of differencies between
rows in tables was decreased by SVD computation. Thus
you can see smaller number of clusters. There is no doubt
method was successful in reducing number of concepts.</p>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>M.</given-names>
            <surname>Berry</surname>
          </string-name>
          and
          <string-name>
            <given-names>M.</given-names>
            <surname>Browne</surname>
          </string-name>
          .
          <article-title>Understanding search engines</article-title>
          .
          <source>Mathematical Modeling and Text Retrieval</source>
          ,
          <year>1999</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>C. Lee C.</given-names>
            <surname>Kiu</surname>
          </string-name>
          .
          <article-title>Discovering ontological semantics using fca and som</article-title>
          .
          <source>M2USIC</source>
          <year>2004</year>
          ,
          <year>2004</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>B.</given-names>
            <surname>Ganter</surname>
          </string-name>
          and
          <string-name>
            <given-names>R.</given-names>
            <surname>Wille</surname>
          </string-name>
          .
          <source>Formal Concept Analysis</source>
          . Springer-Verlag Berlin Heidelberg,
          <year>1999</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>Douglas</given-names>
            <surname>Vogel Karen S.K. Cheung</surname>
          </string-name>
          .
          <article-title>Complexity reduction in lattice-based information retrieval</article-title>
          .
          <source>Information Retrieval</source>
          ,
          <volume>8</volume>
          , page 285299,
          <year>2005</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>T.</given-names>
            <surname>Kohonen</surname>
          </string-name>
          .
          <article-title>Self-organizing maps</article-title>
          . Springer,
          <year>2001</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <surname>T.Letsche M. Berry</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Dumais</surname>
          </string-name>
          .
          <article-title>Computation methods for intelligent information access</article-title>
          .
          <source>Proceedings of the 1995 ACM/IEEE Supercomputing Conference</source>
          ,
          <year>1995</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <surname>T.Letsche M. Berry</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Dumais</surname>
          </string-name>
          .
          <article-title>Computation methods for intelligent information access</article-title>
          .
          <source>Proceedings of the 1995 ACM/IEEE Supercomputing Conference</source>
          ,
          <year>1995</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>P. Moravec P.</given-names>
            <surname>Gajdos</surname>
          </string-name>
          .
          <article-title>Concept lattice generation by singular value decomposition</article-title>
          .
          <source>CLA</source>
          <year>2004</year>
          , pages
          <fpage>13</fpage>
          -
          <lpage>22</lpage>
          ,
          <year>2004</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>J.</given-names>
            <surname>Vesanto</surname>
          </string-name>
          .
          <article-title>Som-based data visualization methods</article-title>
          . pages
          <fpage>6</fpage>
          -
          <lpage>9</lpage>
          ,
          <year>1999</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <surname>Lin</surname>
            <given-names>X.</given-names>
          </string-name>
          <article-title>Map displays for information retrieval</article-title>
          .
          <source>Journal of the American Society of Information Science</source>
          , pages
          <fpage>40</fpage>
          -
          <lpage>54</lpage>
          ,
          <year>1997</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>