<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Cheating to achieve Formal Concept Analysis over a large formal context?</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Victor Codocedo</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Carla Taramasco</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Hernan Astudillo</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Ecole Polytechnique</institution>
          ,
          <addr-line>32 Boulevard Victor 75015 Paris</addr-line>
          ,
          <country country="FR">France</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>LORIA</institution>
          ,
          <addr-line>BP 70239, F-54506 Vandoeuvre-les-Nancy</addr-line>
          ,
          <country country="FR">France</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Universidad Tecnica Federico Santa Mar a</institution>
          ,
          <addr-line>Av. Espan~a 1640. Valpara so</addr-line>
          ,
          <country country="CL">Chile</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Researchers are facing one of the main problems of the Information Era. As more articles are made electronically available, it gets harder to follow trends in the di erent domains of research. Cheap, coherent and fast to construct knowledge models of research domains will be much required when information becomes unmanageable. While Formal Concept Analysis (FCA) has been widely used on several areas to construct knowledge artifacts for this purpose [17] (Ontology development, Information Retrieval, Software Refactoring, Knowledge Discovery), the large amount of documents and terminology used on research domains makes it not a very good option (because of the high computational cost and humanly-unprocessable output). In this article we propose a novel heuristic to create a taxonomy from a large term-document dataset using Latent Semantic Analysis and Formal Concept Analysis. We provide and discuss its implementation on a real dataset from the Software Architecture community obtained from the ISI Web of Knowledge (4400 documents).</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>Research communities are facing one of the main problems of the Information Era
and Formal Concept Analysis is not prepared to solve it. The amount of articles
available online is growing each year yielding di cult to track trends, following
ideas, looking for new terminology, etc. While some communities have
understood the need for an artifact representing the knowledge within the domain
(such as an ontology, a body-of-knowledge or a taxonomy) the problem remains
in its construction since it is hard (highly technical), expensive (researchers are
scarce) and complex (information is dynamic).</p>
      <p>
        Automatic and semi-automatic creation of a terms taxonomy have been
widely boarded in several elds [
        <xref ref-type="bibr" rid="ref13 ref3 ref4 ref5">3,4,5,13,24</xref>
        ]. In this work we focus on the
approach described by Roth et al. [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ] in which a taxonomy is derived from a
corpus of documents by the use of Formal Concept Analysis (FCA). In
particular, they describe an application used to \represent a meaningful structure of
? We would like to thank Chilean project FONDEF D08I1155 ContentCompass,
intrabasal project FB/20SO/10 in the context of the Chilean basal project FB0821 and
ECOS-CONICYT project C09E08 for funding this work.
a given knowledge community in a form of a lattice-based taxonomy". This
application is illustrated using a set of abstracts of the embryologist community
obtained from MedLine spanning 5 years where a random set of 25 authors and
18 terms were analyzed. Although the lattice-based taxonomy obtained was a
fair representation of the domain, real-size corpora of research communities are
rather much larger than this example.
      </p>
      <p>
        Handling large datasets has been de ned as one of the open problems in the
community of FCA4 for two main reasons: rst, the computational costs involved
in the calculation of the concept lattice can make the use of FCA prohibitive
and second, the concept lattice structure yielded could be so complex that its
use may be impossible [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ].
      </p>
      <p>
        Iceberg lattices [21] help in improving readability by eliminating \not
representative" data, but useful information, such as \emerging behaviors [
        <xref ref-type="bibr" rid="ref12 ref15">12,15</xref>
        ],
is lost in the process. Stabilized lattices (using a stability measure [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ]) also
improves readability by eliminating \noisy elements" from data, but being a
post-process tool it also raises computational costs.
      </p>
      <p>We describe in this document a novel heuristic to create a lattice-based
taxonomy from a large corpus using Formal Concept Analysis and a widely used
Information Retrieval technique called Latent Semantic Analysis (LSA). In
particular, we describe a process to compress a formal context into a smaller reduced
context in order to obtain a lattice of terms that can be used to describe the
knowledge on a given research domain. We illustrate our approach using a
realsize dataset from a research community of Computer Sciences.</p>
      <p>The remainder of this paper proceeds as follows: Section 2 explains the basis
of FCA, section 3 presents our approach and section 4, a case study over a real
dataset from a research community. Section 5 presents the results and a
comparison of the obtained taxonomy with a human-expert handmade thesaurus.
Finally, the conclusions are described in section 6.
2</p>
    </sec>
    <sec id="sec-2">
      <title>Formal Concept Analysis</title>
      <p>Formal Concept Analysis, originally developed as a sub eld of applied
mathematics [23], is a method for data analysis, knowledge representation and
information management. It organizes information in a lattice of formal concepts.
A formal concept is constituted by its extension (the objects that compose the
concept) and its intension (the attributes that objects share). Objects and
attributes are placed as rows and columns (resp.) in a cross-table or formal context
where each cell indicates whether the object of that row have the attribute of that
column. In what follows, we describe the Formal Concept Analysis framework
as synthesized by Wille [22].
4 http://www.upriss.org.uk/fca/problems06.pdf
2.1</p>
      <sec id="sec-2-1">
        <title>Framework</title>
        <p>Let G be a set of objects, M a set of attributes and I a binary relation between G
and M (I (G M )) indicating by gIm that the object g contains the attribute
m and K = (G; M; I) be the formal context de ned by G, M and I. For A G
and B M it is de ned the derivation operator (0) as follows:
(1)
(2)
(4)
(5)</p>
      </sec>
      <sec id="sec-2-2">
        <title>2.2 Iceberg Concept Lattices</title>
        <p>Let (A; B) be a concept of B(K), its support is de ned as:</p>
        <p>A
supp(A; B) = j j</p>
        <p>
          jGj
(A; B) = jfC
Stability was proposed by Kuznetsov in [
          <xref ref-type="bibr" rid="ref14 ref16">14,16</xref>
          ] as a mechanism to prune \noisy
concepts". It was extended by Roth et al. We use and provide their de nition
from [
          <xref ref-type="bibr" rid="ref19">19</xref>
          ] and [
          <xref ref-type="bibr" rid="ref15">15</xref>
          ]:
        </p>
        <p>Let K = (G; M; I) be a formal context and (A; B) be a formal concept of K.
The stability index, , of (A; B) is de ned as follows:</p>
        <p>Stability measures how much the intent of a concept depends on particular
objects of its extent, meaning that if the formal context changes and some objects
disappear, then stability indicates how likely it is for a concept to remain in the
concept lattice. Stability can also be used to construct a stabilized lattice for a
given threshold similarly to an iceberg lattice.</p>
        <p>Analogous to de nition 5, the extensional stability of a concept (A; B)
can be de ned as:
e(A; B) = jfD</p>
        <p>B j D0 = Agj
2jBj
(6)</p>
        <p>Extensional stability measures how likely is for a concept to remain if some
attributes are eliminated from the context. We will use both de nitions in this
work di erentiating them as intensional stability (on (5)) and extensional
stability (on (6)).
3</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Reducing a large formal context</title>
      <p>
        Di erent from Roth's approach [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ], we are not interested in tracking groups of
people working on groups of topics, but rather in the relations among topics.
These relations occur in the articles that authors write, where topics or terms
can appear in sets and each one can appear one or more times. To elaborate:
      </p>
      <p>Given a corpus of articles G, a list of terms M and the relation among them
I (G M ) indicating by gIm that the article g contains the term m, the
document-article formal context is de ned as:</p>
      <p>KO = (G; M; I)
(7)
3.1</p>
      <sec id="sec-3-1">
        <title>Rationale</title>
        <p>Even for a small set of terms, the amount of articles for a small research
community can reach thousands of articles making the processing of KO impossible
or useless. The problem gets worse over time, because it can be expected that
each year hundreds of articles will be added to the corpus.</p>
        <p>
          What happens with terms over time? In taxonomy evolution, as described in
[
          <xref ref-type="bibr" rid="ref18">18</xref>
          ], symmetric patterns arise: some elds will progress or decline; some elds
will contain more or less concepts (enrichment or impoverishment ); and some
elds will merge or split. In any case, it is not expected that the amount of terms
would vary greatly.
        </p>
        <p>
          Latent Semantic Analysis (LSA) or Latent Semantic Indexing (LSI) [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ] is a
technique used commonly in Information Retrieval (IR) as a tool for indexation,
clusterization and query answering. LSA is based on the idea that for a given
set of terms and documents, the relation among terms can be explained by a set
of dimensions whose size is much smaller than the amount of documents. We
exploit this feature of LSA to construct a reduced formal context of
dimensions and terms having as conditions that information regarding relations of
terms cannot be lost and that it has to produce a coherent taxonomy using less
computational time. In what follows, we provide a brief description of LSA to
elaborate on how we used it to produce a reduced formal context. For further
reading, please refer to [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ].
3.2
        </p>
      </sec>
      <sec id="sec-3-2">
        <title>Latent Semantic Analysis</title>
        <p>Given a list of m terms and a corpus of n documents, let A be a term-document
matrix of rank-min(m,n) as de ned in 8, where aij is the weight5 of the term i in
the document j. The Single-Value Decomposition of matrix A (in equation (9))
produces its factorization in three matrices where contains the single-values
of matrix A at the diagonal in descending order and the columns of matrices U
and V are called left and right singular vectors of A.
(8)
(9)
(10)
(11)
(12)
Am n = [aij ] ; i = [1::m]; j = [1::n]
Am n = Um m
A0m n = Um k
m n V T</p>
        <p>n n
k k V T
k n</p>
        <p>
          Since singular values drops quickly, we can create a new approximation of
matrix A using k min(m; n) as shown in (10). Matrix A0 A is the closest
k-rank matrix approximation to A by the Frobenius measure [
          <xref ref-type="bibr" rid="ref11">11</xref>
          ]. Two new
matrices can be calculated:
        </p>
        <p>Bm k = Um k</p>
        <p>Cn k = Vn k
k k
k k
where B holds the vector-space representation in k dimensions of terms; and
C the one of documents. Both of these matrices are used for clusterization since,
on them, similar elements are closer on each dimension. In particular, each
dimension on B (each column) has a Gaussian-like distribution where terms group
around the mean value (see gure 1), except for dimension 0 (the di erent
behavior in gure 1(b)) where terms have almost the same value6. We exploit this
feature to de ne a conversion-function that allows us to construct the reduced
context.
3.3</p>
      </sec>
      <sec id="sec-3-3">
        <title>A probabilistic-based conversion-function</title>
        <p>
          Which terms are related within a given dimension? Since each dimension holds
continuous values, it is hard to de ne a region for them. Nevertheless, we know
5 Several weighting functions can be used, being the most used frequency of term and
term frequency-inverse document frequency (tf.idf)
6 We do not use the information in this dimension for our analysis and exclude it from
our results.
00.10 0.05 0.00 0.05 Coordinate Value 0.20 0.25 0.30 0.35
0.10 0.15
00.2
0.1
0.0 Coordinate Value 0.2
0.1
0.3
0.4
(a) Distribution of values in dimension 1
(b) Distribution of values in all dimensions
where function Gk is the probability density function (PDF) for dimension k
and 2 ]0,0.5[ de nes the limits of the \belonging region".
Similar techniques have been proposed before. Gajdos et al[
          <xref ref-type="bibr" rid="ref9">9</xref>
          ] used LSA to
reduce complexity in the structure of the lattice by eliminating noise in the formal
context. While this approach is useful, it does not reduce the amount of data,
but it \tunes it" to get a clearer result. Snasel et al. [
          <xref ref-type="bibr" rid="ref20 ref9">20,9</xref>
          ] proposed a
matrixreduction algorithms based on NMF. 7 and SVD8. While they state that these
methods are successful to reduce the amount of concepts obtained using FCA,
they do not describe a real life use of their technique (their experiment was
performed over a 17x16 matrix) neither do they discuss about the performance
of their approach. Kumar and Srinivas [
          <xref ref-type="bibr" rid="ref1">1</xref>
          ] approach consists of using fuzzy
KMeans clustering 9 to reduce the attributes in a formal term-document context.
In their approach, documents are categorized in k clusters using the cosine
similarity measure. Cheung et al. [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ] introduced term-document lattices complexity
reduction by de ning a set of equivalence relations that allows to reduce the set
of objects. Finally, Dias et al. introduced JBOS [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ] (junction based on objects
similarity) which proposed a similar method where objects where group into
prototype objects by calculating its similarity according to certain weights assigned
manually to attributes.
4
        </p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Case Study: Software Architecture Community</title>
      <p>The Software Architecture Corpus (SAC) was composed by extracting metadata
from papers retrieved by the ISI Web of Knowledge search engine 10 using the
query "software architecture". It is assumed that the keyword "software
architecture" is present in each paper on their titles and/or abstracts.</p>
      <p>While the search engine retrieved 4701 articles, not all of them have an
abstract to work with. Those are excluded from our analysis leaving 4565 articles
spanning from 1990 to 2009 (retrieved documents span from 1973 to 2009).
4.1</p>
      <sec id="sec-4-1">
        <title>Term list</title>
        <p>A term list was assembled by using Natural Language Processing over the
articles' titles and abstracts. In order to avoid common words, a stopword list and a
lexical tagger were used as a lter. A list of candidate terms was then manually
ltered to obtain a nal list of 120 terms, which included words and multi-words
(such as \Uni ed Model Language"). Table 1 shows a sample of selected terms.</p>
        <p>Each term was looked up on each document and its frequency of use was
calculated. Then, a weighting measure was applied (tf.idf11) to each value. The
7 Non-negative matrix factorization
8 Single-Value Decomposition
9 K-Means Clustering is a classic clustering technique for vector-space models
10 http://isiwebofknowledge.com
11 Term Frequency-Inverse Document Frequency is a weighting measure commonly
used on IR based on the notion that term infrequency on a global scale makes it
important.
term-document matrix Aw = aij was constructed using the nal list of terms
(M) and the corpus of documents (G) where aij represents the weight of term i
in document j. We de ned the relation I (G M ) = f(j; i) : 8j 2 G ^ 8i 2
M () aij &gt; 0g to build up the original context KO = (G; M; I) describing
that a document contains a term only if its weight on it is over 0. The formal
context KO was used later to compare our reductions.
As we stated at the end of section 3.3, two parameters had to be set in order to
create the reduced context. Sadly, in LSA there is not a known method to nd
the best value for k, and not knowing that, it is not possible to nd a good value
for . We de ned a set of goals to observe which were the values of k and that
best accomplished them. The goals de ned were:
{ Low Execution time
{ High Stability
{ Few Concepts in the nal lattice
Using three xed values for k we reduced several contexts and processed them
through FCA in order to nd the best value for . As shown in gure 2, it was
found that higher values of (close to 0.5) yields the best results. Repeating
the experience with 3 xed values for (0.45, 0.47 and 0.49) to nd the best
value for k we found a trade-o between stability and execution time as it can
be observed in gure 3. Higher values of k yield higher stability but also a high
execution time, and vice-versa. Since stability drops fast on k=60 and in the
same value the execution time grows greatly, we selected it to obtain our results.
was set on 0.45 and 0.47.
0.30
0.25
1.0
0.8
12 http://coron.loria.fr/</p>
        <p>Results shows that using LSA before FCA performs a clear reduction in
the formal context from a size of 4565 120 (original context) to 60 120
(reduced context), speci cally a reduction of 76 times the amount of data to
be processed. It also lowers the amount of concepts yielded in the nal
lattice (27 and 141 times for equal to 0.45 and 0.47 resp.), and because of that
the time required to calculate the full concept lattice is considerably
reduced, even considering the time required to create the reduced contexts.</p>
        <p>Stability gives more clues about the good quality of the reduction. Figure 4
shows intensional and extensional stability distribution. As it can be observed,
the original context's lattice has a better intensional stability than the reduced
contexts but a worst extensional stability. Mean values for these two measures
are shown in table 2.</p>
        <p>Since we have eliminated redundant data, each dimension is almost equally
important meaning that in reduced contexts we cannot a ord to eliminate a
subset of them without a ecting greatly the structure of the lattice obtained. In
this case, we have eliminated a big part of the noise (k=60 was in fact a very good
choice). On the other hand, the growth in extensional stability re ects that the
structure of the reduced lattices is not tied to some speci c terms. Some terms
can be removed and the structure of the lattice would not vary greatly, which is
what happens each year (see section 3.1).
5.1</p>
      </sec>
      <sec id="sec-4-2">
        <title>A Software Architecture Taxonomy</title>
        <p>
          Figure 5 shows the reduced notation of the lattice for the reduced context (k=60
and = 0:45). This lattice was drawn with Coron-drawer13 a set of scripts
specially written for large lattices. For the sake of space and simplicity we provide
13 http://code.google.com/p/coron-drawer/
4
itsh
3
2
1
00.1 0.2 0.3 0.4 0.5Stability0.6 0.7 0.8 0.9 1.0
00.1 0.2 0.3 0.4 0.5Stability0.6 0.7 0.8 0.9 1.0
(a) Intensional Stability
(b) Extensional Stability
a small comparison of the terms in the reduced lattice-based taxonomy with a
human-expert handmade thesaurus of Software Architecture [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ].
Software Architecture Thesaurus Comparison The thesaurus contains
494 elements (we call them elements to di erentiate them from lattice's
concepts and taxonomy's terms) organized in a hierarchical fashion. They have at
most one parent and the hierarchy has multiple roots. The thesaurus is
exhaustive and comprises mainly de nitions of Software Architecture's concepts and
entities (such as framework's names or important authors in the domain). The
comparison shows:
{ From our 120 term list, 50 terms (42%) match exactly with a term on the
thesaurus. 25 terms (21%) have a semi-match, meaning that they are part
of a term on the thesaurus (database in our hierarchy and shared database
in the thesaurus) and 45 (37%) terms do not have a simile in the thesaurus.
{ The three main concepts design, analysis and framework (with support over
50%) found in our taxonomy, also remain being main elements in the
thesaurus.
{ Even when some elements in the thesaurus are not found in our taxonomy,
they actually exists as relations among terms. For instance, the relation
among the terms design and pattern describe the thesaurus' element design
pattern. This is also true for design decision, information view, knowledge
reuse, quality requirements, business methodology and several more elements.
6
        </p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>Conclusions</title>
      <p>In this work we have presented a method and a technique to apply Formal
Concept Analysis (FCA) to large contexts of data in order to obtain a
latticebased taxonomy. We have outlined that large-size datasets are not suitable to
be processed by FCA and that, this fact is an important problem in the domain.</p>
      <p>The solution presented here, is based on an Information Retrieval technique
called Latent Semantic Analysis which is used to reduce a term-document matrix
to a much smaller matrix where terms are related to a set of dimensions instead
of documents. Using a probabilistic approach, this matrix is converted into a
binary formal context where FCA can be applied.</p>
      <p>The approach was illustrated with a case study using a research domain from
computational sciences called Software Architecture. The corpus created for this
domain consists of more than 4500 documents and 120 terms. We have
compared the characteristics of the lattice obtained through FCA from the original
formal context of terms and documents and the reduced contexts generated by
our approach. We have found that not only our approach is considerably more
economic in execution time as well as in the amount of concepts obtained in
the nal lattice but intensional and extensional stabilities give us elements to be
certain of the quality of our approach.</p>
      <p>A small comparison with a human expert handmade thesaurus of the
community of Software Architecture is provided in order to illustrate that a real and
coherent taxonomy can be obtained using our approach.
21. Gerd Stumme. E cient data mining based on formal concept analysis. In
Abdelkader Hameurlain, Rosine Cicchetti, and Roland Traunmuller, editors, Database and
Expert Systems Applications, volume 2453 of Lecture Notes in Computer Science,
pages 3{22. Springer Berlin / Heidelberg, 2002.
22. Rudolf Wille. Restructuring lattice theory: an approach based on hierarchies of
concepts. In Ivan Rival, editor, Ordered sets, pages 445{470, Dordrecht{Boston,
1982. Reidel.
23. Rudolf Wille. Formal concept analysis as mathematical theory of concepts and
concept hierarchies. In Bernhard Ganter, Gerd Stumme, and Rudolf Wille, editors,
Formal Concept Analysis, volume 3626 of Lecture Notes in Computer Science,
pages 1{33. Springer Berlin / Heidelberg, 2005.
24. Jian-hua Yeh and Naomi Yang. Ontology construction based on latent topic
extraction in a digital library. In George Buchanan, Masood Masoodian, and Sally
Cunningham, editors, Digital Libraries: Universal and Ubiquitous Access to
Information, volume 5362 of Lecture Notes in Computer Science, pages 93{103. Springer
Berlin / Heidelberg, 2008.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Ch</surname>
            . Aswani Kumar and
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Srinivas</surname>
          </string-name>
          .
          <article-title>Concept lattice reduction using fuzzy K-Means clustering</article-title>
          .
          <source>Expert Systems with Applications</source>
          ,
          <volume>37</volume>
          (
          <issue>3</issue>
          ):
          <volume>2696</volume>
          {
          <fpage>2704</fpage>
          ,
          <string-name>
            <surname>March</surname>
          </string-name>
          <year>2010</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Karen</surname>
            <given-names>S. K.</given-names>
          </string-name>
          <string-name>
            <surname>Cheung</surname>
            and
            <given-names>Douglas</given-names>
          </string-name>
          <string-name>
            <surname>Vogel</surname>
          </string-name>
          .
          <source>Complexity Reduction in Lattice-Based Information Retrieval. Information Retrieval</source>
          ,
          <volume>8</volume>
          (
          <issue>2</issue>
          ):
          <volume>285</volume>
          {
          <fpage>299</fpage>
          ,
          <string-name>
            <surname>April</surname>
          </string-name>
          <year>2005</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <given-names>Philipp</given-names>
            <surname>Cimiano</surname>
          </string-name>
          , Andreas Hotho, and
          <article-title>Ste en Staab. Learning concept hierarchies from text corpora using formal concept analysis</article-title>
          .
          <source>J. Artif. Int. Res.</source>
          ,
          <volume>24</volume>
          :
          <fpage>305</fpage>
          {
          <fpage>339</fpage>
          ,
          <year>August 2005</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          <article-title>4. V ctor Codocedo and Hernan Astudillo. No mining, no meaning: relating documents across repositories with ontology-driven information extraction. In Proceeding of the eighth ACM symposium on Document engineering</article-title>
          ,
          <source>DocEng '08</source>
          , pages
          <fpage>110</fpage>
          {
          <fpage>118</fpage>
          , New York, NY, USA,
          <year>2008</year>
          . ACM.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <given-names>Wisam</given-names>
            <surname>Dakka</surname>
          </string-name>
          , Panagiotis G. Ipeirotis, and
          <string-name>
            <surname>Kenneth</surname>
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Wood</surname>
          </string-name>
          .
          <article-title>Automatic construction of multifaceted browsing interfaces</article-title>
          .
          <source>In Proceedings of the 14th ACM international conference on Information and knowledge management</source>
          ,
          <source>CIKM '05</source>
          , pages
          <fpage>768</fpage>
          {
          <fpage>775</fpage>
          , New York, NY, USA,
          <year>2005</year>
          . ACM.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <given-names>Scott</given-names>
            <surname>Deerwester</surname>
          </string-name>
          , Susan T. Dumais, George W. Furnas,
          <string-name>
            <surname>Thomas</surname>
            <given-names>K.</given-names>
          </string-name>
          <string-name>
            <surname>Landauer</surname>
          </string-name>
          , and Richard Harshman.
          <article-title>Indexing by latent semantic analysis</article-title>
          .
          <source>Journal of the american society for Information Science</source>
          ,
          <volume>41</volume>
          (
          <issue>6</issue>
          ):
          <volume>391</volume>
          {
          <fpage>407</fpage>
          ,
          <year>1990</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Sergio</surname>
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Dias</surname>
            and
            <given-names>Newton J.</given-names>
          </string-name>
          <string-name>
            <surname>Vieira</surname>
          </string-name>
          .
          <article-title>Reducing the size of concept lattices: The JBOS Approach</article-title>
          .
          <source>In Proceedings of the 8ht international conference on Concept Lattices and their Applications</source>
          ,
          <source>CLA</source>
          <year>2010</year>
          , pages
          <fpage>80</fpage>
          {
          <fpage>91</fpage>
          ,
          <year>2010</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <given-names>Anabel</given-names>
            <surname>Fraga</surname>
          </string-name>
          and
          <string-name>
            <given-names>Juan</given-names>
            <surname>Llorens</surname>
          </string-name>
          .
          <article-title>Training initiative for new software/enterprise architects: An ontological approach</article-title>
          .
          <source>In WICSA, page 19. IEEE Computer Society</source>
          ,
          <year>2007</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <given-names>Petr</given-names>
            <surname>Gajdos</surname>
          </string-name>
          , Pavel Moravec, and
          <string-name>
            <given-names>Vaclav</given-names>
            <surname>Snasel</surname>
          </string-name>
          .
          <article-title>Concept lattice generation by singular value decomposition</article-title>
          .
          <source>In Vaclav Snasel and Radim Belohlavek</source>
          , editors,
          <source>International Workshop on Concept Lattices and their Applications (CLA)</source>
          , volume
          <volume>110</volume>
          <source>of CEUR Workshop Proceedings. CEUR-WS.org</source>
          ,
          <year>2004</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <given-names>Bernhard</given-names>
            <surname>Ganter</surname>
          </string-name>
          and
          <string-name>
            <given-names>Rudolf</given-names>
            <surname>Wille</surname>
          </string-name>
          .
          <source>Formal Concept Analysis: Mathematical Foundations</source>
          . Springer, Berlin/Heidelberg,
          <year>1999</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>Gene</surname>
            <given-names>H.</given-names>
          </string-name>
          <string-name>
            <surname>Golub and Charles F. Van Loan</surname>
          </string-name>
          .
          <source>Matrix computations (3rd ed.)</source>
          . Johns Hopkins University Press, Baltimore,
          <string-name>
            <surname>MD</surname>
          </string-name>
          , USA,
          <year>1996</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <string-name>
            <surname>Nicolas</surname>
            <given-names>Jay</given-names>
          </string-name>
          , Francois Kohler, and
          <string-name>
            <given-names>Amedeo</given-names>
            <surname>Napoli</surname>
          </string-name>
          .
          <article-title>Analysis of social communities with iceberg and stability-based concept lattices</article-title>
          .
          <source>In Proceedings of the 6th international conference on Formal concept analysis</source>
          ,
          <source>ICFCA'08</source>
          , pages
          <fpage>258</fpage>
          {
          <fpage>272</fpage>
          , Berlin, Heidelberg,
          <year>2008</year>
          . Springer-Verlag.
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13. John Kominek and
          <string-name>
            <given-names>Rick</given-names>
            <surname>Kazman</surname>
          </string-name>
          .
          <article-title>Accessing multimedia through concept clustering</article-title>
          .
          <source>In Proceedings of the SIGCHI conference on Human factors in computing systems, CHI '97</source>
          , pages
          <fpage>19</fpage>
          {
          <fpage>26</fpage>
          , New York, NY, USA,
          <year>1997</year>
          . ACM.
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          14.
          <string-name>
            <given-names>Sergei</given-names>
            <surname>Kuznetsov</surname>
          </string-name>
          .
          <article-title>Stability as an estimate of the degree of substantiation of hypotheses derived on the basis of operational similarity. nauchn. tekh. inf., ser.2 (automat</article-title>
          . document.
          <source>math. linguist.)</source>
          .
          <volume>12</volume>
          :
          <issue>21</issue>
          {
          <fpage>29</fpage>
          ,
          <year>1990</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          15.
          <string-name>
            <surname>Sergei</surname>
            <given-names>Kuznetsov</given-names>
          </string-name>
          , Sergei Obiedkov, and
          <string-name>
            <given-names>Camille</given-names>
            <surname>Roth</surname>
          </string-name>
          .
          <article-title>Reducing the representation complexity of lattice-based taxonomies</article-title>
          . In Uta Priss, Simon Polovina, and Richard Hill, editors, Conceptual Structures:
          <article-title>Knowledge Architectures for Smart Applications</article-title>
          , volume
          <volume>4604</volume>
          of Lecture Notes in Computer Science, pages
          <volume>241</volume>
          {
          <fpage>254</fpage>
          . Springer Berlin / Heidelberg,
          <year>2007</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          16.
          <string-name>
            <surname>Sergei</surname>
            <given-names>O.</given-names>
          </string-name>
          <string-name>
            <surname>Kuznetsov</surname>
          </string-name>
          .
          <article-title>On stability of a formal concept</article-title>
          .
          <source>Annals of Mathematics and Arti cial Intelligence</source>
          ,
          <volume>49</volume>
          :
          <fpage>101</fpage>
          {
          <fpage>115</fpage>
          ,
          <string-name>
            <surname>April</surname>
          </string-name>
          <year>2007</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          17.
          <string-name>
            <given-names>Uta</given-names>
            <surname>Priss</surname>
          </string-name>
          .
          <article-title>Formal concept analysis in information science</article-title>
          .
          <source>Annual Review of Information Science and Technology</source>
          ,
          <volume>40</volume>
          (
          <issue>1</issue>
          ):
          <volume>521</volume>
          {
          <fpage>543</fpage>
          ,
          <year>September 2007</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          18.
          <string-name>
            <given-names>Camille</given-names>
            <surname>Roth</surname>
          </string-name>
          and
          <string-name>
            <given-names>Paul</given-names>
            <surname>Bourgine</surname>
          </string-name>
          .
          <article-title>Lattice-based dynamic and overlapping taxonomies: The case of epistemic communities</article-title>
          .
          <source>SCIENTOMETRICS</source>
          ,
          <volume>69</volume>
          (
          <issue>2</issue>
          ):
          <volume>429</volume>
          {
          <fpage>447</fpage>
          ,
          <string-name>
            <surname>NOV</surname>
          </string-name>
          <year>2006</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          19.
          <string-name>
            <surname>Camille</surname>
            <given-names>Roth</given-names>
          </string-name>
          , Sergei Obiedkov, and
          <string-name>
            <given-names>Derrick</given-names>
            <surname>Kourie</surname>
          </string-name>
          .
          <article-title>Towards concise representation for taxonomies of epistemic communities</article-title>
          .
          <source>In Proceedings of the 4th international conference on Concept lattices and their applications</source>
          ,
          <source>CLA'06</source>
          , pages
          <fpage>240</fpage>
          {
          <fpage>255</fpage>
          , Berlin, Heidelberg,
          <year>2008</year>
          . Springer-Verlag.
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          20.
          <string-name>
            <surname>Vaclav</surname>
            <given-names>Snasel</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>Martin</given-names>
            <surname>Polovincak</surname>
          </string-name>
          , and
          <string-name>
            <surname>Hussam</surname>
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Dahwa</surname>
          </string-name>
          .
          <article-title>Concept lattice Reduction by Singular Value Decomposition</article-title>
          .
          <source>Proceedings of the Spring Young Researcher's Colloquium on Database and Information Systems</source>
          ,
          <year>2007</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>