<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Concept interestingness measures: a comparative study</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Sergei O. Kuznetsov</string-name>
          <email>skuznetsov@hse.ru</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Tatiana P. Makhalova</string-name>
          <email>t.makhalova@gmail.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>ISIMA, Complexe scientifique des C ́ezeaux</institution>
          ,
          <addr-line>63177 Aubi`ere Cedex</addr-line>
          ,
          <country country="FR">France</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>National Research University Higher School of Economics</institution>
          ,
          <addr-line>Kochnovsky pr. 3, Moscow 125319</addr-line>
          ,
          <country country="RU">Russia</country>
        </aff>
      </contrib-group>
      <fpage>59</fpage>
      <lpage>72</lpage>
      <abstract>
        <p>Concept lattices arising from noisy or high dimensional data have huge amount of formal concepts, which complicates the analysis of concepts and dependencies in data. In this paper, we consider several methods for pruning concept lattices and discuss results of their comparative study. c paper author(s), 2015. Published in Sadok Ben Yahia, Jan Konecny (Eds.): CLA 2015, pp. 59-72, ISBN 978-2-9544948-0-7, Blaise Pascal University, LIMOS laboratory, Clermont-Ferrand, 2015. Copying permitted only for private and academic purposes.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Introduction</title>
      <p>Formal Concept Analysis (FCA) underlies several methods for rule mining,
clustering and building taxonomies. When constructing a taxonomy one often deals
with high dimensional or/and noisy data which results in a huge amount of
formal concepts and dependencies given by implications and association rules. To
tackle this issue different approaches were proposed for selecting most important
or interesting concepts. In this paper we consider existing approaches which fall
into the following groups: pre-processing of a formal context, modification of the
closure operator, and concept filtering based on interestingness indices
(measures). We mostly focus on comparison of interestingness measures and study
their correlations.</p>
    </sec>
    <sec id="sec-2">
      <title>FCA framework</title>
      <p>
        Here we briefly recall FCA terminology [
        <xref ref-type="bibr" rid="ref20">20</xref>
        ]. A formal context is a triple (G, M, I),
where G is called a set objects, M is called a set attributes and I ⊆ G × M is a
relation called incidence relation, i.e. (g, m) ∈ I if the object g has the attribute
m. The derivation operators (·)0 are defined for A ⊆ G and B ⊆ M as follows:
A0 = {m ∈ M |∀g ∈ A : gIm}
      </p>
      <p>B0 = {g ∈ G|∀m ∈ B : gIm}
A0 is the set of attributes common to all objects of A and B0 is the set of objects
sharing all attributes of B. The double application of (·)0 is a closure operator,
i.e. (·)00 is extensive, idempotent and monotone. Sets A ⊆ G, B ⊆ M , such that
A = A00 and B = B00 are said to be closed.</p>
      <p>A (formal) concept is a pair (A, B), where A ⊆ G, B ⊆ M and A0 = B,
B0 = A. A is called the (formal) extent and B is called the (formal) intent of the
concept (A, B). A partial order 6 is defined on the set of concepts as follows:
(A, B) ≤ (C, D) iff A ⊆ C (D ⊆ B), a pair (A, B) is a subconcept of (C, D),
while (C, D) is a superconcept of (A, B).
3</p>
    </sec>
    <sec id="sec-3">
      <title>Methods for simplifying a lattice structure</title>
      <p>
        With the growth of the dimension of a context the size of a lattice can increase
exponentially, it becomes almost impossible to deal with the huge amount of
formal concepts. With this respect a wide variety of methods have been
proposed. Classification of them was presented in [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ]. Authors proposed to divide
techniques for lattice pruning into three classes: redundant information removal,
simplification, selection. In this paper, we consider also other classes of methods
and their application to concept pruning.
3.1
      </p>
      <sec id="sec-3-1">
        <title>Pre-processing</title>
        <p>
          Algorithms for concept lattice are time consuming. To decrease computation
costs one can reduce the size of a formal context. Cheung and Vogel [
          <xref ref-type="bibr" rid="ref13">13</xref>
          ] applied
Singular Value Decomposition (SVD) to obtain a low-rank approximation of
Term-Document matrix and construct concept lattice using pruned concepts.
Since this method is also computationally complex [
          <xref ref-type="bibr" rid="ref25">25</xref>
          ], alternative methods
such as spherical k-Means [
          <xref ref-type="bibr" rid="ref14">14</xref>
          ] and fuzzy k-Means [
          <xref ref-type="bibr" rid="ref17">17</xref>
          ], Non-negative Matrix
Decomposition [
          <xref ref-type="bibr" rid="ref33">33</xref>
          ] were proposed.
        </p>
        <p>Dimensionality reduction can dramatically decrease the computational load
and simplify the lattice structure, but in most cases it is very difficult to interpret
the obtained results.</p>
        <p>
          Another way to solve described problems without changing the dimension of
the context was proposed in [
          <xref ref-type="bibr" rid="ref18">18</xref>
          ], where an algorithm that significantly improves
the lattice structure by making small changes of context was presented. The
central notion of the method is the concept incomparability w.r.t. ≤ relation.
The goal of the proposed method is to diminish total incomparability of the
concepts in the lattice.
        </p>
        <p>The authors note that the result is close to that of fuzzy k-Means, but the
former is achieved with fewer context changes than required by the latter.
However, such transformations do not always lead to the decrease of a number of
formal concepts, the transformations of a context are aimed at increasing the
share of comparable concepts, thus this method does not ensure a significant
simplification of the lattice structure.</p>
        <p>
          Context pruning by clustering objects was introduced in [
          <xref ref-type="bibr" rid="ref15">15</xref>
          ]. The similarity
of objects is defined as the weighted sum of shared attributes. Thus, the original
context is replaced by the reduced one. Firstly, we need to assign weights wm
for each attribute m ∈ M . The similarity between objects is defined as weighted
sum of shared attributes.
        </p>
        <p>Objects are considered similar if sim(g, h) ≥ ε, where ε is a predefined
threshold. In order to avoid the generation of large clusters another threshold α was
proposed. Thus, the algorithm is an agglomerative clustering procedure, such
that at each step clusters are brought together if the similarity between them is
less than ε and the volume of clusters is less than α|G| objects.
3.2</p>
      </sec>
      <sec id="sec-3-2">
        <title>Reduction based on a background knowledge or predefined constraints</title>
        <p>
          Another approach to tackle computation and representation issues is to
determine constraints on the closure operator. It can be done using background
knowledge of attributes. In [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ] the extended closure operator was presented. It is
based on the notion of AD-formulas (attribute-dependency formulas), which
establish dependence of attributes and their relative importance. Put differently,
the occurrence of certain attributes implies that more important ones should
also occur. Concepts which do not satisfy this condition are not included in the
lattice.
        </p>
        <p>
          In [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ] a numerical approach to defining attribute importance was proposed.
The importance of a formal concept can be defined by various aggregation
functions (average, minimum, maximum) and different intent subsets (generator,
minimal generator or intent itself). It was shown [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ] that there is a
correspondence between this numerical approach and AD-formulas.
        </p>
        <p>
          Carpineto and Romano [
          <xref ref-type="bibr" rid="ref12">12</xref>
          ] considered document-term relation and proposed
to use a thesaurus of terms to prune the lattice. Two different attributes are
considered as same if there is a common ancestor in the hierarchy. To enrich the
set of attributes they used a thesaurus, but in general, it may be quite difficult
to establish such kind of relationship between arbitrary attributes.
        </p>
        <p>
          Computing concepts with extents exceeding a threshold was proposed in [
          <xref ref-type="bibr" rid="ref26">26</xref>
          ]
and studied in relation to frequent itemset mining in [
          <xref ref-type="bibr" rid="ref34">34</xref>
          ]. The main drawback
of this approach, called “iceberg lattice” mining, is missing rare and probably
interesting concepts.
        </p>
        <p>
          Several polynomial-time algorithms for computing Galois sub-hierarchies were
proposed, see [
          <xref ref-type="bibr" rid="ref3 ref9">9, 3</xref>
          ].
3.3
        </p>
      </sec>
      <sec id="sec-3-3">
        <title>Filtering concepts</title>
        <p>Selecting most interesting concepts by means of interestingness measures
(indices) is the most widespread way of dealing with the huge number of concepts.
The situation is aggravated by complexity of computing some indices. However,
this approach may be fruitful, since it provides flexible tools for exploration of
a derived taxonomy. In this section we consider different indices for filtering
formal concepts. These indices can be divided into the following groups:
measures designed to assess closed itemsets (formal concepts), arbitrary itemsets and
measures for assessing the membership in a basic level (a psychology-motivated
approach).</p>
      </sec>
      <sec id="sec-3-4">
        <title>Indices for formal concepts</title>
        <p>
          Stability Stability indices were introduced in [
          <xref ref-type="bibr" rid="ref27 ref28">27, 28</xref>
          ] and modified in [
          <xref ref-type="bibr" rid="ref29">29</xref>
          ]. One
distinguishes intensional and extensional stability. The first one allows estimating
the strength of dependence of an intent on each object of the respective extent.
Extensional stability is defined dually.
        </p>
        <p>Stabi (A, B) = | {C ⊆ A|C0 = B} |
2|A|</p>
        <p>
          The problem of computing stability is #P -complete [
          <xref ref-type="bibr" rid="ref28">28</xref>
          ] and hence it makes
this measure impractical for large contexts. In [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ] its Monte Carlo approximation
was introduced, a combination of Monte Carlo and upper bound estimate was
proposed in [
          <xref ref-type="bibr" rid="ref10">10</xref>
          ]. Since for large contexts the stability is close to 1 [
          <xref ref-type="bibr" rid="ref21">21</xref>
          ] the
logarithmic scale of stability (inducing the same ranking as stability) [
          <xref ref-type="bibr" rid="ref10">10</xref>
          ] is
often used:
        </p>
        <p>LStab (c) = −log2 (1 − Stab (c))
The bounds of stability are given by
Δmin (c) − log2 (|M |) ≤ −log2</p>
        <p>2−Δ(c,d) ≤ LStab (c) ≤ Δmin (c) ,</p>
        <p>X
d∈DD(c)
where Δmin (c) = mind∈DD(c)Δ (c, d), DD (c) is a set of all direct descendants
of c in the lattice and Δ (c, d) is the size of the set-difference between extents of
formal concepts c and d.</p>
        <p>In our experiments we used the bounds of logarithmic stability, because the
combined method is still computationally demanding.</p>
        <p>
          Concept Probability Stability of a formal concept may be interpreted as
probability of retaining its intent after removing some objects from the extent, taking
that all subsets of the extent have equal probability. In [
          <xref ref-type="bibr" rid="ref24">24</xref>
          ] it was noticed that
some interesting concepts with small number of object usually have low stability
value. To ensure selection of interesting infrequent closed patterns, the concept
probability was introduced. It is equivalent to the probability of a concept
introduced earlier by R. Emilion [
          <xref ref-type="bibr" rid="ref19">19</xref>
          ].
        </p>
        <p>The probability that an arbitrary object has all attributes from the set B is
defined as follows
pB =</p>
        <p>Y pm
m∈B</p>
        <p>Concept probability is defined as the probability of B being closed:
p (B = B00) =
n n "
X p (|B0| = k, B = B00) = X
k=0 k=0
pkB (1 − pB)n−k Y</p>
        <p>k
1 − pm
m∈/B
#
where n = |G|.</p>
        <p>The concept probability has the following probabilistic components: the
occurrence of each attribute from B in all k objects, the absence of at least one
attribute from B in other objects and the absence of other attributes shared by
all k objects.</p>
        <p>
          Robustness Another probabilistic approach to assessing a formal concept was
proposed in [
          <xref ref-type="bibr" rid="ref35">35</xref>
          ]. Robustness is defined as the probability of a formal concept
intent remaining closed while deleting objects, where every object of a formal
context is retained with probability α. Then for a formal concept c = (A, B) the
robustness is given as follows:
r (c, α) =
        </p>
        <p>
          X (−1)|Bd|−|Bc| (1 − α)|Ac|−|Ad|
d c
Separation The separation index was considered in [
          <xref ref-type="bibr" rid="ref24">24</xref>
          ]. The main idea behind
this measure is to describe the area covered by a formal concept among all
nonzero elements in the corresponding rows and columns of the formal context.
Thus, the value characterizes how specific is the relationship between objects
and attributes of the concept with respect to the formal context.
        </p>
        <p>
          s (A, B) = Pg∈A |g0| + P|Am||∈BB| |m0| − |A||B|
Basic Level Metrics The group of so-called “basic level” measures was
considered by Belohlavek and Trnecka [
          <xref ref-type="bibr" rid="ref6 ref7">6, 7</xref>
          ]. These measures were proposed to formalize
the existing psychological approach to defining basic level of a concept [
          <xref ref-type="bibr" rid="ref31">31</xref>
          ].
Similarity approach (S) A similarity approach to basic level was proposed in
[
          <xref ref-type="bibr" rid="ref32">32</xref>
          ] and subsequently formalized and applied to FCA in [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ]. The authors defined
basic level as combination of three fuzzy functions that correspond to formalized
properties outlined by Rosch: high cohesion of concepts, considerably greater
cohesion with respect to upper neighbor and a slightly less cohesion with respect
to lower neighbors. The membership degree of a basic level is defined as follows:
        </p>
        <p>
          BLS = coh∗∗ (A, B) ⊗ coh∗u∗n (A, B) ⊗ cohl∗n∗ (A, B) ,
where αi is a fuzzy function that corresponds to the conditions defined above,
⊗ is t-norm [
          <xref ref-type="bibr" rid="ref23">23</xref>
          ].
        </p>
        <p>A cohesion of a formal concept is a measure of pairwise similarity of all object
in the extent. Various similarity measures can be used for cohesion functions:
simSMC (B1, B2) = |B1 ∩ B2| + |M − (B1 ∪ B2) |</p>
        <p>|M |
simJ (B1, B2) = |B1 ∩ B2|
|B1 ∪ B2|</p>
        <p>The first similarity index simSMC takes into account the number of
common attributes, while Jaccard similarity simJ takes exactly the proportion of
attributes shared by two sets. There are two ways to compute cohesion of formal
concepts: taking average or minimal similarity among sets of attributes of the
concept extent, the formulas are represented below (for average and minimal
similarity respectively).</p>
        <p>coh.a.. (A, B) =
coh.m.. (A, B) =</p>
        <p>P x1,x2 ⊆A,x16=x2 sim... (x01, x02)</p>
        <p>|A| (|A| − 1) /2
min sim... (x01, x02)
x1,x2∈A
The Rosch’s properties for upper and lower neighbors take the following forms:
coh.a.∗.,un (A, B) = 1 −
coh.a.∗.,ln (A, B) =</p>
        <p>Pc∈UN(A,B) coh.∗.. (c) /coh.∗.. (A, B)</p>
        <p>|U N (A, B) |
Pc∈LN(A,B) coh.∗.. (A, B) /coh.∗.. (c)</p>
        <p>|LN (A, B) |
coh.m..∗,un (A, B) = 1 −
coh.m..∗,ln (A, B) =</p>
        <p>max
c∈UN(A,B)
min
c∈LN(A,B)</p>
        <p>coh.∗.. (c) /coh.∗.. (A, B)
coh.∗.. (A, B) /coh.∗.. (c)
where U N (A, B) and LN (A, B) are upper and lower neighbors of a formal
concept (A, B) respectively.</p>
        <p>As the authors noted, experiments revealed that the type of cohesion function
does not affect the result, while the choice of similarity measure can greatly
affect the outcome. More than that, in some cases upper (lower) neighbors may
have higher (lower) cohesion than the formal concept itself (for example, some
boundary cases, when a neighbors’s extent (an intent) consists of identical rows
(columns) of a formal context). To tackle this issue of non-monotonic neighbors
w.r.t. similarity function authors proposed to take coh.∗.∗.,ln and coh.∗.∗.,un as 0, if
the rate of non-monotonic neighbors is larger that a threshold.</p>
        <p>
          In our experiments we used the following notation: SMC∗∗ and J∗∗, where
the first star is replaced by a cohesion type, the second one is replaced by the
type of a similarity function. Below, we consider another four metrics that were
introduced in [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ].
        </p>
        <p>Predictability approach (P) Predictability of a formal concept is computed in a
quite similar way to BLS. A cohesion function is replaced by a predictability
function:</p>
        <p>
          P (A, B) = pred∗∗ (A, B) ⊗ pred∗u∗n (A, B) ⊗ predl∗n∗ (A, B)
The main idea behind this approach is to assign high score to concept (A, B)
with low conditional entropy of the presence of attributes not in B in intents of
objects from A (i.e., requiring few attributes outside B in objects from A)[
          <xref ref-type="bibr" rid="ref7">7</xref>
          ]:
pred (A, B) = 1 −
        </p>
        <p>X
y∈M−B</p>
        <p>E (I [hx, yi ∈ I] |I [x ∈ A])
|M − B|
.</p>
        <p>
          Cue Validity (CV), Category Feature Collocation (CFC), Category Utility (CU)
The following measures based on the conditional probability of object g ∈ A
given that y ⊆ g0 were introduced in [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ]:
        </p>
        <p>CV (A, B) = X P (A|y0) =
y∈B</p>
        <p>A
X | |
y∈B |y0|
CF C (A, B) = X p (A|y0) p (y0|A) =
y∈M
y∈M
X |A ∩ y0| |A ∩ y0|
|y0| |A|
CU (A, B) = p (A) X hp (y0|A)2 − p (y0)2i = |A| X
y∈M
|G| y∈M
"
|A ∩ y0|
|y0|
2
−
|y0|
|G|
2#</p>
        <p>
          The main intuition behind CV is to express probability of extent given
attributes from intent, CFC index takes into account the relationship between all
attributes of the concept and intent of the formal concept, while CU evaluates
how much an attribute in an intent is characteristic for a given concept rather
than for the whole context [
          <xref ref-type="bibr" rid="ref36">36</xref>
          ].
        </p>
      </sec>
      <sec id="sec-3-5">
        <title>Metrics for arbitrary itemsets</title>
        <p>Frequency(support) It is one of the most popular measures in the theory of
pattern mining. According to this index the most “interesting” concepts are
frequent ones (having high support). For an arbitrary formal concept the support
is defined as follows</p>
        <p>
          A
supp (A, B) = | |
|G|
The support provides efficient level-wise algorithms for constructing semilattices
since it exhibits anti-monotonicity (a priori property [
          <xref ref-type="bibr" rid="ref2 ref30">2, 30</xref>
          ]):
        </p>
        <p>
          B1 ⊂ B2 → supp (B1) ≥ supp (B2)
Lift In the previous section different methods with background knowledge were
considered. Another way to add additional knowledge to data is proposed in [
          <xref ref-type="bibr" rid="ref11">11</xref>
          ].
Under assumption of attributes independence it is possible to compute individual
frequencies of attributes and take their product as the expected frequency. The
ratio of the observed frequency to its expectation is defined as lift. The lift of a
formal concept (A, B) is defined as follows:
lif t (B) = Qb∈B P (b0)
        </p>
        <p>P (A)</p>
        <p>
          |A|/|G|
= Qb∈B |b0|/|G|
Collective Strength The collective strength [
          <xref ref-type="bibr" rid="ref1">1</xref>
          ] combines ideas of comparing the
observed data and expectation under the assumption of independence of
attributes. To calculate this measure for a formal concept (A, B) one needs to
define for B the set of objects VB that has at least one attribute in B, but not
all of them at once. Denote q = Qb∈B supp (b0) and supp (VB) = v, the collective
strength of a formal concept has the following form:
cs (B) = 1 − v
v
        </p>
        <p>q
1 − q
4</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Experiments</title>
      <p>In this section, we compare measures with respect to their ability to help selecting
most interesting concepts and filtering concepts coming from noisy datasets. For
both goals, one is interested in a ranking of concepts rather than in particular
values of the measures.
4.1</p>
      <sec id="sec-4-1">
        <title>Formal Concept Mining</title>
        <p>
          Usually concept lattices constructed from empirical data have huge amount of
formal concepts, many of them being redundant, excessive and useless. In this
connection the measures can be used to estimate how meaningful a concept is.
Since the “interestingness” of a concept is a fairly subjective measure, the correct
comparison of indices in terms of ability to select meaningful ones is impossible.
With this respect we focus on similarity of indices described above. To identify
how similar indices are, we use the Kendall tau correlation coefficient [
          <xref ref-type="bibr" rid="ref22">22</xref>
          ]. Put
differently, we consider pairwise similarity of two lists of the same concepts that
are ordered by values of the chosen indices. A set of strongly correlated measures
can be replaced by one with the lowest computational complexity.
        </p>
        <p>We randomly generated 100 formal contexts of random sizes. The number of
attributes was in range between 10 and 40, while the number of objects varied
from 10 to 70. For generated contexts we calculated pairwise Kendall tau for
all indices of each context.The averaged values of correlations coefficients are
represented in Table 1.</p>
        <p>
          In [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ] it was shown that the CU, CFC and CV are correlated, while S and
P are not strongly correlated to other metrics. The results of our simulations
allow us to conclude that CU, CFC and CV are also pairwise correlated to
separation and support. Moreover, support is strongly correlated to separation
and probability. Since the computational complexity of support is less than that
of separation and probability, it is preferable to use support. It is worth noting
that predictability (P) and robustness are not correlated to any other metrics
and hence they can not be replaced by the metrics introduced so far.
        </p>
        <p>
          Thus, based on the correlation analysis, it is possible to reduce
computationally complexity by choosing the most easily computable index within the class
of correlated metrics.
In practice, we often have to deal with noisy data. In this case, the number of
formal concepts can be very large and the lattice structure becomes too
complicated [
          <xref ref-type="bibr" rid="ref24">24</xref>
          ]. To test the ability to filter out noise we took 5 lattices of different
structure. Four of them are quite simple (Fig. 1) and the fifth one is the
binarized fragment of the Mushroom data set 1 on 500 objects and 14 attributes, its
concept lattice consists of 54 formal concepts.
        </p>
        <p>(a)
(b)
(c)
(d)
1 https://archive.ics.uci.edu/ml/datasets/Mushroom</p>
        <p>For a generated 0-1 datatable we changed table elements (0 to 1 and 1 to
0) with a given probability. The rate of noise (the probability of replacement)
varied in the range from 0.05 to 0.5. We test the ability of a measure to filter
redundant concepts in terms of precision and recall. For top-n (w.r.t. a measure)
formal concepts, the recall and precision are defined as follows:
recalltop−n = |original conceptstop−n|</p>
        <p>|original concepts|
precisiontop−n = |original conceptstop−n|</p>
        <p>|top − n concepts|</p>
        <p>Figures 2 show the ROC curve for the measures. The curves that are close
to the left upper corner correspond to the most powerful measures.</p>
        <p>The best and most stable results correspond to the high estimate of stability
(stabilityh). The similar precision has the lower estimate of stability (Table 2),
whereas precision of separation and probability depends on the proportion of
noise and lattice structure as well. The measures of basic level that utilize
similarity and predictability approaches become zero for some concepts. The rate
of vanished concepts (including original ones) increases as the noise probability
gets bigger. In our study we take such concepts as “false negative”, so in this
case ROC curves do not pass through the point (1,1). More than that, recall and
precision are unstable with respect to the noise rate and lattice structure. This
group of measures is inappropriate for noise filtering.</p>
        <p>The other basic level measures, such as CU, CFC and CV, demonstrate much
better recall compared to previous ones. However, in general the precision of CU,
CFC and CV is determined by lattice structure (Table 2).</p>
        <p>Frequency has the highest precision among the indices that are applicable for
the assessment of arbitrary sets of attributes. Frequency is stable with respect
to the noise rate, but can vary under different lattice structures. For the lift
and the collective strength precision depends on the lattice structure, and the
collective strength also has quite unstable recall.</p>
        <p>Precision of robustness depends on both lattice structure and value of α
(Fig. 2). In our study we have got the highest precision for α close to 0.5.</p>
        <p>Thus, the most preferred metrics for noise filtering are stability estimates,
CV, frequency and robustness (where α is greater than 0.4).</p>
        <p>
          In [
          <xref ref-type="bibr" rid="ref24">24</xref>
          ] it was noticed that the combination of the indices can improve the
filtering power of indices. In this regard, we have studied top-n concepts selected
by pairwise combination of measures. As it was shown by the experiments, the
combination of measures may improve recall of the top-n set, while precision
gets lower with respect to a more accurate measure. Figure 3 shows recall and
precision of different combination of measures. In the best case it is possible to
improve the recall, the precision on small sets of top-n concepts is lower than
the precision of one measure by itself.
5
        </p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>Conclusion</title>
      <p>In this paper we have considered various methods for selecting interesting
concepts and noise reduction. We focused on the most promising and well
interpretable approach based on interestingness measures of concepts. Since
“interestingness” of a concept is a subjective measure, we have compared several
measures known in the literature and identified groups of most correlated ones. CU,
CFC, CV, separation and frequency make up the first group. Frequency is
correlated to separation and probability.</p>
      <p>Another part of our experiments was focused on the noise filtering. We have
found that the stability estimates work perfectly with data of various noise rate
and different structure of the original lattice. Robustness and 3 of basic level
metrics (cue validity, category utility and category feature collocation approaches)
could also be applied to noise reduction. The combination of measures can also
improve the recall, but only in the case of high noise rate.</p>
      <sec id="sec-5-1">
        <title>Acknowledgments</title>
        <p>The authors were supported by the project “Mathematical Models, Algorithms,
and Software Tools for Mining of Structural and Textual Data” supported by
the Basic Research Program of the National Research University Higher School
of Economics.</p>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Aggarwal</surname>
            ,
            <given-names>C.C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yu</surname>
            ,
            <given-names>P.S.:</given-names>
          </string-name>
          <article-title>A new framework for itemset generation</article-title>
          .
          <source>In: Proceedings of the seventeenth ACM SIGACT-SIGMOD-SIGART symposium on Principles of database systems</source>
          . pp.
          <fpage>18</fpage>
          -
          <lpage>24</lpage>
          . ACM (
          <year>1998</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Agrawal</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Srikant</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          , et al.:
          <article-title>Fast algorithms for mining association rules</article-title>
          .
          <source>In: Proc. 20th int. conf. very large data bases, VLDB</source>
          . vol.
          <volume>1215</volume>
          , pp.
          <fpage>487</fpage>
          -
          <lpage>499</lpage>
          (
          <year>1994</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3. Ar´evalo, G.,
          <string-name>
            <surname>Berry</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Huchard</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Perrot</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sigayret</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Performances of galois sub-hierarchy-building algorithms</article-title>
          .
          <source>In: Formal Concept Analysis</source>
          , pp.
          <fpage>166</fpage>
          -
          <lpage>180</lpage>
          . Springer (
          <year>2007</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Babin</surname>
            ,
            <given-names>M.A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kuznetsov</surname>
            ,
            <given-names>S.O.</given-names>
          </string-name>
          :
          <article-title>Approximating concept stability</article-title>
          . In: Domenach,
          <string-name>
            <given-names>F.</given-names>
            ,
            <surname>Ignatov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            ,
            <surname>Poelmans</surname>
          </string-name>
          ,
          <string-name>
            <surname>J</surname>
          </string-name>
          . (eds.)
          <source>Formal Concept Analysis. Lecture Notes in Computer Science</source>
          , vol.
          <volume>7278</volume>
          , pp.
          <fpage>7</fpage>
          -
          <lpage>15</lpage>
          . Springer Berlin Heidelberg (
          <year>2012</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Belohlavek</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Macko</surname>
          </string-name>
          , J.:
          <article-title>Selecting important concepts using weights</article-title>
          . In: Valtchev,
          <string-name>
            <given-names>P.</given-names>
            ,
            <surname>Jschke</surname>
          </string-name>
          ,
          <string-name>
            <surname>R</surname>
          </string-name>
          . (eds.)
          <source>Formal Concept Analysis, Lecture Notes in Computer Science</source>
          , vol.
          <volume>6628</volume>
          , pp.
          <fpage>65</fpage>
          -
          <lpage>80</lpage>
          . Springer Berlin Heidelberg (
          <year>2011</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Belohlavek</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Trnecka</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Basic level of concepts in formal concept analysis</article-title>
          . In: Domenach,
          <string-name>
            <given-names>F.</given-names>
            ,
            <surname>Ignatov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            ,
            <surname>Poelmans</surname>
          </string-name>
          ,
          <string-name>
            <surname>J</surname>
          </string-name>
          . (eds.)
          <source>Formal Concept Analysis, Lecture Notes in Computer Science</source>
          , vol.
          <volume>7278</volume>
          , pp.
          <fpage>28</fpage>
          -
          <lpage>44</lpage>
          . Springer Berlin Heidelberg (
          <year>2012</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Belohlavek</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Trnecka</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Basic level in formal concept analysis: Interesting concepts and psychological ramifications</article-title>
          .
          <source>In: Proceedings of the Twenty-Third International Joint Conference on Artificial Intelligence</source>
          . pp.
          <fpage>1233</fpage>
          -
          <lpage>1239</lpage>
          . IJCAI '13, AAAI Press (
          <year>2013</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>Belohlavek</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Vychodil</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          :
          <article-title>Formal concept analysis with background knowledge: attribute priorities</article-title>
          .
          <source>Systems, Man, and Cybernetics</source>
          , Part C:
          <article-title>Applications</article-title>
          and Reviews,
          <source>IEEE Transactions on 39(4)</source>
          ,
          <fpage>399</fpage>
          -
          <lpage>409</lpage>
          (
          <year>2009</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Berry</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Huchard</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>McConnell</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sigayret</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Spinrad</surname>
          </string-name>
          , J.:
          <article-title>Efficiently computing a linear extension of the sub-hierarchy of a concept lattice</article-title>
          . In: Ganter,
          <string-name>
            <given-names>B.</given-names>
            ,
            <surname>Godin</surname>
          </string-name>
          ,
          <string-name>
            <surname>R</surname>
          </string-name>
          . (eds.)
          <source>Formal Concept Analysis, Lecture Notes in Computer Science</source>
          , vol.
          <volume>3403</volume>
          , pp.
          <fpage>208</fpage>
          -
          <lpage>222</lpage>
          . Springer Berlin Heidelberg (
          <year>2005</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Buzmakov</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kuznetsov</surname>
            ,
            <given-names>S.O.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Napoli</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Scalable estimates of concept stability</article-title>
          . In: Glodeanu,
          <string-name>
            <given-names>C.</given-names>
            ,
            <surname>Kaytoue</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            ,
            <surname>Sacarea</surname>
          </string-name>
          , C. (eds.)
          <source>Formal Concept Analysis, Lecture Notes in Computer Science</source>
          , vol.
          <volume>8478</volume>
          , pp.
          <fpage>157</fpage>
          -
          <lpage>172</lpage>
          . Springer International Publishing (
          <year>2014</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>Cabena</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Choi</surname>
            ,
            <given-names>H.H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kim</surname>
            ,
            <given-names>I.S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Otsuka</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Reinschmidt</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Saarenvirta</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          :
          <article-title>Intelligent miner for data applications guide</article-title>
          .
          <source>IBM RedBook SG24-5252-00</source>
          <volume>173</volume>
          (
          <year>1999</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <string-name>
            <surname>Carpineto</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Romano</surname>
          </string-name>
          , G.:
          <article-title>A lattice conceptual clustering system and its application to browsing retrieval</article-title>
          .
          <source>Machine Learning</source>
          <volume>24</volume>
          (
          <issue>2</issue>
          ),
          <fpage>95</fpage>
          -
          <lpage>122</lpage>
          (
          <year>1996</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13.
          <string-name>
            <surname>Cheung</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Vogel</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          :
          <article-title>Complexity reduction in lattice-based information retrieval</article-title>
          .
          <source>Information Retrieval</source>
          <volume>8</volume>
          (
          <issue>2</issue>
          ),
          <fpage>285</fpage>
          -
          <lpage>299</lpage>
          (
          <year>2005</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          14.
          <string-name>
            <surname>Dhillon</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Modha</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          :
          <article-title>Concept decompositions for large sparse text data using clustering</article-title>
          .
          <source>Machine Learning</source>
          <volume>42</volume>
          (
          <issue>1-2</issue>
          ),
          <fpage>143</fpage>
          -
          <lpage>175</lpage>
          (
          <year>2001</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          15.
          <string-name>
            <surname>Dias</surname>
            ,
            <given-names>S.M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Vieira</surname>
          </string-name>
          , N.:
          <article-title>Reducing the size of concept lattices: The JBOS approach</article-title>
          .
          <source>In: Proceedings of the 7th International Conference on Concept Lattices and Their Applications</source>
          , Sevilla, Spain,
          <source>October 19-21</source>
          ,
          <year>2010</year>
          . pp.
          <fpage>80</fpage>
          -
          <lpage>91</lpage>
          (
          <year>2010</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          16.
          <string-name>
            <surname>Dias</surname>
            ,
            <given-names>S.M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Vieira</surname>
            ,
            <given-names>N.J.:</given-names>
          </string-name>
          <article-title>Concept lattices reduction: Definition, analysis and classification</article-title>
          .
          <source>Expert Systems with Applications</source>
          <volume>42</volume>
          (
          <issue>20</issue>
          ),
          <fpage>7084</fpage>
          -
          <lpage>7097</lpage>
          (
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          17.
          <string-name>
            <surname>Dobˇsa</surname>
          </string-name>
          , J., Dalbelo-Baˇsi´c, B.:
          <article-title>Comparison of information retrieval techniques: latent semantic indexing and concept indexing</article-title>
          .
          <source>Journal of Inf. and Organizational Sciences</source>
          <volume>28</volume>
          (
          <issue>1-2</issue>
          ),
          <fpage>1</fpage>
          -
          <lpage>17</lpage>
          (
          <year>2004</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          18. Du¨ntsch, I.,
          <string-name>
            <surname>Gediga</surname>
          </string-name>
          , G.:
          <article-title>Simplifying contextual structures</article-title>
          . In: Kryszkiewicz,
          <string-name>
            <given-names>M.</given-names>
            ,
            <surname>Bandyopadhyay</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            ,
            <surname>Rybinski</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            ,
            <surname>Pal</surname>
          </string-name>
          , S.K. (eds.)
          <source>Pattern Recognition and Machine Intelligence, Lecture Notes in Computer Science</source>
          , vol.
          <volume>9124</volume>
          , pp.
          <fpage>23</fpage>
          -
          <lpage>32</lpage>
          . Springer International Publishing (
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          19.
          <string-name>
            <surname>Emilion</surname>
          </string-name>
          , R.:
          <article-title>Concepts of a discrete random variable</article-title>
          . In: Brito,
          <string-name>
            <given-names>P.</given-names>
            ,
            <surname>Cucumel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            ,
            <surname>Bertrand</surname>
          </string-name>
          , P.,
          <string-name>
            <surname>de Carvalho</surname>
          </string-name>
          , F. (eds.)
          <source>Selected Contributions in Data Analysis and Classification</source>
          , pp.
          <fpage>247</fpage>
          -
          <lpage>258</lpage>
          . Studies in Classification,
          <source>Data Analysis, and Knowledge Organization</source>
          , Springer Berlin Heidelberg (
          <year>2007</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          20.
          <string-name>
            <surname>Ganter</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wille</surname>
          </string-name>
          , R.:
          <article-title>Contextual attribute logic</article-title>
          . In: Tepfenhart,
          <string-name>
            <given-names>W.</given-names>
            ,
            <surname>Cyre</surname>
          </string-name>
          , W. (eds.)
          <source>Conceptual Structures: Standards and Practices, Lecture Notes in Computer Science</source>
          , vol.
          <volume>1640</volume>
          , pp.
          <fpage>377</fpage>
          -
          <lpage>388</lpage>
          . Springer Berlin Heidelberg (
          <year>1999</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          21.
          <string-name>
            <surname>Jay</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kohler</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Napoli</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Analysis of social communities with iceberg and stability-based concept lattices</article-title>
          . In: Medina,
          <string-name>
            <given-names>R.</given-names>
            ,
            <surname>Obiedkov</surname>
          </string-name>
          , S. (eds.)
          <source>Formal Concept Analysis, Lecture Notes in Computer Science</source>
          , vol.
          <volume>4933</volume>
          , pp.
          <fpage>258</fpage>
          -
          <lpage>272</lpage>
          . Springer Berlin Heidelberg (
          <year>2008</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          22.
          <string-name>
            <surname>Kendall</surname>
            ,
            <given-names>M.G.</given-names>
          </string-name>
          :
          <article-title>A new measure of rank correlation</article-title>
          . Biometrika pp.
          <fpage>81</fpage>
          -
          <lpage>93</lpage>
          (
          <year>1938</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          23.
          <string-name>
            <surname>Klement</surname>
            ,
            <given-names>E.P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mesiar</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pap</surname>
          </string-name>
          , E.: Triangular norms. Springer Netherlands (
          <year>2000</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          24.
          <string-name>
            <surname>Klimushkin</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Obiedkov</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Roth</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          :
          <article-title>Approaches to the selection of relevant concepts in the case of noisy data</article-title>
          . In: Kwuida,
          <string-name>
            <given-names>L.</given-names>
            ,
            <surname>Sertkaya</surname>
          </string-name>
          ,
          <string-name>
            <surname>B</surname>
          </string-name>
          . (eds.)
          <source>Formal Concept Analysis, Lecture Notes in Computer Science</source>
          , vol.
          <volume>5986</volume>
          , pp.
          <fpage>255</fpage>
          -
          <lpage>266</lpage>
          . Springer Berlin Heidelberg (
          <year>2010</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          25.
          <string-name>
            <surname>Kumar</surname>
            ,
            <given-names>C.A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Srinivas</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          :
          <article-title>Latent semantic indexing using eigenvalue analysis for efficient information retrieval</article-title>
          .
          <source>Int. J. Appl. Math. Comput. Sci</source>
          <volume>16</volume>
          (
          <issue>4</issue>
          ),
          <fpage>551</fpage>
          -
          <lpage>558</lpage>
          (
          <year>2006</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          26.
          <string-name>
            <surname>Kuznetsov</surname>
            ,
            <given-names>S.O.:</given-names>
          </string-name>
          <article-title>Interpretation on graphs and complexity characteristics of a search for specific patterns</article-title>
          .
          <source>Automatic Documentation and Mathematical Linguistics</source>
          <volume>24</volume>
          (
          <issue>1</issue>
          ),
          <fpage>37</fpage>
          -
          <lpage>45</lpage>
          (
          <year>1989</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          27.
          <string-name>
            <surname>Kuznetsov</surname>
            ,
            <given-names>S.O.</given-names>
          </string-name>
          :
          <article-title>Stability as an estimate of degree of substantiation of hypotheses derived on the basis of operational similarity</article-title>
          .
          <source>Nauchn</source>
          . Tekh. Inf.,
          <source>Ser</source>
          .
          <volume>2</volume>
          (
          <issue>12</issue>
          ),
          <fpage>21</fpage>
          -
          <lpage>29</lpage>
          (
          <year>1990</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          28.
          <string-name>
            <surname>Kuznetsov</surname>
            ,
            <given-names>S.O.</given-names>
          </string-name>
          :
          <article-title>On stability of a formal concept</article-title>
          .
          <source>Annals of Mathematics and Artificial Intelligence</source>
          <volume>49</volume>
          (
          <issue>1-4</issue>
          ),
          <fpage>101</fpage>
          -
          <lpage>115</lpage>
          (
          <year>2007</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          29.
          <string-name>
            <surname>Kuznetsov</surname>
            ,
            <given-names>S.O.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Obiedkov</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Roth</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          :
          <article-title>Reducing the representation complexity of lattice-based taxonomies</article-title>
          . In: Conceptual Structures:
          <article-title>Knowledge Architectures for Smart Applications</article-title>
          , pp.
          <fpage>241</fpage>
          -
          <lpage>254</lpage>
          . Springer Berlin Heidelberg (
          <year>2007</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          30.
          <string-name>
            <surname>Mannila</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Toivonen</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Verkamo</surname>
            ,
            <given-names>A.I.</given-names>
          </string-name>
          :
          <article-title>Efficient algorithms for discovering association rules</article-title>
          .
          <source>In: KDD-94: AAAI workshop on Knowledge Discovery in Databases</source>
          . pp.
          <fpage>181</fpage>
          -
          <lpage>192</lpage>
          (
          <year>1994</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref31">
        <mixed-citation>
          31.
          <string-name>
            <surname>Murphy</surname>
            ,
            <given-names>G.L.</given-names>
          </string-name>
          :
          <article-title>The big book of concepts</article-title>
          .
          <source>MIT</source>
          press (
          <year>2002</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref32">
        <mixed-citation>
          32.
          <string-name>
            <surname>Rosch</surname>
          </string-name>
          , E.: Principles of categorization pp.
          <fpage>27</fpage>
          -
          <lpage>48</lpage>
          (
          <year>1978</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref33">
        <mixed-citation>
          33.
          <string-name>
            <surname>Snasel</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Polovincak</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Abdulla</surname>
            ,
            <given-names>H.M.D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Horak</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          :
          <article-title>On concept lattices and implication bases from reduced contexts</article-title>
          .
          <source>In: ICCS</source>
          . pp.
          <fpage>83</fpage>
          -
          <lpage>90</lpage>
          (
          <year>2008</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref34">
        <mixed-citation>
          34.
          <string-name>
            <surname>Stumme</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Taouil</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bastide</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pasquier</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lakhal</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          :
          <article-title>Computing iceberg concept lattices with titanic</article-title>
          .
          <source>Data Knowl. Eng</source>
          .
          <volume>42</volume>
          (
          <issue>2</issue>
          ),
          <fpage>189</fpage>
          -
          <lpage>222</lpage>
          (
          <year>Aug 2002</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref35">
        <mixed-citation>
          35.
          <string-name>
            <surname>Tatti</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Moerchen</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Calders</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          :
          <article-title>Finding robust itemsets under subsampling</article-title>
          .
          <source>ACM Transactions on Database Systems (TODS) 39(3)</source>
          ,
          <volume>20</volume>
          (
          <year>2014</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref36">
        <mixed-citation>
          36.
          <string-name>
            <surname>Zeigenfuse</surname>
            ,
            <given-names>M.D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lee</surname>
            ,
            <given-names>M.D.:</given-names>
          </string-name>
          <article-title>A comparison of three measures of the association between a feature and a concept</article-title>
          .
          <source>In: Proceedings of the 33rd Annual Conference of the Cognitive Science Society</source>
          . pp.
          <fpage>243</fpage>
          -
          <lpage>248</lpage>
          (
          <year>2011</year>
          )
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>