<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>AA FFoorrmmaalCloCnocnepcteAptnaAlynsiaslAyspipsrAoapchprtooaDcihscotvoer AssociatioDniRscuolveserfrom Data Association Rules from Data</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Mondher Maddouri</string-name>
          <email>mondher.maddouri@fst.rnu.tn</email>
        </contrib>
      </contrib-group>
      <fpage>10</fpage>
      <lpage>21</lpage>
      <abstract>
        <p>The Discovery of association rules is a non-supervised task of data mining. Its mostly hard step is to look for the frequent itemsets embedded into large amounts of data. Based on the theory of Formal Concept Analysis, we suggest that the notion of Formal Concept generalizes the notion of itemset, since it takes into account the itemset (as the intent) and the support (as the cardinality of the extent). Accordingly, we propose a new approach to mine interesting item-sets as the optimal concepts covering a binary table (concept coverage). This approach uses a new quality-criteria of a rule: the gain that generalizes the support criteria.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1 Introduction</title>
      <p>
        The volume of stored data increases rapidly. In the past, it doubles each three
centuries. Nowadays, it doubles each 72 days. The analysis of huge volumes of data
by computers leads to the emerging field of Data-Mining and Knowledge-Discovery.
This includes techniques to create knowledge from data sets serving for classification,
clustering, prediction, estimation and optimization tasks [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ].
      </p>
      <p>
        The notion of “Concept” as a couple of intent and extent (serving to represent
nuggets of knowledge) was introduced and formalized many decades ago [
        <xref ref-type="bibr" rid="ref2 ref3 ref4">2, 3, 4</xref>
        ].
Recently, its use in Data-Mining and Knowledge-Discovery is in growth [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. Our
ConceptMiner project (a development environment in MS-Visual C++) aims to design
new heuristic algorithms for Data-Mining based on Formal Concept Analysis.
      </p>
      <p>
        In this paper, we study the clustering problem that we resolve by the induction of
association rules. In fact, the experts (biologists) want to know which sub-sequences
of proteins that appear together in Tall Like Receptor (TLR) sequences or even in
Similar TLR sequences [
        <xref ref-type="bibr" rid="ref7">8</xref>
        ].
      </p>
      <p>In section 2, we resume the basic concepts of association rules and we outline the
well-known techniques. In section 3, we introduce the basic definitions of Formal
Concept Analysis that we use later. In section 4, we give two heuristic algorithms to
calculate the optimal Item-sets from rows of data. Illustrative examples are given
throughout the paper. Finally, we present our concluding remarks and the future
perspectives of this work.</p>
    </sec>
    <sec id="sec-2">
      <title>2 Basics in Association Rules</title>
      <p>
        Discovering association rules is a non-supervised data-mining task that aims to induce
symbolic conditional rules from a data file [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. The conditional rule has the following
syntax: “IF conditions THEN results”. The combination of conditions leads to a
descriptive rule like: “IF A AND NOT B THEN C AND D”. It can lead, also, to a
predictive rule including the time dimension like: “IF condition at time_1 THEN
result at time_2”. This looks like: “IF he_bays(TV, today) THEN he_bayes(recorder,
in_one_year)”.
      </p>
      <p>
        The association rules are used in a wide range of applications, due to their
comprehensive format. Mainly, it is used to analyze the supermarket tickets. This
helps in optimal stock management, in merchandizing and marketing [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ].
      </p>
      <p>
        During the few last years, a wide variety of algorithms have been proposed. The
most popular algorithm is certainly the iterative one: Apriori [
        <xref ref-type="bibr" rid="ref8">9</xref>
        ]. It consists in three
steps. First aboard, it considers all the attributes/properties describing the set of data
as the initial itemsets. Then, it combines the couples of item-sets in order to obtain
larger item-sets. Finally, it searchs for the rules that can be derived from each itemset.
      </p>
      <p>
        MaxEclat and MaxClique [
        <xref ref-type="bibr" rid="ref9">10</xref>
        ] use graph theory to decompose the lattice into
sublattices in order to mine maximal frequent item-sets. MaxMiner [
        <xref ref-type="bibr" rid="ref10">11</xref>
        ] is an exploring
algorithm for the maximal frequent item-sets with a look-ahead pruning strategy.
PincerSearch [
        <xref ref-type="bibr" rid="ref11">12</xref>
        ] follows both a bottom-up and a top-down search in the lattice at
the same time while looking for the maximal frequent patterns. GenMax [
        <xref ref-type="bibr" rid="ref12">13</xref>
        ] uses a
backtracking search to enumerate all maximal frequent item-sets and eliminates
progressively non maximal ones. Mafia [
        <xref ref-type="bibr" rid="ref13">14</xref>
        ] works with a depth first search and uses
pruning strategies to mine maximal frequent item-sets. DepthProject [
        <xref ref-type="bibr" rid="ref14">15</xref>
        ] follows
dynamic re-ordering to reduce the search space. Salleb &amp; al. [
        <xref ref-type="bibr" rid="ref15">16</xref>
        ] uses a boolean
algebra approach that loads the data base in main memory as a binary decision
diagram, and searchs for maximal frequent item-sets by making iteratively the product
of vectors representing the items. The reader can find further description on these
algorithms in [
        <xref ref-type="bibr" rid="ref12 ref15">13, 16</xref>
        ].
      </p>
      <p>
        Generally, the number of association rules (induced by these algorithms) grows in
an exponential manner with the number of data-rows (or the attributes). This can
reach hundreds’ of thousands using only some thousands of data rows [
        <xref ref-type="bibr" rid="ref16">17</xref>
        ]. So, it
becomes very hard to read and interpret them. The induced rules can be classified in
three categories:
- Useful rules: that a human expert can understand and use.
- Trivial rules: that represents evidence. They are valid but never used.
- Weak rules: those are not acceptable by the expert and not understandable.
      </p>
      <p>
        To remedy to this problem, many methods were proposed [
        <xref ref-type="bibr" rid="ref16">17</xref>
        ]. The basic one
consists of calculating a minimal set of rules that allows us to re-produce the others.
This can be done while calculating the rules [
        <xref ref-type="bibr" rid="ref17">18</xref>
        ], or after that using the concept lattice
[
        <xref ref-type="bibr" rid="ref18 ref19">19, 20</xref>
        ]. Another method searches only the rules where the condition or the result part
matches a particular group of terms [
        <xref ref-type="bibr" rid="ref20">21</xref>
        ]. Agrawal et al. [
        <xref ref-type="bibr" rid="ref8">9</xref>
        ] maintains a partial order
relation between the rules using two measures: the support and the confidence.
Consider the rule: X → Y , where X (conditions) and Y (results) are sets of atomic
propositions:
      </p>
      <p>- Support (X → Y): the percentage of examples satisfying both X and Y.
- Confidence (X → Y): the number of examples satisfying both X and Y,
devided by the number of examples satisfying only the condition X.</p>
      <p>The authors introduce two thresholds defined by the end-user:
- MinSup: the minimal value of the rule support that can be accepted.
- MinConf: the minimal value of the rule confidence that can be accepted.</p>
      <p>The method discards all the item-sets with a support less than MinSup, and all the
rules with a confidence less than MinConf.</p>
      <p>
        To filter more efficiently the association rules, some authors [
        <xref ref-type="bibr" rid="ref21">22</xref>
        ] introduce the
conviction measure and [
        <xref ref-type="bibr" rid="ref16">17</xref>
        ] propose five different measures like the benefit (interest)
and the satisfaction.
      </p>
      <p>
        In this paper, we introduce a new mesure: the gain. It combines the support of the
rule with the length of the itemset. This is done thanks to the theory of Formal
Concept Analysis [
        <xref ref-type="bibr" rid="ref3 ref4">3, 4</xref>
        ], since we suggest that an item-set is completely represented
by a formal concept as a couple of intent (the classic item-set) and extent (its support).
3. Basics in Formal Concept Analysis
Let O be a set of objects (or examples), P a set of properties (or attributes) and R a
binary relation defined between O and P [
        <xref ref-type="bibr" rid="ref3 ref4">3, 4</xref>
        ].
      </p>
      <p>
        Definition 1 [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]: A formal context (O, P, R) consists of two sets O and P and a
relation R between O and P. The elements of O are called the objects and the elements
of P are called the properties of the context. In order to express that an object o is in a
relation R with a property p, we write oRp or (o, p)∈R and read it as "the object o has
the property p".
      </p>
      <p>
        Definition 2 [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]: For a set A⊆O of objects and a set B⊆P of properties, we define :
The set of properties common to the objects in A :
      </p>
      <p>A4={p∈P | oRp for all o∈A} (1)
The set of objects which have all properties in B :</p>
      <p>
        B3={o∈O | oRp for all p∈B}
The couple of operators (4,3) is a Galois Connection.
(2)
Definition 3 [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]: A formal concept of the context (O, P, R) is a pair (A, B) with
      </p>
      <p>A⊆O, B⊆P, A4=B and B3=A. (3)
We call A the extent and B the intent of the concept (A, B).</p>
      <p>
        Definition 4 [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]: The set of all concepts of the context (O, P, R) is denoted by ³(O,
P, R). An ordering relation (&lt;&lt;) is easily defined on this set of concepts by :
(A1, B1) &lt;&lt; (A2, B2) ⇔ A1⊆A2 ⇔ B2⊆B1
1.a Formal context (O, P, R)
FC1
      </p>
      <p>∅
o1,o2, o3, o4, o5
FC2</p>
      <p>B
o1, o2, o3, o4
FC4</p>
      <p>A, B
o1, o2</p>
      <p>B, C
o3, o4
A, B, C, D
∅</p>
      <p>C
o3, o4, o5
B, C, D
o4</p>
      <p>FC3
C, D
o4, o5
FC7</p>
      <p>FC6
P A
1
1
0
0
0
O
o1
o2
o3
o4
o5
FC5</p>
      <p>
        FC8
1.b Concept Lattice of the context (O, P, R)
Basic Theorem for Concept Lattices [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] :
(³ (O, P, R), &lt;&lt;) is a complete lattice. It is called the concept lattice (or GALOIS
lattice) of (O, P, R), for which infimum and supremum can be described as follow:
Sup
      </p>
      <p>
        i∈I(Ai, Bi)=((∪i∈IAi)43, (∩i∈IBi))
Inf
i∈I(Ai, Bi)=( ∩i∈IAi, (∪i∈IBi)34)
(5)
(6)
Example : Figure 1 (graphic 1.a) is used to illustrate the notion of formal context (O,
P, R). This context includes five objects {o1, o2, o3, o4, o5} described by four
properties {A, B, C, D}. The concept lattice of this context is drawn in Figure 1
(graphic 1.b). It contains eight formal concepts.
Definition 5 [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]: Let (o, p) be a couple in the context (O, P, R). The pseudo-concept
PC containing the couple (o, p) is the union of all the formal concepts containing
(o,p).
      </p>
      <p>
        Definition 6 [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]: A coverage of a context (O, P, R) is defined as a set of formal
concepts CV={RE1, RE2, ..., REn} in ³ (O, P, R), such that any couple (o, p) in the
context (O, P, R) is included in at least one concept of CV.
      </p>
      <p>Example: To illustrate the notion of pseudo-concept and coverage, we consider the
same formal context (O, P, R) from figure 1. Figure 2 (graphic 2.a) represents the
pseudo-concept containing the couple (o3, B). It is the union of the concepts FC2 and
FC5 (see figure 1).</p>
      <p>
        A coverage of the context (O, P, R) can be formed by the three concepts: {FC4,
FC5 , FC6}. FC4 is the concept containing the items ({o1, o2}, {A, B}). FC5 is the
concept containing the items ({o3, o4}, {B, C}). FC6 is the concept containing the
items ({o4, o5}, {C, D}). The lattice constitutes also a concept coverage.
4 Discovery of Optimal Item-sets
The most computationally expensive step to discover association rules is the
calculation of the frequent item-sets [
        <xref ref-type="bibr" rid="ref14">15</xref>
        ]. Generally, this step consists of applying,
iteratively, some heuristics to calculate candidate item-sets. At the iteration i, we
combine the item-sets of the iteration i-1. After that, the support threshold (minsupp)
is used to discard non-frequent candidates. The item-sets of iteration i-1, are also
discarded. We keep the remaining item-sets of the latest iteration n (where n is the
number of properties in the formal context).
      </p>
      <p>In this step, we think that many approaches use, in an implicit manner, two basic
characteristics of the association rules:
- General rules: we should ignore the item-sets with high support and low
cardinality. General rules are week, since they are valid for numerous
examples (or observations).</p>
      <p>Specific rules: we should ignore the item-sets with high cardinality and low
support. Specific rules are week, since they are valid only for few examples
(or observations).</p>
      <p>The two characteristics are defined based on the cardinality of two special sets:
- Support: that is the cardinality of the set of examples verifying the rule. In</p>
      <p>Formal Concept Analysis, this represents the extent of a formal concept.
- Cardinality of the Item-set: that is the number of properties in the item-set.</p>
      <p>In Formal Concept Analysis, this represents the intent of a formal concept.</p>
      <p>Consequently, we think that an association rule should not be represented only by
the item-set (that is the intent). An association rule is completely and explicitly
represented by a formal concept (as a couple of intent and extent).</p>
      <p>Also, we think that the selection of “good” item-sets should not be based only on
the support (that is the cardinality of the extent). A powerful selection should be
based on the cardinalities of both the extent and the intent of its formal concept.</p>
      <p>To formalize the new criterions, we give the following definitions.</p>
      <p>Definition 7: Let FCi= (Ai, Bi) be a formal concept. We define:
- Length of a concept FCi: the number of properties in the intent Bi of the concept.
- Width of a concept FCi: the number of examples in the extent Ai of the concept.
- Gain of a concept: is a function of the width and the length, given by:</p>
      <p>Gain(FCi)=width(FCi)*length(FCi)–[width(FCi)+length(FCi)] (7)
In a manner that the gain function cannot be maximised when one of the two
parameters is low and the other is high. It becomes maximal only when the two
parameters are high (a lot of examples and a lot of properties). Hence, we are sure that
the rule is not a general one, neither a specific one. Note also that the Width indicates
the support of the itemset (or the corresponding association rule).</p>
      <p>
        Definition 8 [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]: A formal concept FCi= (Ai, Bi) containing a couple (o, p) is said to
be optimal if it maximizes the gain function.
      </p>
      <p>
        Definition 9 [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]: A coverage CV={ FC1, FC2, ..., FCk} of a context (O, P, R) is
optimal if it is composed by optimal concepts.
      </p>
      <p>Example: Figure 2 (graphic 2.b) represents the optimal concept FC5 containing the
couple (o3, B) and having the gain value 0. Figure 2 (graphic 2.c) represents the non
optimal concept FC2 containing the couple (o3, B) and having the gain value -1.</p>
      <p>The optimal coverage of the context (O, P, R) is formed by three optimal
concepts: {FC4., FC5 , FC6}. FC4 is the concept containing the items ({o1, o2}, {A,
B}). FC5 is the concept containing the items ({o3, o4}, {B, C}). FC6 is the concept
containing the items ({o4, o5}, {C, D}).
4.1 Heuristic Searching for Optimal Concept
The pseudo-concept containing the couple (o, p), introduced in definition 5, is the
union of all the concepts containing (o, p). The importance of the notion of
pseudoconcept, indicated by PCF, is that it can be calculated in a simple manner as the
restriction of the relation R by the set of objects/examples described by p, so {p}4,
and the set of properties describing the object o, so {o}3. Where (4,3) is the Galois
connection of the context (O, P, R).</p>
      <p>Once we obtain the pseudo-concept PCF, two cases are considered:
- Case 1: PCF forms a formal concept. This means that no zero is found in the
relation/matrix representing PCF. Then, PCF is the optimal concept that we
are looking for. The algorithm stops.
- Case 2: PCF is not a formal concept. This means, simply, that we can find
some zero entries in the relation/matrix representing PCF. In this case, we
will look for more restraint pseudo-concepts within the pseudo-concept PCF.
So, we consider the pseudo-concepts containing the couples like (X, p) or (o,
Y). These concepts contain, certainly, the couple (o, p). The heuristic that we
use here is that: the optimal concept is included in the optimal
pseudoconcept. So, we should calculate the Gain value of the different
pseudoconcepts containing the couples like (X, p) or (o, Y). Then we choose the
pseudo-concept having the greatest value of the Gain function to be the new
PCF.</p>
      <p>This heuristic procedure (of case 2) is repeated until we meat the case 1: until PCF
becomes a formal concept. To calculate the gain of a pseudo-concept, we introduce
the general form of the previous function:
Definition 10: Let PCFi= (Ai, Bi, Ri) be a pseudo-concept, where Ri is the restriction
of the binary relation R, to the subsets Ai and Bi. We define the:
- Length of a pseudo-concept PCFi: the number of properties in Bi.
- Width of a pseudo-concept PCFi: the number of examples in Ai.
- Size of a pseudo-concept PCFi: the number of couples (of values equal to 1) in
the pseudo-concept. When PCFi is a formal concept, we have:</p>
      <p>Size(PCF)=Length(PCF)×Width(PCF) (8)</p>
      <p>Gain of a pseudo-concept: is a function of the width, the length and the size given
by (9) :</p>
      <p>Size(PCF) ⎤×[Size(PCF)−(Length(PCF)+Width(PCF))]
Gain(PCF )=⎡⎢⎣ Length(PCF)×Width(PCF) ⎦⎥</p>
      <sec id="sec-2-1">
        <title>4.2 Algorithm for Optimal Coverage</title>
        <p>The problem of covering a binary relation by a set of optimal concepts is known as
the problem of covering a binary matrix by a number of its complete sub-matrix. A
complete sub matrix is a matrix where all its entries are equal to '1'. This was found as
an NP-hard problem. Here, we use a polynomial heuristic solution.</p>
        <p>Let R be the binary relation to cover. The heuristic solution consists of dividing R
into p packages (subsets): P1, ..., Pp. Each package represents one or more couples.
The idea is to construct step by step the optimal coverage of R. In the first step, we
cover the relation R1=P1 by CV1. In the kth step, let Rk-1=P1 ∪ ... ∪ Pk-1 and let CVk-1
be its optimal coverage. The task is to construct the optimal coverage CVk of Rk=Rk-1
∪ Pk using only the initial coverage CVk-1 and the package Pk. In the pth step, we
obtain a set of concepts covering the relation R.</p>
        <p>Algorithm for building the Coverage
Begin
Let R be partitioned to p packages P1, ..., Pp.
Let CV0:=∅.</p>
        <p>FOR k=1 to p DO</p>
        <p>Sort the couples of Pk by the pertinence of
their pseudo-concepts
While (Pk≠∅) Do
-Select a couple (a, b) in Pk by the sorted
order of the gain function (9)
-Search PC : the pseudo-concept containing
(a, b) within Rk=CVk-1∪Pk
-Search FC: the optimal concept
containing (a, b) within PC
CVk:=(CV k-1-{r∈CV k-1/ r⊆FC }) ∪{FC}: Delete
all the redundant concepts from CVk</p>
        <p>Pk:=Pk-{(X,Y)∈Pk/(X,Y)∈FC}</p>
        <p>End While
End FOR
End.</p>
        <p>Example: Let R be the relation of figure 1. R is partitioned into five packages:
P1={o1}x{A, B}, P2={o2}x{A, B}, P3={o3}x{B, C}, P4={o4}x{B, C, D} and
P5={o5}x{C, D}. Initially, R is an empty relation and in each step we add a package.
Figure 3 presents the incrementation step when adding P5.</p>
        <p>In this step R4 encloses the four rows P1, ..., P4. The initial optimal coverage CV4
encloses the formal concepts FC4=({o1, o2}, {A, B}) and FC5=({o3, o4}, {B, C})
and FC7=({o4}, {B, C, D}) (graphic 3.a). The package P5 encloses only two couples.
The pseudo concept containing the couple (o5, C) represents the union of two formal
concepts FC3 and FC6. So, the optimal concept containing (o5, C) and (o5, D) is the
concept FC6. The concept FC7 is redundant. Thus, the final coverage of R contains
the concepts FC4, FC5 and FC6.</p>
        <p>In conclusion, from the data set of figure 1, we discover three Item-sets : {A, B},
{B, C} and {C, D}. These Item-sets will be processed later by the Apriori algorithm,
to calculate the corresponding association rules.</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>5 Experiments</title>
      <p>
        The experiments are done with a PC PENTIUM II, 233 MHZ, 144 Mo of RAM and
running under MS-Windows 98. The analyzed data sets are organized in two groups.
The first group is made by five data sets representing biological sequences coding
macromolecules (table 1). The first two data sets are proprietary sets for the TLR
macromolecules. The three others are taken from the machine-learning repository at
UCI University. Classic tables of data make the second group of data sets (table 2).
The first two data sets are also taken from the machine-learning repository at UCI
University. The other sets are taken from a proprietary database.
To evaluate the performance of the proposed method (IAR: Incremental induction of
Association Rules), we compare it to the existing approach Aprior [
        <xref ref-type="bibr" rid="ref22 ref8">9, 23</xref>
        ]. The
Apriori method was programmed in C++ and integrated in our software package. We
      </p>
      <sec id="sec-3-1">
        <title>Identifier</title>
        <p>DB1
DB2
DB3
DB4
DB5</p>
      </sec>
      <sec id="sec-3-2">
        <title>Identifier</title>
        <p>DB6
DB7
DB8
DB9
DB10</p>
      </sec>
      <sec id="sec-3-3">
        <title>Rows</title>
        <p>763
619
106
1150
500
Rows
958
1191
1000
5000
1000
use the two methods to analyze the ten data sets. We choose two sets of parameters.
The first one with MinSup=0.6 and MinConf=0.5 is used for the biological data sets
(table 1). The second one with MinSup=0.35 and MinConf=0.75 is used for the
standard data sets (table 2).
)
c
e
s
(
e
m
it
U
P</p>
        <p>C
IAR
Apriori
)
%
(
s
e
s
a
b
e
g
d
e
l
w
o
n
K
f
o
y
iilitr
a
m</p>
        <p>S
IAR/Apriori</p>
        <p>DB1 DB2 DB3 DB4
0,21 0,13 0,49 0,03
MinSup=0.6 and MinConf=0.5</p>
        <p>DB5
0,16</p>
        <p>IAR/Apriori</p>
        <p>DB6
0,42</p>
        <p>DB7
0,12</p>
        <p>DB8
0,3</p>
        <p>DB9
0,03</p>
        <p>DB10
0,24
MinSup=0.35 and MinConf=0.75
Fig. 5: Similarity between the knowledge bases of the two methods</p>
        <p>DB1 DB2 DB3
3930 304 9
12451 916 12</p>
        <p>MinSup=0.6 and MinConf=0.5</p>
        <sec id="sec-3-3-1">
          <title>5.1. Comparing the CPU time</title>
          <p>Figure 4 presents the CPU time measured in seconds for the three methods. We
remark that the three methods keep the same behavior overall the data sets. The
Apriori method takes the greater time, since it is based on an exhaustive approach to
test all the possible combinations. The proposed method IAR takes always a CPU
time better than Apriori.
)
c
e
s
(
e
m
it
U
P</p>
          <p>C
IAR
Apriori
)
%
(
s
e
s
a
b
e
g
d
e
l
w
o
n
K
f
o
y
iliitr
a
m
S</p>
          <p>A FMoromnadlhCeronMceapdtdAounrailysis Approach to Discover Association Rules from Data</p>
        </sec>
        <sec id="sec-3-3-2">
          <title>5.2. Measuring the similarity of rules</title>
        </sec>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>6 Conclusion</title>
      <p>
        In this paper, we presented a new Formal Concept Analysis approach to search for
optimal item-sets. We gave two heuristic algorithms to build the coverage of formal
concepts and to create optimal item-sets from data records. The association rules can
be induced from the Item-sets in the same manner as Apriori [
        <xref ref-type="bibr" rid="ref8">9</xref>
        ].
      </p>
      <p>We assume that a Formal Concept generalizes the notion of Item-set, since it is
more powerful and more explicit. In fact, we note that an item-set represents only the
intent of a Formal Concept. Also, the support of an item-set is the cardinality of the
concept extent. Hence, the Formal Concept encloses more information than the
itemset itself.</p>
      <p>
        The first experimentation of the proposed method is done on a variety of data sets
having different data formats (structured and unstructured). We compared the
proposed method with a known one: Apriori [
        <xref ref-type="bibr" rid="ref8">9</xref>
        ]. We measured their CPU time, the
number of association rules, and the percentage of similar rules. We remarked that
Apriori gives a great number of rules and takes a great time. The proposed method
IAR is faster than Apriori. It gives a low number of rules that have the advantage of
being the common rules.
      </p>
      <p>
        Future work will focus on the quality of the created association rules. In fact, we
plan to study the significance of the discovered rules for human experts. For this
reason we propose to present the rules concerning the TLR macromolecules to our
biologists in order to get their points of view. Accordingly, we plan to study quality
measures other than the support and the confidence [
        <xref ref-type="bibr" rid="ref16 ref21">17, 22</xref>
        ], in order to identify the
measure that gives expert-acceptable rules.
      </p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <given-names>P.</given-names>
            <surname>Adriaans</surname>
          </string-name>
          et
          <string-name>
            <given-names>D.</given-names>
            <surname>Zantinge</surname>
          </string-name>
          .
          <article-title>Data mining</article-title>
          .
          <source>Addison Wesley Longman, England</source>
          ,
          <year>1996</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <given-names>R.</given-names>
            <surname>Wille</surname>
          </string-name>
          .
          <source>Restructuring Lattice Theory : an Approach Based on Hierarchies of Concepts</source>
          .
          <source>In Proc. Symposium on Ordered Sets</source>
          , pp.
          <fpage>445</fpage>
          -
          <lpage>470</lpage>
          , Dordrecht-Boston: reidel,
          <year>1982</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Ganter</surname>
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wille</surname>
            <given-names>R.</given-names>
          </string-name>
          ,
          <source>Formal Concept Analysis: Mathematical Foundations</source>
          , Edition: Springer, Heidelberg-Berlin-New York,
          <year>1999</year>
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Jaoua</surname>
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Beaulieu</surname>
            <given-names>J. M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Belkhiter</surname>
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Deshernais</surname>
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Reguig</surname>
            <given-names>M.</given-names>
          </string-name>
          ,
          <article-title>Optimal rectangular decomposition of a finite binary relation</article-title>
          ,
          <source>International Conference on Discrete Mathematics (sixth conference</source>
          ,
          <year>June 1992</year>
          , Vancouver-Canada).
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Wille</surname>
            <given-names>R.</given-names>
          </string-name>
          ,
          <source>Why can Concept Lattice Support Knowledge Discovery in DataBases, First Workshop on Formal Concept Analysis Advances for KDD</source>
          ,
          <year>2001</year>
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          7.
          <string-name>
            <given-names>M.</given-names>
            <surname>Maddouri</surname>
          </string-name>
          .
          <article-title>Contribution to Concept Learning : an Incremental approach for Inducing Production Rules from examples</article-title>
          .
          <source>PhD dissertation</source>
          from the university of TUNIS II, May
          <year>2000</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          8.
          <string-name>
            <given-names>M.</given-names>
            <surname>Maddouri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Elloumi</surname>
          </string-name>
          ,
          <article-title>A Data Mining Approach based on Machine Learning Techniques to Classify Biological Sequences, Knowledge Based Systems Journal</article-title>
          ,
          <year>March 2002</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          9.
          <string-name>
            <surname>Agrawal</surname>
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Imielinski</surname>
            <given-names>T</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Swami</surname>
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Mining association rules between sets of items in large databases</article-title>
          .
          <source>ACM SIGMOD Int. Conf. on Management of Data</source>
          (
          <year>1993</year>
          )
          <fpage>207</fpage>
          -
          <lpage>213</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          10.
          <string-name>
            <surname>Zaki M. J.:</surname>
          </string-name>
          <article-title>Scalable algorithms for association mining</article-title>
          .
          <source>Int. J. IEEE Transactions on Knowledge and Data Engineering</source>
          , Vol
          <volume>12</volume>
          (
          <issue>3</issue>
          ). (May/June 2000)
          <fpage>372</fpage>
          -
          <lpage>390</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          11.
          <string-name>
            <surname>Bayardo</surname>
            <given-names>R.</given-names>
          </string-name>
          :
          <article-title>Efficiently mining long patterns from databases</article-title>
          .
          <source>ACM SIGMOD Int. Conf. on Management of Data</source>
          (
          <year>1998</year>
          )
          <fpage>85</fpage>
          -
          <lpage>93</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          12.
          <string-name>
            <surname>Lin</surname>
            <given-names>D.</given-names>
          </string-name>
          , Kedem Z.:
          <article-title>Pincer search: a new algorithm for discovering the maximum frequent set</article-title>
          .
          <source>Int. Conf. on Extending Database Technology, (March</source>
          <year>1998</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          13.
          <string-name>
            <surname>Gouda</surname>
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zaki</surname>
            <given-names>M. J.:</given-names>
          </string-name>
          <article-title>Efficiently mining maximal frequent itemsets</article-title>
          .
          <source>IEEE Int. Conf. on Data Mining, (November 01).</source>
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          14.
          <string-name>
            <surname>Burdick</surname>
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Calimlim</surname>
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gehrke</surname>
            <given-names>J. E.</given-names>
          </string-name>
          :
          <article-title>Mafia: a maximal frequent itemset algorithm for transactional databases</article-title>
          .
          <source>Int. Conf. on Data Engineering</source>
          , (
          <year>April 2001</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          15.
          <string-name>
            <surname>Agarwal</surname>
            <given-names>R. C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Aggarwal</surname>
            <given-names>C. C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Prasad</surname>
            <given-names>V. V. V.</given-names>
          </string-name>
          :
          <article-title>Depth first generation of long patterns</article-title>
          .
          <source>Int. Conf. on Knowledge Discovery and Data Mining</source>
          , (
          <year>August 2000</year>
          )
          <fpage>108</fpage>
          -
          <lpage>118</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          16.
          <string-name>
            <surname>Salleb</surname>
            <given-names>A.</given-names>
          </string-name>
          , Maazouzi
          <string-name>
            <given-names>Z.</given-names>
            ,
            <surname>Vrain</surname>
          </string-name>
          <string-name>
            <surname>C.</surname>
          </string-name>
          :
          <article-title>Mining maximal frequent item-sets by a Boolean based approach</article-title>
          .
          <source>European Conf. on Artificial Intelligence</source>
          , Lyon France (
          <year>July 2002</year>
          )
          <fpage>285</fpage>
          -
          <lpage>289</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          17.
          <string-name>
            <surname>Cherfi</surname>
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Toussaint</surname>
            <given-names>Y.</given-names>
          </string-name>
          :
          <article-title>Text mining by combining association rules and statistics</article-title>
          .
          <source>Int. Workshop on Text Mining CIFT-02</source>
          , Hammamet - Tunisia
          <source>(October 21-23</source>
          ,
          <year>2002</year>
          )
          <fpage>67</fpage>
          -
          <lpage>80</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          18.
          <string-name>
            <surname>Guigues</surname>
            <given-names>J. L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Duquenne</surname>
            <given-names>V.</given-names>
          </string-name>
          :
          <article-title>Familles minimales d'implication informatives resultant d'un tableau de données binaries</article-title>
          .
          <source>J. Mathematiques, informatique et Sciences Humaines</source>
          , Vol.
          <volume>95</volume>
          (
          <year>1986</year>
          )
          <fpage>5</fpage>
          -
          <lpage>18</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          19.
          <string-name>
            <surname>Stumme</surname>
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Taouil</surname>
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bastide</surname>
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pasquier</surname>
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lakhal</surname>
            <given-names>L.</given-names>
          </string-name>
          :
          <article-title>Intelligent structuring and reducing of association rules with formal concept analysis</article-title>
          .
          <source>Lecture Notes in Advances in Artificial Intelligence</source>
          , KI-2000
          <source>Proceedings</source>
          , Vol.
          <volume>2174</volume>
          . Springer-Verlag (
          <year>2001</year>
          )
          <fpage>335</fpage>
          -
          <lpage>350</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          20.
          <string-name>
            <surname>Toussaint</surname>
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Simon</surname>
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Building and interpreting term dependencies using association rules extracted from Galois lattices</article-title>
          .
          <source>Int. Conf. on Content-Based Multimedia Information Access</source>
          , Paris France (
          <year>2000</year>
          )
          <fpage>1686</fpage>
          -
          <lpage>1693</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          21.
          <string-name>
            <surname>Klemettinen</surname>
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Manilla</surname>
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ronkainen</surname>
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Toivonen</surname>
            <given-names>H</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Verkamo</surname>
            <given-names>A. I.</given-names>
          </string-name>
          :
          <article-title>Finding interesting rules from large sets of discovered association rules</article-title>
          .
          <source>Int. Conf. on Knowledge Management</source>
          ,
          <string-name>
            <surname>Gaithersburg</surname>
            <given-names>USA</given-names>
          </string-name>
          (
          <year>1994</year>
          )
          <fpage>401</fpage>
          -
          <lpage>407</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          22.
          <string-name>
            <surname>Bayardo</surname>
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Agrawal</surname>
            <given-names>R.</given-names>
          </string-name>
          :
          <article-title>Mining the most interesting rules</article-title>
          .
          <source>ACM SIGKDD Int. Conf. on Knowledge Discovery and Data Mining</source>
          (
          <year>1999</year>
          )
          <fpage>145</fpage>
          -
          <lpage>154</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          23.
          <string-name>
            <given-names>D. A.</given-names>
            <surname>Zighed</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. P.</given-names>
            <surname>Auray</surname>
          </string-name>
          ,
          <string-name>
            <surname>G. Duru.</surname>
          </string-name>
          <article-title>SIPINA: méthode et logiciel</article-title>
          .
          <source>Alexandre Lacassagne</source>
          , Lyon-France,
          <year>1992</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          24.
          <string-name>
            <surname>Dougherty</surname>
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kohavi</surname>
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sahami</surname>
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Supervised and Unsupervised Discretization of Continuous Features</article-title>
          .
          <source>Int. Conf. on Machine Learning</source>
          (
          <year>1995</year>
          ).
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>