<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>A local discretization of continuous data for lattices: Technical aspects</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Nathalie Girard</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Karell Bertet</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Muriel Visani</string-name>
          <email>mvisani@univ-lr.fr</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Laboratory L3i - University of La Rochelle - FRANCE ngirar02</institution>
          ,
          <addr-line>kbertet</addr-line>
        </aff>
      </contrib-group>
      <abstract>
        <p>Since few years, Galois lattices (GLs) are used in data mining and de ning a GL from complex data (i:e: non binary) is a recent challenge [1,2]. Indeed GL is classically de ned from a binary table (called context), and therefore in the presence of continuous data a discretization step is generally needed to convert continuous data into discrete data. Discretization is classically performed before the GL construction in a global way. However, local discretization is reported to give better classi cation rates than global discretization when used jointly with other symbolic classi cation methods such as decision trees (DTs). Using a result of lattice theory bringing together set of objects and speci c nodes of the lattice, we identify subsets of data to perform a local discretization for GLs. Experiments are performed to assess the e ciency and the effectiveness of the proposed algorithm compared to global discretization.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>(1)
Finally for one discretization step, the attribute V is divided into two intervals:
[v1; cV ] and ]cV ; vn] and the process is repeated.</p>
      <p>
        This process can be run using, at each step, all the objects in the training set.
This is global discretization. It can also be run during model construction
considering, at each step, only a part of the training set. This is local
discretization. In [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ], Quinlan shows that local discretization improves supervised
classi cation with decision trees (DTs) as compared with global
discretization. In DT construction, the growing process is iterated until S is satis ed.
Local discretization is performed on the subset of objects in the current node to
select its best attribute (V (node)), according to the splitting criterion. Given
the structural links between DTs and Galois lattices (GLs) [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ], we propose a
local discretization algorithm for GL and compare its performances with a global
discretization.
2
      </p>
    </sec>
    <sec id="sec-2">
      <title>Local discretization for Galois lattices</title>
      <p>
        A GL is generally de ned from a binary relation R between objects O and
binary attributes I - i.e. a binary data set also called a formal context - denoted
as a triplet T = (O; I; R). A GL is composed of a set of concepts - a concept
(A; B) is a maximal objects-attributes subset in relation - ordered by a
generalization/specialization relation. For more details on GL theory, notation and their
use in classi cation tasks, please refer to [
        <xref ref-type="bibr" rid="ref10 ref9">9,10</xref>
        ]. To de ne a local discretization
for GL, we have to identify at each discretization step the subset of concepts
to be processed. Given a subset of objects A 2 P (O), there always exists a
smallest concept M containing this subset and identi ed in lattice theory as a
meet-irreducible concept of the GL [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ]. Moreover, it is possible to compute
the set of meet-irreducibles directly from the context, thus the generation of the
lattice is useless [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ]. Consequently, local discretization is performed on the set
of meet-irreducible concepts M I which does not satisfy S. Attributes in M I are
locally discretized: the best attribute V (M ) for each M 2 M I is computed
according to eq. (3); then the best one V (M I) (eq. (4),(5)) for the whole set
M I is split into two intervals as explain before. The context T is then updated
with these new intervals; and its M I are computed. The process is iterated until
all M 2 M I verify the stopping criterion S. The context T is initialized with,
for each continuous attribute, an interval -i.e. a binary attribute- containing all
continuous values observed in D; thus each object is in relation with every
binary attributes of T . The GL of the inital context T contains only one concept
(O; I) being a meet-irreducible concept, which is used to initialize M I. See [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ]
for more details on the algorithm.
      </p>
      <p>The main di erence with DT is that splitting an attribute in a GL impacts
all the other concepts of the GL that contain this attribute, and due to the
order relation between concepts , the structure of the GL is also modi ed.
Whereas, when an attribute is split in a DT node, predecessors and others
branches are not impacted. In order to select the best V (M I) over all the
concepts sharing this attribute, we introduce di erent computing of V (M I).
Let M I = fDq = (Aq; Bq); q Qgg be the set of meet-irreducible concepts not
satisfying S. The best attribute V (Dq) associated to its best cut-point is rst
computed for each concept Dq 2 M I:</p>
      <p>V (Dq) = argmaxV 2Bq (gain(V; cV ; Dq))
where cV is de ned by (1) for Dq instead of D.</p>
      <p>Let us de ne IMI = fV (D1); : : : ; V (DQ)g the set of best attributes associated
to each concept in M I. The best attribute V (M I) among IMI can be de ned
in two di erent ways:
By local discretization: Local discretization selects the best attribute V 2
IMI as the one having the best gain for M I:</p>
      <p>V (M I) = argmaxV (Dq)2IMI (gain(V (Dq); cV (Dq); Dq))
By linear local discretization: Linear local discretization takes into account
that the split of one attribute V 2 IMI in a concept Dq can impact the other
concepts. So we compute a linear combination of the criterion as the sum of
the gain for each concept Dq0 2 M I containing this attribute V . The selected
attribute is the one that gives the best linear combination:
(3)
(4)
V (M I) = argmaxV 2IMI (</p>
      <p>X
Dq0 2MIjV 2Bq0</p>
      <p>jAq0 j
PDq2MI jAqj
gain(V; cV ; Dq0 )) (5)
3</p>
    </sec>
    <sec id="sec-3">
      <title>Experimental comparison</title>
      <p>
        The study is performed on three supervised databases of the UCI Machine
Learning Repository1: the Image Segmentation database (Image1), the Glass Identi
cation Database (GLASS) and the Breast Cancer Database (BREAST Cancer).
We also use one supervised data set stemming from GREC 2003 database2
described by the statistical Radon signature (GREC Radon). Table 1 presents the
complexity of each lattice structure associated to each discretization
algorithm and the classi cation performance using each GL by navigation [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ]
and using CHAID as DT classi er [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. Discretization is performed in each case
with 2 as a splitting and stopping supervised criterion.
4
      </p>
    </sec>
    <sec id="sec-4">
      <title>Conclusion</title>
      <p>
        The study [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] shows that for DTs, local discretization induces more complex
structures compared to global discretization; Table 1 shows that for GL, on
the contrary, local discretization allows to reduce the structures'
complexity. In [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ], Quinlan proves that local discretization improves classi cation
performance of DTs compared to global discretization; as in DTs, Table 1 shows
that local discretization improves GLs classi cation performances.
1 http://archive:ics:uci:edu/ml 2 www.cvc.uab.es/grec2003/symreccontest/index.htm
      </p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Ganter</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kuznetsov</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          :
          <article-title>Pattern structures and their projections</article-title>
          . In Delugach, H.,
          <string-name>
            <surname>Stumme</surname>
          </string-name>
          , G., eds.: Conceptual Structures:
          <article-title>Broadening the Base</article-title>
          . Volume
          <volume>2120</volume>
          of LNCS. (
          <year>2001</year>
          )
          <volume>129</volume>
          {
          <fpage>142</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Kaytoue</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kuznetsov</surname>
            ,
            <given-names>S.O.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Napoli</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Duplessis</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          :
          <article-title>Mining gene expression data with pattern structures in formal concept analysis</article-title>
          .
          <source>Inf. Sci</source>
          .
          <volume>181</volume>
          (
          <year>2011</year>
          )
          <year>1989</year>
          {
          <fpage>2001</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Dougherty</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kohavi</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sahami</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Supervised and unsupervised discretization of continuous features</article-title>
          .
          <source>In: Machine Learning: Proc. of the Twelfth International Conference</source>
          , Morgan Kaufmann (
          <year>1995</year>
          )
          <volume>194</volume>
          {
          <fpage>202</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Muhlenbach</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rakotomalala</surname>
          </string-name>
          , R.:
          <article-title>Discretization of continuous attributes</article-title>
          . In Reference, I.G., ed.:
          <article-title>Encyclopedia of Data Warehousing and Mining</article-title>
          . J.
          <string-name>
            <surname>Wang</surname>
          </string-name>
          (
          <year>2005</year>
          )
          <volume>397</volume>
          {
          <fpage>402</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Breiman</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Friedman</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Olshen</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Stone</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          :
          <article-title>Classi cation and regression trees</article-title>
          .
          <source>Wadsworth Inc</source>
          ., 358 pp (
          <year>1984</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Kass</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          :
          <article-title>An exploratory technique for investigating large quantities of categorical data</article-title>
          .
          <source>Applied Statistics</source>
          <volume>29</volume>
          (
          <issue>2</issue>
          ) (
          <year>1980</year>
          )
          <volume>119</volume>
          {
          <fpage>127</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Quinlan</surname>
          </string-name>
          , J.:
          <article-title>Improved use of continuous attributes in C4.5</article-title>
          .
          <source>Journal of Arti cial Intelligence Research</source>
          <volume>4</volume>
          (
          <year>1996</year>
          )
          <volume>77</volume>
          {
          <fpage>90</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>Guillas</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bertet</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Visani</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ogier</surname>
            ,
            <given-names>J.M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Girard</surname>
            ,
            <given-names>N.:</given-names>
          </string-name>
          <article-title>Some links between decision tree and dichotomic lattice</article-title>
          .
          <source>In: Proc. of the Sixth International Conference on Concept Lattices and Their Applications</source>
          ,
          <string-name>
            <surname>CLA</surname>
          </string-name>
          <year>2008</year>
          (
          <year>2008</year>
          )
          <volume>193</volume>
          {
          <fpage>205</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Ganter</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wille</surname>
          </string-name>
          , R.:
          <article-title>Formal concept analysis</article-title>
          ,
          <source>Mathematical foundations</source>
          . Springer Verlag, Berlin, 284 pp (
          <year>1999</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Fu</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Fu</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Njiwoua</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Nguifo</surname>
            ,
            <given-names>E.M.:</given-names>
          </string-name>
          <article-title>A comparative study of fca-based supervised classi cation algorithms</article-title>
          .
          <source>In: Concept Lattices</source>
          . Volume LNCS
          <volume>2961</volume>
          .
          <article-title>(</article-title>
          <year>2004</year>
          )
          <volume>219</volume>
          {
          <fpage>220</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>Birkho</surname>
          </string-name>
          , G.:
          <article-title>Lattice theory</article-title>
          .
          <source>Third edn. Volume 25. American Mathematical Society</source>
          , 418 pp (
          <year>1967</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <string-name>
            <surname>Wille</surname>
          </string-name>
          , R.:
          <article-title>Restructuring lattice theory : an approach based on hierarchies of concepts</article-title>
          .
          <source>Ordered sets</source>
          (
          <year>1982</year>
          )
          <volume>445</volume>
          {470 I. Rival (ed.), Dordrecht-Boston, Reidel.
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13.
          <string-name>
            <surname>Girard</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bertet</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Visani</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Local discretization of numerical data for galois lattices</article-title>
          .
          <source>In: Proceedings of the 23rd IEEE International Conference on Tools with Arti cial Intelligence</source>
          ,
          <string-name>
            <surname>ICTAI</surname>
          </string-name>
          <year>2011</year>
          (
          <year>2011</year>
          ) to appear.
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          14.
          <string-name>
            <surname>Visani</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bertet</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ogier</surname>
            ,
            <given-names>J.M.:</given-names>
          </string-name>
          <article-title>Navigala: an original symbol classi er based on navigation through a galois lattice</article-title>
          .
          <source>International Journal of Pattern Recognition and Arti cial Intelligence, IJPRAI</source>
          <volume>25</volume>
          (
          <year>2011</year>
          )
          <volume>449</volume>
          {
          <fpage>473</fpage>
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>