<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Applied multi-layer clustering to the diagnosis of complex agro-systems</article-title>
      </title-group>
      <contrib-group>
        <aff id="aff0">
          <label>0</label>
          <institution>CNRS, LAAS</institution>
          ,
          <addr-line>7 avenue du colonel Roche, F-31400 Toulouse</addr-line>
          ,
          <country country="FR">France</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Univ de Toulouse</institution>
          ,
          <addr-line>INSA, LAAS, F-31400 Toulouse</addr-line>
          ,
          <country country="FR">France</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>involving the companies VIVADOUR</institution>
          ,
          <addr-line>CACG, GEOSYS, ME- TEO FRANCE, PIONEER and laboratories CESBIO, LAAS- CNRS</addr-line>
        </aff>
      </contrib-group>
      <fpage>19</fpage>
      <lpage>26</lpage>
      <abstract>
        <p>In many fields, such as medical, environmental, a lot of data are produced every day. In many cases, the task of machine learning is to analyze these data composed of very heterogeneous types of features. We developed in previous work a classification method based on fuzzy logic, capable of processing three types of features (data): qualitative, quantitative, and more recently intervals. We propose to add a new one: the object type which is a meaningful combination of other features yielding the possibility of developing hierarchical classifications. This is illustrated by a real-life case study taken from the agriculture area1.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>
        Nowadays, large scale datasets are produced in various
different fields such as social networks, medical, process
operation, agricultural/environmental,... Many studies
relate to data mining with the intention of analyzing and if
possible extracting knowledge from these data. The data
classification has to provide a relevant and well-fitted
representation of reality. In this context, the issue of
representing of data is crucial since the formalisms must be
generic yet well suited to every new problem. For machine
learning, the concern is to be able to detect adequate
patterns from heterogeneous, large, and sometimes uncertain
datasets. In diagnosis, the necessity to quickly recognize a
problem to provide a sure solution to solve it appears to be
essential. One of the main challenges is the necessity to
process heterogeneous data (qualitative, quantitative...) and
sometimes to merge data obtained in different contexts.
We developed a classification method based on fuzzy logic
[
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] capable of processing heterogeneous data types and
noisy data. The LAMDA (Learning Algorithm for
Multivariate Data Analysis) method is a classification method,
capable to process three types of data: qualitative,
quantitative, and intervals [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. We addressed one of the main
difficulties encountered in data analysis tasks: the diversity of
information types. Such information types are given by
qualitative valued data, which can be nominal or ordinal,
mixed with quantitative and interval data. Many situations
leading to well-conditioned algorithms for quantitative
valued information become very complex whenever there
are several data given in qualitative form. In a
nonexhaustive list, we can mention, rule based deduction,
classification, clustering, dimensionality reduction…
During the last decades, few research works have been
directed to defy the issue of representing multiplicity for data
analysis purposes [
        <xref ref-type="bibr" rid="ref11 ref3">3, 11</xref>
        ]. However, no standard principle
has been proposed in the literature to handle in a unified
way heterogeneous data. Indeed, a lot of proposed
techniques process separately quantitative and qualitative data.
In data reduction tasks for example, they are either based
on distance measures for the former type [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ] and on
information or consistency measures for the later one.
Whereas in classification and clustering tasks, eventually
only a Hamming distance is used to handle qualitative data
[
        <xref ref-type="bibr" rid="ref11 ref14 ref4">4,11,14</xref>
        ]. Other approaches are originally designed to
process only quantitative data and therefore arbitrary
transformations of qualitative data into a quantitative space are
proposed without taking into account their nature in the
original space [
        <xref ref-type="bibr" rid="ref12 ref15 ref16">12,15,16</xref>
        ]. For example, the variable shape
can take values in a discrete unordered set {round, square,
triangle}. These values are transformed respectively to
quantitative values 1, 2, and 3. However, we can also
choose to transform them to 3, 2 and 1. Another inverse
practice is to enhance the qualitative aspect and discretize
the quantitative value domain into several intervals, then
objects in the same interval are labeled by the same
qualitative value [
        <xref ref-type="bibr" rid="ref17 ref18">17,18</xref>
        ]. Obviously, both approaches introduce
distortion and end up with information loss with respect to
the original data. Moreover, none of the previously
proposed approaches combines in a fully adequate way, the
processing of symbolic intervals simultaneously with
quantitative and qualitative data. Although extensive
studies were performed to process this type of data in the
Symbolic Data Analysis framework [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ], they were focused
generally on the clustering tasks [
        <xref ref-type="bibr" rid="ref10 ref8">8, 10</xref>
        ] and no unified
principle was given to handle simultaneously the three
types of data for different analysis purposes. In [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ], a new
general principle, was introduced as “Simultaneous
Mapping for Single Processing (SMSP)”, enabling the
reasoning in a unified way about heterogeneous data for several
data analysis purposes. The fact that SMSP together with
LAMDA can process simultaneously these three types of
data without pre-processing is one of its principal
advantages compared to other classical machine learning
methods such as SVM (Support Vector Machine [
        <xref ref-type="bibr" rid="ref20">20</xref>
        ]),
KNN [
        <xref ref-type="bibr" rid="ref21">21</xref>
        ]. Decision trees are very powerful tools for
classification and diagnosis [
        <xref ref-type="bibr" rid="ref22">22</xref>
        ] but their sequential approach is
still not advisable to process multidimensional data since,
by their very nature, they cannot be processed as
efficiently as totally independent information [
        <xref ref-type="bibr" rid="ref23">23</xref>
        ]. A complete
description of the LAMDA method and comparison with
other classification techniques on various well known data
sets can be found in [
        <xref ref-type="bibr" rid="ref24 ref25 ref26">24, 25, 26</xref>
        ]. Its other main
characteristic is the fuzzy formalism which enables an element to
belong to several classes simultaneously. It is also possible
to perform clustering (i.e. with no a priori knowledge of
the number and the class prototypes).
      </p>
      <p>Besides the three existing types, we propose to add
another type: the class type which can be processed
simultaneously with the three former ones: quantitative,
qualitative, intervals thanks to the “SMSP”. In this configuration
the class feature represents a meaningful aggregation of
other features. This aggregation can be defined by a class
determined by a previous classification, or the result of an
abstraction. This new type gives the possibility to develop
hierarchical classifications or to fuse different
classifications. It allows an easier representation of many various
and complex types of data, like multi-dimensional data,
while being realistic and conserving their constraints. In a
first part, the LAMDA method is briefly explained. The
second part is devoted to the new type of data introduced:
the object type. Finally, this new method is exemplified
through an agronomical project.
2</p>
    </sec>
    <sec id="sec-2">
      <title>The LAMDA method</title>
      <p>
        The LAMDA method is an example of fuzzy logic based
classification methods [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]. The classification method takes
as input a sample x made up of N features. The first step is
to compute for each feature of x, an adequacy degree to
each class Ck , k = 1..K where K is the total number of
classes. This is obtained by the use of a fuzzy adequacy
function providing K vectors of Marginal Adequacy
Degree vectors (MAD). This degree estimates the closeness
of every single sample feature to the prototype
corresponding to its class. At this point, all the features are in a
common space. Then the second step is to aggregate all these
marginal adequacy degrees into one global adequacy
degree (GAD) by means of a fuzzy aggregation function.
Thus the K MAD vectors become K GADs. Fuzzy logic[
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]
is here used to express MADs and GADs, since the
membership degree of a sample to a given class is not binary
but takes a value in [
        <xref ref-type="bibr" rid="ref1">0,1</xref>
        ]. Classes can be known a priori,
commonly determined by an expert and the learning
process is therefore supervised, or classes can created during
the learning itself (unsupervised mode or clustering).
Three types of features can be processed by the LAMDA
method: quantitative, qualitative and intervals for the
MAD calculation [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. The membership functions µ(x) used
by LAMDA are based on the generalization of a
probabilistic rule defined on 0, 1 to the [
        <xref ref-type="bibr" rid="ref1">0,1</xref>
        ]-space.
2.1
      </p>
    </sec>
    <sec id="sec-3">
      <title>Calculation of MAD for quantitative features</title>
      <p>The quantitative type allows the representation of
numerical values, assuming that the including space is known as a
defined interval. For this type of descriptor, membership
functions can be used, such as the Gaussian membership
function so that the membership function for the xth sample
descriptor to the kth class is:
µ ki (xi ) = exp
−(xi −ρ ki )2</p>
      <p>
        2σ i 2
or the binomial membership function:
µ ki (xi ) = ρ ki xi (1 − ρ ki )(1− xi )
(1)
(2)
ρ i ∈ [
        <xref ref-type="bibr" rid="ref1">0, 1</xref>
        ] is the mean of the ith feature based on the
k
samples belonging to the class Ck, xi∈ [
        <xref ref-type="bibr" rid="ref1">0, 1</xref>
        ] is the
normalized xth feature and σi the standard deviation of the ith
feature value based on the samples belonging to the class Ck.
In case of qualitative feature, the possible values of the ith
feature forms a set of modalities such as Di= Q1i , Q2i Qmi
with m the total number of modalities. The qualitative type
permits to express by words the different modalities of a
criterion.
      </p>
      <p>
        The frequency of a modality Qli of the ith feature for the
class Ck is the quantity of samples belonging to Ck whose
modality for their ith feature is Qli [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ].So each modality
Qi ∈ Di has an associated frequency. Let θ kij be the
frel
quency of a modality Qij for the class Ck. The membership
function concerning the ith feature is:
µ ki (xi ) = (θ ki1)q1i * (θ ki 2 )q2i ** (θ kim )qmi
(3)
where qli =1 if xi = Qli and qli = 0 otherwise, for l=1, ..m.
2.3
      </p>
    </sec>
    <sec id="sec-4">
      <title>Calculation of MAD for interval features</title>
      <p>
        Finally, to take in account the potential uncertainties or
noises in data, we can use the interval representation [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ].
The membership function for the interval type descriptors
is regarded as being the similarity S (xi,ρ ki ) between the
symbolic interval value for the ith feature xi and the interval
[ ρ ki− ,ρ ki+ ] which represents the value of the ith feature for
the class Ck, so that:
µ ki (xi ) = S (xi,ρ k )
i
(4)
Let ω be defined as the scalar cardinal of a fuzzy set in a
discrete universe asϖ [X ] = ∑x ∈V µ x (x i ) .
      </p>
      <p>In case of a crisp interval, it becomes:
ϖ [X ] = upperBound(X)- lowerBound(X).</p>
      <p>Given two intervals A=[a-, a+] and B=[b-, b+], the distance
is defined as:</p>
      <p>δ [A, B] = max[0, (max{a − , b− }− min{a + , b+ })]
and the definition of the similarity measure between two
crisp intervals:
S (I1, I 2 ) = 12 ϖϖ [[II11 ∩∪ II 22 ]] + 1 − δ ϖ[I1[V,I]2 ]
The similarity combines the Jaccard's similarity measure
which computes the similarity when the intervals overlapp,
and a second term which allows taking into account the
case where the intervals are not straddled.
2.4</p>
    </sec>
    <sec id="sec-5">
      <title>Calculation of feature weights</title>
      <p>
        It is possible to determine the relevance of a feature to
optimize the separation between classes. The MEMBAS
method [
        <xref ref-type="bibr" rid="ref8 ref9">8, 9</xref>
        ] is a feature weighting method based on a
membership margin. A distinguishable property of this
method is its capability to process problems characterized
by mixed-type data (quantitative, qualitative and interval).
It lies on the maximization of the margins between two
closest classes for each sample. It can be expressed as:
Max ∑ Jj =1 β j (w f ) = 1/N ∑ Jj =1∑iN=1 w fiµ ci (xi( j) )  (7)
w f − ∑iN=1 w fiµ c~i (xi( j) )
Subject to the following constraints: || w f ||22 = 1, w f ≥ 0 .
The first constraint is the normalized bound for the
modulus of wf so that the maximization ends up with
noninfinite values, whereas the second guarantees the
nonnegative property of the obtained weight vector. Then
can be simplified as:
s = 1/ N ∑ Jj =1{U jc − U j~c }
U jc = µ 1c  xi( j) ,,µ cN  xi( j)  , µ ki  xi( j)  is the
membership function of class c (c corresponds to the
“right” class for sample x(j), c~ the closest class evaluated
at the given value xi( j) of the ith feature of pattern x(j). s is
computed with respect to all samples contained in the data
base excluding x(j) (“leave-one-out margin”).
(5)
(6)
(8)
and
      </p>
      <p>
        This optimization problem has an analytical solution
determined by the classical Lagrangian method. Details of
the method can be found in [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ].
3
      </p>
    </sec>
    <sec id="sec-6">
      <title>The new object type</title>
      <p>In order to allow the combination of various data types
into one single global object and therefore to support
multi-dimensional features, we develop a novel data type.
Each feature of an object descriptor can be described by a
measured value and an extrinsic object-related weight. A
sample GAD calculus formula is then the weighted mean
of all MADs:</p>
      <p>
        GADkj = ∑  MADkji .w~ fi  for j=1…J
(9)
where MADkji = MAD of the jth sample for the ith feature
to class k and w~ fi ∈ [
        <xref ref-type="bibr" rid="ref1">0,1</xref>
        ] = Normalized value of weight
w fi of the ith feature determined by the MEMBAS
method, and J is the total number of samples which have been
classified.
to constitute a new sample set called B. Then a
classification of the B samples is processed. Once the classification
of the B samples has been done, its results are used to
compute the classification of A. If the samples of the A set
have C complex features, the second classification level
implies C distinct sample sets B1, B2, … BC thus C distinct
classifications.
      </p>
      <p>
        The MEMBAS algorithm [
        <xref ref-type="bibr" rid="ref8 ref9">8, 9</xref>
        ] can then calculate the
weights of every feature for the classes definition. It is
applied on the B samples so that its involved features
become the weighted components of a meaningful object.
The complex features of an A sample is then a balanced
combination of attributes.
      </p>
      <p>As explained in the Figure 2, the sample Sample1 is
described by X features, including the object-type feature
Desc1,1 . Desc1,1 is described by Desc1,α , Desc1,β, etc.</p>
      <p>
        To get their respective importance Wα , Wβ etc in Desc1,1
description, a previous classification is performed
regarding Desc1,1 as a sample (Sample1,p), so that each weight
can be calculated using the MEMBAS algorithm [
        <xref ref-type="bibr" rid="ref8 ref9">8, 9</xref>
        ].
Once the respective weights of each feature are known,
objects are automatically instantiated to be involved in the
main classification. Desc1,1 is then described in line with
the obtained weights Wα , Wβ and the known values
V1,α , V1,β.
2.5
      </p>
    </sec>
    <sec id="sec-7">
      <title>Evaluation of a classification quality</title>
      <p>The comparison of two classifications can be performed
by measuring their respective compactness and their
separation. Better the classes are compact and separated easier
will be the recognition process.</p>
      <p>
        A method to measure the quality of a partition has been
proposed by [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ]. This index measures the quality partition
in terms of classes compactness and separation. This
partition index is the Clusters Validity Index (CV, Eq.(10))
which depends only on the GADs (membership degree of
an individual to a class) and not explicitly on data values.
      </p>
      <p>CV =</p>
      <p>Dis
N</p>
      <p>*
⋅ Dmin . K
(10)
where N is the total number of individuals in the data base
and K the total number of classes.</p>
      <p>Dis represents the dispersion given by:</p>
      <p>J
∑δ kj ⋅ exp(δ kj )
j =1</p>
      <p>K
Dis = ∑ 1−</p>
      <p>k =1 N ⋅ GADMk ⋅ exp(GADMk )
with: δ kj = GADMk − GADkj ∀j, j ∈ [1, J ]
(11)
and</p>
      <p>GADMk = maxGAD j </p>
      <p>
         k 
*
Dmin is the minimum distance between two classes. This
distance is computed by using the distance d*(A,B)
between two fuzzy sets A and B [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] defined by:
(12)
(13)
      </p>
      <p>J j
d *(A, B) = 1− M [A ∩ B] = 1− j∑= 1min(GADAj ,GADB ) (14)
M [A ∪ B] J j
∑ max(GADAj ,GADB )
j = 1
The highest value of CV corresponds to a better partition.
4</p>
    </sec>
    <sec id="sec-8">
      <title>Application to an agronomical project</title>
      <p>
        The agronomical project aims at developing a diagnosis
system for an optimized water management system and an
efficient distinctive guidance for corn farmers in order to
decrease the use of phytosanitary products and the water
consumption for irrigation. The project involves two
aspects. The first one aims at complementing the benefits of
adopting and implementing the cultural profile techniques
[
        <xref ref-type="bibr" rid="ref28 ref29">28, 29</xref>
        ]. In this context, we perform a classification of
plots based on various agronomic and SAFRAN
meteorological data [
        <xref ref-type="bibr" rid="ref30">30</xref>
        ], so that each plot should mostly belong to
one particular class whose features are known. Thanks to
the provided information stemmed from the classification
results, advice can be offered to the corn farmers
concerning the corn variety they should sow and the schedule they
should follow for an optimized yield. This study includes
two steps which are described in figure 3. The first one
concerns the clustering of a training set of 50 plots, using
the unsupervised LAMDA classification.
      </p>
      <p>The data used for this classification are six distinctive
agronomical descriptors, describing the plots' features and
that are highly involved in their capacity for yield and
water retention, and twenty-one weather features, defining
the meteorological class in which the plot is situated. The
second part of the project will be repeated annually to
update and improve the clustering performed previously by
adding new information returned by the farmers after
harvest. In the following, only the first part is presented.</p>
      <p>Firstly a previous meteorological clustering (A) is
required to realize a realistic plot classification since the
yield of seedling is highly related to the meteorological
conditions. The weather is then regarded as a complex
entity so that it is only one of a plot features. It is based on
the historical meteorological data of the geographical
position corresponding to the studied plot. Those descriptors
refer to the temperature, the quantity of rainfall, and the
evapotranspiration which occurred during three crucial
periods of the year. Each feature is described in several
distinctive ways. For instance, one period temperature is
evaluated according three types of information. This
meteorological clustering is an unsupervised classification
based on weather data covering every single days of the
determined periods during the fifty last years for all the
geolocalized points belonging to the area studied in this
project (South-West of France). In the event that the plot is
part of the training set (studied area), the weather type of
its area is known and the plot classification can be done
directly. Otherwise, the weather type is obtained thanks to
a supervised classification mode (B') delivering the most
appropriate context. In any cases, the weather type is an
object-feature. This hierarchical treatment permits to
regard each meteorological type as a whole and let the
weather contexts follow their natural evolution
independently of agronomical variations. Moreover,
considering the meteorological features as a single global object
permits taking into account the environmental constraints
and getting a realistic model. As we can observe in the
Figure 4, the meteorological clustering (B) has permitted
to divide the area in three sub-areas. The results of
clustering (B) and the meteorological supervised classification
(B’) have been first performed with every sample of the set
and the distribution of the weights between the
meteorological features has been determined.</p>
      <p>The result of this classification is consistent and so, we
can use the obtained classes and weights of the
meteorological features (obtained with MEMBAS) as
objectfeatures in classification (A). To analyze the benefit of
using hierarchical classification, a clustering (A') has been
performed by using the twenty-one meteorological features
separately and the agronomical features (twenty-seven
features taken indistinctly). We can notice that the
prototypes of the classes are highly dependent on the
meteorological classes for clustering (A) while clustering (A') is
mainly influenced by the ground type.</p>
      <p>To enlighten this, we chose arbitrarily two very close
classes containing the similar plots in both clustering. Each
class prototype is described by the mean value of its
marginal degree memberships (MAD). We represent in Figure
5 these prototype parameters for meteorological features
only for both cases (A with diamond and A' with square)
with in abscises, the marginal membership degree for class
1 and in ordinate the same marginal membership degree
for class 2. For a better quantification of the benefits that
the use of the object representation brings, the CV is
systematically calculated in order to determine the better
partition quality. The results are very encouraging since
CV = 0.69 when the meteorological data are regarded as a
whole object and 0.2 when they are treated separately. The
object type representation enables to multiply by more
than 3 this index and therefore the compactness of the
obtained partition.</p>
      <p>The second aspect of our implication in the project deals
with the water utilization of various clusters of farmers
with the aim of forecasting the needs of each cluster and
adjusting the repartition. From this perspective, we realize
an unsupervised classification of a training data-set of
2900 samples described by seven features: distance to the
closest waterway, orientation, altitude… Orientation
concerns cardinal points and we assume that it is not
expressible with different modalities since continuity cannot be
represented by qualitative descriptors. It cannot be a
number nor an interval because of the cyclic form to be kept.
Thus we choose to regard a cluster orientation as an object
composed of two descriptors that correspond to the
coordinates of its cardinal point in a trigonometric circle base.
The orientation of each cluster can take eight different
values: N, NE, E, SE, S, SW, W, and NW, which bring us
to consider eight different combinations. In accordance
with the trigonometrical circle, these eight combinations
are respectively: (0,1), (√2, √2), (1,0), (- √2, √2), (0,-1),
2 2 2 2
(- √2,- √2), (1,0), (√2,- √2).</p>
      <p>2 2 2 2</p>
      <p>Once our results are validated by an expert, the
classification is experimented twice: firstly treating each
descriptor separately and secondly involving the object type.
Such as meteorological data in the first example, the CV is
calculated in order to determine the better partition quality.</p>
      <p>In this case, which implies 2900 samples, CV= 0.08
when abscissa and ordinate are separated, and CV= 0.13
when using an orientation object. As in the first example,
these results show a qualitative gain for the partition when
the object type is used to express the semantically
connected data.</p>
    </sec>
    <sec id="sec-9">
      <title>Conclusion</title>
      <p>This modular architecture allows more flexibility and a
more precise treatment of data. As we can notice with the
previous agronomical classification, the object approach
makes each module able to be managed independently of
the others so that they can evolve autonomously,
depending on their own specific features and contexts. The object
representation permits to preserve multi-dimensionality
and makes fusion of datasets easier. A better overview is
offered since we can percept the variations of each module
distinctively and the evolution of their influences.</p>
      <p>
        As a perspective, an agent-oriented architecture, based
on the multi-agents theory [
        <xref ref-type="bibr" rid="ref31">31</xref>
        ] will be developed so that
each sample could be considered independently of the
others. They would be so able to create classes acting
simultaneously and comparing themselves to the others, so
that the classes definition won’t depend on the samples
order in the file anymore but will directly result from the
samples set definition. This orientation will assure that the
classification result of our method is unique and stable for
a given samples set. We aim at developing some methods
to allow a semantic data processing also.
      </p>
      <p>Proceedings of the 26th International Workshop on Principles of Diagnosis
26</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>D.</given-names>
            <surname>Dubois</surname>
          </string-name>
          and H. Prade editor.
          <source>The three semantics of fuzzy sets</source>
          ,
          <source>Fuzzy sets and systems</source>
          , vol.
          <volume>90</volume>
          ,N°
          <volume>2</volume>
          ,
          <fpage>pp141</fpage>
          -
          <lpage>150</lpage>
          ,Elsevier, London,
          <year>1997</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>L.</given-names>
            <surname>Hedjazi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Aguilar-Martin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.V.</given-names>
            <surname>Le Lann</surname>
          </string-name>
          and
          <string-name>
            <given-names>T.</given-names>
            <surname>Kempowsky</surname>
          </string-name>
          ,
          <article-title>Towards a unified principle for reasoning about heterogeneous data: a fuzzy logic framework</article-title>
          ,
          <source>International Journal of Uncertainty, Fuzziness and Knowledge-Based Systems</source>
          , vol.
          <volume>20</volume>
          ,N°2,pp.
          <fpage>281</fpage>
          -
          <lpage>302</lpage>
          , World Scientific,
          <year>2012</year>
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>R.S.</given-names>
            <surname>Michalski</surname>
          </string-name>
          and R.E. Stepp,
          <article-title>automated construction of classifications: Conceptual clustering versus numerical taxonomy</article-title>
          ,
          <source>IEEE Trans. Pattern Anal. Machine Intell</source>
          ., vol. PAMI-
          <volume>5</volume>
          , no.
          <issue>4</issue>
          (
          <issue>1980</issue>
          ), pp.
          <fpage>396</fpage>
          -
          <lpage>410</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>D.W.</given-names>
            <surname>Aha</surname>
          </string-name>
          ,
          <article-title>Tolerating noisy, irrelevant and novel attributes in instance based learning algorithms</article-title>
          , Int.
          <source>ManMachine Studies</source>
          <volume>36</volume>
          (
          <year>1992</year>
          ), pp.
          <fpage>267</fpage>
          -
          <lpage>287</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>S.</given-names>
            <surname>Cost</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Salzberg</surname>
          </string-name>
          ,
          <article-title>A weighted nearest neighbor algorithm for learning with symbolic features</article-title>
          ,
          <source>Machine learning (10)</source>
          (
          <year>1993</year>
          ), pp.
          <fpage>57</fpage>
          -
          <lpage>78</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>T.</given-names>
            <surname>Mohri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Hidehiko</surname>
          </string-name>
          ,
          <article-title>An optimal Weighting Criterion of case indexing for both numeric and symbolic attributes</article-title>
          , in D.W. Aha (Ed.),
          <article-title>Case-based Reasoning: papers from the 1994 workshop</article-title>
          . Menlo Park, CA:AIII Press, pp.
          <fpage>123</fpage>
          -
          <lpage>127</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>C.</given-names>
            <surname>Giraud-Carrier</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Tony</surname>
          </string-name>
          ,
          <article-title>An Efficient Metric for heterogeneous Inductive Learning Applications in the Attribute-Value Language, Intelligent systems (</article-title>
          <year>1995</year>
          )
          <fpage>341</fpage>
          -
          <lpage>350</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>K.C.</given-names>
            <surname>Gowda</surname>
          </string-name>
          , E. Diday,
          <article-title>Symbolic clustering using a new similarity measure</article-title>
          ,
          <source>IEEE Trans. SMC</source>
          <volume>22</volume>
          (
          <issue>2</issue>
          ) (
          <year>1992</year>
          )
          <fpage>368</fpage>
          -
          <lpage>378</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>Q.H</given-names>
            <surname>Hu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.X.</given-names>
            <surname>Xie</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.R.</given-names>
            <surname>Yu</surname>
          </string-name>
          ,
          <article-title>Hybrid attribute reduction based on a novel fuzzy-rough model and information granulation</article-title>
          ,
          <source>Pattern Recognition</source>
          <volume>40</volume>
          (
          <year>2007</year>
          )
          <fpage>3509</fpage>
          -
          <lpage>3521</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>F.A.</given-names>
            <surname>T. De</surname>
          </string-name>
          <string-name>
            <given-names>Carvalho</given-names>
            ,
            <surname>R.M.C.R. De Souza</surname>
          </string-name>
          ,
          <article-title>Unsupervised Pattern Recognition Models for Mixed Feature-Type Symbolic Data</article-title>
          ,
          <source>Pattern Recognition Letters</source>
          <volume>31</volume>
          (
          <year>2010</year>
          )
          <fpage>430</fpage>
          -
          <lpage>443</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <surname>I. Kononenko</surname>
          </string-name>
          , Estimating Attributes:
          <article-title>Analysis and Extensions of Relief</article-title>
          ,
          <source>Proc. European Conf. Mach. Learning ECML</source>
          (
          <year>1994</year>
          ), pp.
          <fpage>171</fpage>
          -
          <lpage>182</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>K.</given-names>
            <surname>Kira</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Rendell</surname>
          </string-name>
          ,
          <article-title>A practical approach to feature selection</article-title>
          .
          <source>In proced. 9th Int'l Workshop on Machine Learning</source>
          (
          <year>1992</year>
          ), pp.
          <fpage>249</fpage>
          -
          <lpage>256</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>M.</given-names>
            <surname>Dash</surname>
          </string-name>
          , H. Liu,
          <article-title>Consistency-based search in feature selection</article-title>
          ,
          <source>Artif. Intell</source>
          .
          <volume>151</volume>
          (
          <year>2003</year>
          )
          <fpage>155</fpage>
          -
          <lpage>176</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>D.W.</given-names>
            <surname>Aha</surname>
          </string-name>
          ,
          <article-title>Incremental, instance-based learning of independent and graded concept descriptions</article-title>
          ,
          <source>in Proced. Of the 6th int'l Mach. Learning Workshop</source>
          . (
          <year>1989</year>
          )
          <fpage>387</fpage>
          -
          <lpage>391</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>J.</given-names>
            <surname>Weston</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Mukherjee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Chapelle</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Pontil</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Vapnik</surname>
          </string-name>
          , Feature Selection for SVMs,
          <source>Advances in Neural Information Processing Systems</source>
          (
          <year>2001</year>
          ), pp.
          <fpage>668</fpage>
          -
          <lpage>674</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>T.</given-names>
            <surname>Cover</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Hart</surname>
          </string-name>
          ,
          <article-title>Nearest neighbor pattern classification</article-title>
          ,
          <source>IEEE Trans. Inf. Theory</source>
          <volume>13</volume>
          (
          <year>1967</year>
          ), pp.
          <fpage>21</fpage>
          -
          <lpage>27</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>H.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Hussian</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.L.</given-names>
            <surname>TAM</surname>
          </string-name>
          , M. Dash,
          <article-title>Discretization: an enabling technique</article-title>
          ,
          <source>J. Data Mining and Knowledge Discovery</source>
          <volume>6</volume>
          (
          <issue>4</issue>
          ) (
          <year>2002</year>
          )
          <fpage>393</fpage>
          -
          <lpage>423</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>M.A.</given-names>
            <surname>Hall</surname>
          </string-name>
          ,
          <article-title>Correlation-based Feature Selection for Discrete and Numeric Class Machine Learning</article-title>
          ,
          <source>Int. Conf. Mach. Learning ICML</source>
          (
          <year>2000</year>
          ), pp.
          <fpage>359</fpage>
          -
          <lpage>366</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>H.H.</given-names>
            <surname>Bock</surname>
          </string-name>
          ,
          <string-name>
            <surname>Diday E.</surname>
          </string-name>
          <article-title>Analysis of Symbolic Data, Exploratory methods for extracting statistical information from complex data</article-title>
          . (Springer, Berlin Heidelberg,
          <year>2000</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>V.</given-names>
            <surname>Vapnik</surname>
          </string-name>
          ,
          <source>The Nature of Statistical Learning Theory Data Mining and Knowledge Discovery</source>
          ,pp.
          <fpage>1</fpage>
          -
          <issue>47</issue>
          , 6, Springer Verlag,
          <year>1995</year>
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>T.</given-names>
            <surname>Cover</surname>
          </string-name>
          and
          <string-name>
            <given-names>P.</given-names>
            <surname>Hart</surname>
          </string-name>
          ,
          <article-title>Nearest neighbor pattern classification Information Theory</article-title>
          , IEEE Transactions on,
          <volume>13</volume>
          ,
          <issue>1</issue>
          ,pp.
          <fpage>21</fpage>
          -
          <lpage>27</lpage>
          ,
          <year>1967</year>
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <surname>Michie</surname>
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Spiegelhalter</surname>
            <given-names>D.J.</given-names>
          </string-name>
          , Taylor C.C.,
          <string-name>
            <surname>Machine</surname>
            <given-names>Learning</given-names>
          </string-name>
          ,
          <source>Neural and Statistical Classification, Ellis Horwood series in Artificial Intelligence, february</source>
          ,
          <year>1994</year>
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <surname>Rakotomalala</surname>
            <given-names>R.</given-names>
          </string-name>
          ,
          <source>Decision Trees, review MODULAD</source>
          ,
          <volume>33</volume>
          ,
          <year>2005</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [24]
          <string-name>
            <given-names>J. C.</given-names>
            <surname>Aguado</surname>
          </string-name>
          and
          <string-name>
            <given-names>J.</given-names>
            <surname>Aguilar-Martin</surname>
          </string-name>
          ,
          <article-title>A mixed qualitative-quantitative self-learning classification technique applied to diagnosis</article-title>
          ,
          <source>The Thirteenth International Workshop on Qualitative Reasoning</source>
          , (QR'99)pp.
          <fpage>124</fpage>
          -
          <lpage>128</lpage>
          ,
          <year>1999</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          [25]
          <string-name>
            <given-names>L.</given-names>
            <surname>Hedjazi</surname>
          </string-name>
          , J. Aguilar- Martin,
          <string-name>
            <given-names>M.V.</given-names>
            <surname>Le Lann</surname>
          </string-name>
          ,
          <article-title>Similarity-margin based feature selection for symbolic interval data</article-title>
          ,
          <source>Pattern Recognition Letters</source>
          , Vol.
          <volume>32</volume>
          , N°4, pp.
          <fpage>578</fpage>
          -
          <lpage>585</lpage>
          ,
          <year>2012</year>
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          [26]
          <string-name>
            <given-names>L.</given-names>
            <surname>Hedjazi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Aguilar-Martin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.V.</given-names>
            <surname>Le Lann</surname>
          </string-name>
          , and
          <string-name>
            <surname>Tatiana</surname>
          </string-name>
          Kempowsky-Hamon,
          <article-title>Membership-Margin based Feature Selection for Mixed Type and High-dimensional Data: Theory and Applications</article-title>
          , Information Sciences, accepted to be published,
          <year>2015</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          [27]
          <string-name>
            <given-names>C. V.</given-names>
            <surname>Isaza</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H. O.</given-names>
            <surname>Sarmiento</surname>
          </string-name>
          , ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Kempowsky-Hamon</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.V.</given-names>
            <surname>Le Lann</surname>
          </string-name>
          ,
          <article-title>Situation prediction based on fuzzy clustering for industrial complex processes</article-title>
          ,
          <source>Information Sciences</source>
          , Volume
          <volume>279</volume>
          ,
          <issue>20</issue>
          <year>September 2014</year>
          , pp.
          <fpage>785</fpage>
          -
          <lpage>804</lpage>
          ,
          <year>2014</year>
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          [28]
          <string-name>
            <surname>Henin</surname>
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gras</surname>
            <given-names>R.</given-names>
          </string-name>
          , Monnier G.,
          <year>1969</year>
          ,
          <article-title>Le profil cultural (2e edition</article-title>
          ), Masson Ed. Paris.
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          [29]
          <string-name>
            <surname>Gautronneau</surname>
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gigleux</surname>
            <given-names>C.</given-names>
          </string-name>
          ,
          <year>2002</year>
          ,
          <article-title>Towards holistic approaches for soil diagnosis in organic orchards</article-title>
          ,
          <source>Proceedings of the 14th IFOAM Organic World Congress, Victoria</source>
          , p
          <fpage>34</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          [30]
          <string-name>
            <surname>Quintana-Seguí</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Le</surname>
            <given-names>Moigne</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            ,
            <surname>Durand</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            ,
            <surname>Martin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            ,
            <surname>Habets</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            ,
            <surname>Baillon</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            , ... &amp;
            <surname>Morel</surname>
          </string-name>
          ,
          <string-name>
            <surname>S.</surname>
          </string-name>
          (
          <year>2008</year>
          ).
          <article-title>Analysis of near-surface atmospheric variables: Validation of the SAFRAN analysis over France</article-title>
          .
          <source>Journal of applied meteorology and climatology</source>
          ,
          <volume>47</volume>
          (
          <issue>1</issue>
          ),
          <fpage>92</fpage>
          -
          <lpage>107</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref31">
        <mixed-citation>
          [31]
          <string-name>
            <surname>Ferber</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          (
          <year>1999</year>
          ).
          <article-title>Multi-agent systems: an introduction to distributed artificial intelligence</article-title>
          (Vol.
          <volume>1</volume>
          ). Reading: Addison-Wesley.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>