<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>C. Marsala) - ORCID ID:</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Building Intuitionistic Fuzzy Sets in Machine Learning</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Christophe Marsala</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Sorbonne Université</institution>
          ,
          <addr-line>CNRS, LIP6, F-75005 Paris</addr-line>
          ,
          <country country="FR">France</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2021</year>
      </pub-date>
      <volume>000</volume>
      <fpage>0</fpage>
      <lpage>0002</lpage>
      <abstract>
        <p>The construction of intuitionistic fuzzy sets is a di!cult task. Some approaches have been proposed in the literature and they have been used successfully in some application domains. However, these approaches do barely take into account the representativeness of the data used to build the intuitionistic fuzzy set. In this paper, a new approach is proposed to build intuitionistic fuzzy sets (IFS). This approach is based on the use of a representativeness degree of the data. This approach enables to build an IFS with an intuitionistic fuzzy index that is a good indicator of the lack of knowledge associated with the data that make it a good approach to be used in a Machine learning setting.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Intuitionistic fuzzy sets</kwd>
        <kwd>Machine learning</kwd>
        <kwd>lack of knowledge</kwd>
        <kwd>representativeness</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>In this paper, we propose a new approach to construct IFS from a dataset that takes care of the
speci#city of the Machine learning setting. Firstly, in Section 2 a recall of existing approaches
is done. Secondly, in Section 3, the proposed approach is introduced. Thirdly, in Section 4, an
experimental study is presented to bring out some main properties of the proposed approach.
Finally, a conclusion and some future work are presented.
2. Construction of intuitionistic fuzzy sets: existing approaches
In this section, we focus on existing approaches to build IFS. The following notations are used.
Let U = {u1, . . . , un} be a discrete universe and let A be a subset of U , and let P = {p1, . . . , pn}
be a probability distribution on U with 0 ≤ pi+1 &lt; pi ≤ 1 and Pin=1 pi = 1.</p>
      <p>The question that should be answered is “how to de"ne the intuitionistic fuzzy set A of U
from P ?” There does not exist a lot of approaches in the literature to answer this question, we
present hereafter the 2 main ones. First of all we present the approach based on the de#nition
of a mass assignment [7], and afterwards we present the approach by [6] based on the use of an
intuitionistic fuzzy generator.</p>
      <sec id="sec-1-1">
        <title>2.1. Basic recalls</title>
        <sec id="sec-1-1-1">
          <title>Intuitionistic fuzzy sets have been introduced by Atanassov [1].</title>
          <p>An IFS A of U is de#ned as: A = {(u, μA(u), νA(u)) | u ∈ U } with μA : U → [0, 1] and
νA : U → [0, 1] such that for all u ∈ U , 0 ≤ μA(u) + νA(u) ≤ 1. The values μA(u) and νA(u)
are, respectively, the membership degree and the non-membership degree of u to A.</p>
          <p>The margin of hesitancy (or intuitionistic fuzzy index) of u to A is de#ned as πA(u) = 1 −
(μA(u) + νA(u)), it informs about the lack of knowledge about A when it exists. When the
margin of hesitancy according to A is equal to zero for all u, that is to say when μA(u)+νA(u) =
1, ∀u ∈ U , then A is a Zadeh’s fuzzy set.</p>
        </sec>
      </sec>
      <sec id="sec-1-2">
        <title>2.2. The mass assignment approach</title>
        <p>This approach has been introduced in [7] and detailed also in [8]. It has been used in Machine
learning applications presented in [9] and [5].
2.2.1. Mass assignment
From a probability distribution P , a mass assignment mA of a fuzzy set A of U can be de#ned
thanks to the mass assignment theorem introduced in [10] that enables such a construction:
where Fi = {u ∈ U | p(u) ≥ pi} and
mA(Fi) =
mA(Fn) =
μi − μi+1 for i = 1, . . . , n − 1
μn
μi = |Fi|pi +</p>
        <p>n
X (|Fj | − |Fj−1|)pj .</p>
        <p>j=i+1
2.2.2. Algorithm to build an IFS from a mass assignment
In [7], the building of an IFS A = {(u, μA(u), νA(u)) | u ∈ U } could be done from two
independent probability distributions on U : P + the probability distribution connected to the
membership of the elements of U to the IFS A, and P − the probability distribution connected
to their non-membership to the IFS A. The process is done according to the following steps [7]:
1. On the one hand, the mass assignment mA+ is build from P + with the mass assignment
theorem (see Section 2.2.1). As stated in [7]: mA+(u) is the possibility that u has the value
mA+(u) and thus, it is considered that mA+(u) = μA(u) + πA(u).
2. On the other hand, the mass assignment mA− is build from P − with the mass assignment
theorem. mA−(u) is the possibility that u has the value mA−(u) and it is considered that
mA−(u) = νA(u) + πA(u).
3. Finally, the aggregation of mA+ and mA− enables to obtain μA and νA taking into account
that, for all u ∈ U , μA(u) + νA(u) + πA(u) = 1, it gives mA+(u) + mA−(u) = 1 + πA(u)
and thus for all u ∈ U , πA(u) = mA+(u) + mA−(u) − 1 that leads to the determination of
the values μA(u) and νA(u).</p>
        <p>By means of the above process, the IFS A ⊆ U can be completely de#ned from P + and P −.
2.2.3. Discussion
This approach to de#ne the IFS A could have the drawback to produce a negative value for
πA(u) as there is no guarantee that mA+(u) + mA−(u) ≥ 1 for all u ∈ U . Indeed, an IFS could
not necessarily be build from any probability distributions and IFS could not be a convenient
model in this case.</p>
        <p>However, one solution when facing a negative value for πA(u) is to set it to 0, it is the solution
that has been used in the experimental part of this paper.</p>
        <p>In a Machine learning setting, this approach to de#ne the IFS A is interesting because it does
not need any hyper-parameter to be set.</p>
      </sec>
      <sec id="sec-1-3">
        <title>2.3. The fuzzy generator approach</title>
        <p>The fuzzy generator approach to build an IFS that is described in this section has been proposed
in [6]. This approach has been mainly used in image segmentation [11] [4], or in clustering
problems [12] to cite some examples of its use.
2.3.1. Intuitionistic fuzzy generator
The de#nition of an intuitionistic fuzzy generator (IFG) has been introduced in [6]: a function
Φ : [0, 1] → [0, 1] is called an intuitionistic fuzzy generator if Φ(x) ≤ 1 − x for all x ∈ [0, 1].</p>
        <p>An example of such a generator is:
1 − x
1 + λx
N (x) =
, with λ &gt; 0
where N (0) = 1 and N (1) = 0.
2.3.2. Construction of an IFS by means of an IFG
An IFS can be de#ned from a fuzzy set by means of an IFG [6]: let A = {(u, μA(u)) | u ∈
U } be a fuzzy sets on U , and let Φ be an intuitionistic fuzzy generator, then the set A˜ =
{(u, μA(u), Φ(μA(u))) | u ∈ U } is an intuitionistic fuzzy set on U .
2.3.3. Discussion
In [4] and in [12], the IFG presented in Section 2.3.1 is used to build the IFS, but another IFG
can be used, see for instance [11].</p>
        <p>In a Machine learning setting, this approach to build IFS is very promising, in particular
in non-supervised Machine learning problems. However, the choice of at least one
hyperparameter, the IFG, should be done. Moreover, if the IFG given in Section 2.3.1 is used, it is also
mandatory to choose a good value for λ.</p>
      </sec>
    </sec>
    <sec id="sec-2">
      <title>3. A new approach to build an IFS</title>
      <p>The proposed approach is dedicated to the construction of an IFS in a machine learning setting
(but not limited to this kind of application domain). As a consequence, we consider that the
information provided to build the IFS is not only probability distributions but a complete set of
instances separated into two classes.</p>
      <p>Let X be a universe of values and U = {u1, . . . , un} ⊆ X be a discrete subset of X , and let
C = {+, −} be a set of classes over the elements of X . For i = 1, . . . , n, let ni+ be the number
of instances of ui that are associated with class +, and let ni− be the number of instances of
ui that are associated with class −. We denote by n+ = Pin=1 ni+ and n− = Pin=1 ni− and we
assume in the following that n+ 6= 0 and n− 6= 0. Moreover, we denote by ni = ni+ + ni−,
∀i = 1, . . . , n and we assume that ni 6= 0, ∀i = 1, . . . , n.</p>
      <p>In the following, a training set is the tuple hU , C, (n1+, . . . , nn+), (n1−, . . . , nn−)i.
The set of values n1+, . . . , nn+ (resp. +n1−, . . . , nn−) de#nes a probability distribution P +
−
(resp. P −) over U such that P +(ui) = nni+ (resp. P −(ui) = nni− ).</p>
      <p>These two probability distributions provide us with information about the elements of U and
their membership (P +) or non-membership (P −) to a set A ⊆ U that we want to build.</p>
      <p>Our approach aims at de#ning A as an intuitionistic fuzzy set of U according to the knowledge
that is provided by a training set. This approach is composed of two main steps. First of all,
from the training set two corresponding weighted distributions are built taking into account
the representativeness of the training set. Secondly, the IFS A is built using these two weighted
distributions. The process is detailed in the following.</p>
      <sec id="sec-2-1">
        <title>3.1. Representativeness and weighted distributions</title>
        <p>Given a training set T S = hU , C, (n1+, . . . , nn+), (n1−, . . . , nn−)i, our approach aims at building
an IFS A over U . As usual in machine learning, the training set provides only a restricted view
about X , thus this brings out the question of its representativeness.
3.1.1. Representativeness degree
To highlight the representativeness of a training set, we introduce the use of a degree that
should be set either by a user that knows the problem under study, or automatically by means
of an objective decision process. This degree highlights how the training set can be considered
as su!ciently representative to infer knowledge that could be generalised to X .</p>
        <p>The representativeness degree ρ ∈ [0, 1] of the training set hU , C, (n1+, . . . , nn+), (n1−, . . . , nn−)i
is such that:
• ρ = 0 when the training set is not representative of X . In this case, the knowledge it
provides are not usable for any elements of X not in U .
• ρ = 1 when the training set is completely representative of X . In this case, the knowledge
it provides is completely usable for any elements of X .</p>
        <p>• the greater ρ, the more representative the training set.</p>
        <p>The representativeness degree evaluates how much we could be con#dent in the fact that
the probability distributions induced by (ni+)i=1,..,n and (ni−)i=1,..,n re$ects the respective
probability distributions on X . This representativeness degree is either given by the user
that appreciate how the knowledge provided by U could be generalised to X , or it could be
determined automatically (to be studied in future work).
3.1.2. Lack-of-knowledge degree
The representativeness degree is an information about the training set. It could be used to
weight the information associated with any example u ∈ U provided in the training set.
De!nition 1. Let T S = hU , C, (n1+, . . . , nn+), (n1−, . . . , nn−)i be a training set of X , and ρ ∈ [0, 1]
be the representativeness degree of T S. The lack-of-knowledge degree of ui ∈ U is de"ned as
ni
l(ui) = ρ ∗ nmax with nmax = sup{i=1,...,n} ni.</p>
        <p>The lack-of-knowledge degree takes into account not only the representativeness of the
training set, but also, for a given property (membership or non-membership) the number of
elements that possess this property. This is represented by the use of nmax = sup{i=1,...,n} ni to
weight the value ni. This degree enables us to take into account the representativeness of U to
reduce the in$uence of an example used for the construction of the IFS A from P + and P −.
ni</p>
        <p>It is easy to see that, for all i = 1, . . . , n, l(ui) ∈ [0, 1] as ρ ∈ [0, 1] and 0 ≤ nmax ≤ 1.
3.1.3. Weighted probability distributions
The lack-of-knowledge degree is used to weight the in$uence of the information related to an
example u with regard to the global information brought out by the training set.
De!nition 2. Let T S = hU , C, (n1+, . . . , nn+), (n1−, . . . , nn−)i be a training set of X , and ρ ∈ [0, 1]
be the representativeness degree of T S. The weighted probability distributions over U related to
+ −
the classes C are de"ned ∀i = 1, . . . , n as: Pw+(ui) = l(ui) nnii and Pw−(ui) = l(ui) nnii .
+ −</p>
        <p>It is easy to see that, for all i = 1, . . . , n, as l(ui) ∈ [0, 1] and nnii ∈ [0, 1] (resp. nnii ∈ [0, 1]),
we have Pw+(ui) ∈ [0, 1] (resp. Pw−(ui) ∈ [0, 1]).
3.2. Construction of an IFS
From a training set, associated with a representativeness degree, it is possible to de#ne an
intuitionistic fuzzy set taking into account the information related to the classes: class +
representing the information related to the membership of elements to A and class − representing
the information related to their non-membership to A.</p>
        <p>De!nition 3. Let T S = hU , C, (n1+, . . . , nn+), (n1−, . . . , nn−)i be a training set of X , and ρ ∈ [0, 1]
be the representativeness degree of T S. The intuitionistic fuzzy set A is de"ned as</p>
        <p>A = {(u, Pw+(u), Pw−(u)) | u ∈ U }.</p>
        <sec id="sec-2-1-1">
          <title>Indeed, it is easy to see that A is an IFS as we have for all ui ∈ U :</title>
          <p>• Pw+(ui) ∈ [0, 1] and Pw−(ui) ∈ [0, 1] (see Section 3.1.3);</p>
          <p>+ −
• Pw+(ui) + Pw−(ui) = l(ui) nnii + l(ui) nnii = l(ui) and 0 ≤ l(ui) ≤ 1 (see Section 3.1.2).</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>4. Experimental study</title>
      <p>A preliminary experimental study is presented in this section to study the in$uence of the
representativeness degree, and to compare the IFS built by means of each of the presented
approaches.</p>
      <p>The experiments are conducted on a training set proposed in [7]. The set U contains 10
elements associated with di"erent sizes ni. In [7], two probability distributions on U are
given to represent the membership and the non-membership to the IFS A to build. In our
experiments, these probability distributions have been used to generate the values (n1+, . . . , nn+),
and (n1−, . . . , nn−) by considering that the total size of the set + (resp. −) is 1000 elements. The
resulting training set is presented in the 2 #rst columns in Table 1.</p>
      <p>A global analysis of this training set leads us to consider that, among the elements u, there
are some elements that are more representative than others. For instance, u1 has only a size of
125 where u5 is associated with a size of 351. Indeed, this highlights the fact that we could be
less con#dent in the decision that could be drawn from this training set for u1 than for u5.</p>
      <sec id="sec-3-1">
        <title>4.1. Comparison of approaches</title>
        <p>In Table 1, we present the IFS built by each of the 3 approaches presented in this paper: our
proposed approach with ρ = 1 (complete representativeness of the training set), the approach
by [7], and the approach by [6] with λ = 0.5. Concerning this last approach, as it takes into
account only one probability distribution, the given IFS is obtained with only the information
related to +.</p>
        <p>The three IFS built by these three approaches are very di"erent. It should be seen that
• the IFS built by [6] does not provide a great hesitancy (intuitionistic indices π) unlike the
other 2 approaches. In this approach, this hesitancy is related to the choice of λ, but it
takes into account only one information from the training set: the probability distribution
associated with the membership of the elements.</p>
        <p>• the IFS built by [7] provides an hesitancy, however, it does not take into account the
representativeness of each element.
• the IFS built by the proposed approach highlights the importance of the information
provided on each element. This leads to an hesitancy more important for u1 than for u5
for instance.</p>
      </sec>
      <sec id="sec-3-2">
        <title>4.2. Influence of the representativeness degree</title>
        <p>In order to study the in$uence of the representativeness degree, the proposed approach has
been used with di"erent values of this degree for the same training set. The training is similar
to the one detailed in the previous section. The resulting IFS obtained with the approach used
with increasing values of the representativeness degree ρ is presented in Table 2. This table
should be completed with the result for ρ = 1 given in Table 1.</p>
        <p>These results show the great in$uence of the representativeness degree in the resulting IFS,
this in$uence can be highlighted by the values of the intuitionistic indices π of the resulting
IFS: the greater the representativeness degree, the lower the values of the intuitionistic indices.
This perfectly highlights the behaviour of the approach to build an IFS that is needed in
a Machine learning setting: the intuitionistic fuzzy index is a representation of the lack of
knowledge that is associated with the membership of an element u to the IFS A [13], therefore
the representativeness of the training set induces the lack of knowledge that is associated with
the data belonging to this training set.</p>
        <p>A main question arises here on ”how to choose a good value for ρ?”. This value could be set by
the user if he/she has su!cient knowledge to evaluate the representativeness of the training set.
Otherwise, ρ could also be set by means of an automatic approach, for instance, by considering
the size of the training set with regard to the dimension of X .</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>5. Conclusion</title>
      <p>In this paper, the problem of building intuitionistic fuzzy sets is tackled. After a survey of the
two main existing approaches from the literature to build IFS from probability distributions, a
new approach is proposed. The aim of this new approach is to be used in a Machine learning
setting where it is important to take into account the representativeness of the training set.</p>
      <p>The proposed approach is based on the use of a representativeness degree of the given training
set and the determination of a lack-of-knowledge degree to weight the importance of each
element of the training set. This approach enables to build an IFS where the intuitionistic fuzzy
index is a good indicator of the lack of knowledge associated with the given training set.</p>
      <p>In future work, even if a good choice could be to set ρ to 1, a #rst study to conduct will be to
#nd a way to make the choice of a good value for ρ, for instance by means of some automatic
ways to set this value. Secondly, the resulting IFS will be used in a complete Machine learning
approach, to build a classi#cation model from a given training set. Moreover, future work
will also deepen the study on the proposed approach in term of convergence property and the
complexity</p>
    </sec>
    <sec id="sec-5">
      <title>Acknowledgments References</title>
      <p>The author would like to thank the anonymous reviewers for their valuable suggestions.
[1] K. T. Atanassov, Intuitionistic fuzzy sets, Fuzzy Sets and Systems 20 (1986) 87–96.
[2] I. Couso, H. Bustince, From fuzzy sets to interval-valued and Atanassov intuitionistic
fuzzy sets: A uni#ed view of di"erent axiomatic measures, IEEE Transactions on Fuzzy
Systems 27 (2019) 362–371.
[3] Y. Bengio, Y. Lecun, G. Hinton, Deep learning for AI, Communications of the ACM 64
(2021) 58–65.
[4] A. Bouchet, S. Montes, V. Ballarin, I. Díaz, Intuitionistic fuzzy set and fuzzy mathematical
morphology applied to color leukocytes segmentation, Signal, Image and Video Processing
14 (2020) 557–564.
[5] E. Szmidt, J. Kacprzyk, P. Bujnowski, Three term attribute description of Atanassov’s
intuitionistic fuzzy sets as a basis of attribute selection, in: Int. Conf. FUZZ-IEEE, 2021.
[6] H. Bustince, J. Kacprzyk, V. Mohedano, Intuitionistic fuzzy generators application to
intuitionistic fuzzy complementation, Fuzzy Sets and Systems 114 (2000) 485–504.
[7] E. Szmidt, J. F. Baldwin, Intuitionistic fuzzy set functions, mass assignment theory,
possibility theory and histograms, in: IEEE Int. Conf. on Fuzzy Systems, 2006, pp. 35–41.
[8] E. Szmidt, M. Kukier, A new approach to classi#cation of imbalanced classes via atanassov’s
intuitionistic fuzzy sets, in: Intelligent Data Analysis: Developing New Methodologies
Through Pattern Discovery and Recovery, IGI Global, 2009, pp. 85–101.
[9] E. Szmidt, J. Kacprzyk, P. Bujnowski, Attribute selection for sets of data expressed by
intuitionistic fuzzy sets, in: 2020 IEEE International Conference on Fuzzy Systems
(FUZZIEEE), 2020, pp. 1–7.
[10] J. F. Baldwin, J. Lawry, T. P. Martin, The application of generalised fuzzy rules to
machine learning and automated knowledge discovery, International Journal of Uncertainty,
Fuzziness and Knowledge-Based Systems 6 (1998) 459–487.
[11] H. Bustince, V. Mohedano, E. Barrenechea, M. Pagola, An algorithm for calculating the
threshold of an image representing uncertainty through A-IFSs, in: Int. Conf. IPMU, 2006,
pp. 2383–2390.
[12] F. Yang, Z. Liu, X. Bai, Y. Zhang, An improved intuitionistic fuzzy c-means for ship
segmentation in infrared images, IEEE Transactions on Fuzzy Systems (2020).
[13] C. Marsala, B. Bouchon-Meunier, Interpretable monotonicities for entropies of intuitionistic
fuzzy sets or interval-valued fuzzy sets, in: Joint Proc. of the 19th World Congress of the
IFSA, the 12th Conf. of the EUSFLAT, and the 11th International Summer School on AGOP,
Atlantis Press, 2021, pp. 48–54.</p>
    </sec>
  </body>
  <back>
    <ref-list />
  </back>
</article>