<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>E. Di Nardo); angelo.ciaramella@uniparthenope.it (A. Ciaramella)
~ https://sites.google.com/view/ciss-angelociaramella/home (A. Ciaramella)</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Advanced Fuzzy Relational Neural Network</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>E. Di Nardo</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>A. Ciaramella</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Department of Computer Science, University of Milan</institution>
          ,
          <addr-line>Milan 20122</addr-line>
          ,
          <country country="IT">Italy</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Department of Science and Technology, University of Naples Parthenope, Centro Direzionale Isola C4</institution>
          ,
          <addr-line>I-80143, Napoli</addr-line>
          ,
          <country country="IT">Italy</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2021</year>
      </pub-date>
      <volume>000</volume>
      <fpage>0</fpage>
      <lpage>0002</lpage>
      <abstract>
        <p>Nowadays most of the researches aimed for studying artificial neural networks and in particular convolutional neural networks for the impressive results in several scientific fields. However, these methodologies need of post-hoc technique for improving their interpretability and explainability. In the last years, fuzzy systems are raising great interest for the simplicity to develop trustworthy and explainable systems. This work aims to introduce a fuzzy relational neural network based model for extrapolating relevant information from images data permitting to obtain a clearer indication on the classification processes. Encouraging results are obtained on benchmark data sets.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Deep Learning</kwd>
        <kwd>Fuzzy Logic</kwd>
        <kwd>Fuzzy Relational Neural Network</kwd>
        <kwd>Computational Intelligence</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
    </sec>
    <sec id="sec-2">
      <title>2. Fuzzy Relational Neural Network model</title>
      <p>
        Fuzzy Rule-based Systems (FRSs), are raising great interest in XAI in the last years as ante-hoc
methodologies [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. The main components of any FRS are the knowledge base (KB) and the
inference engine module. The KB comprises all the fuzzy rules within a rule base (RB), and
the definition of the fuzzy sets in the data base. The inference engine includes a fuzzification
interface, an inference system, and a defuzzification interface [
        <xref ref-type="bibr" rid="ref4 ref5">5, 4</xref>
        ]. Fuzzy Relational Neural
Network (FRNN) [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] is and adaptive model based on a FRS. FRNN can be developed with
diferent norms and a backpropagation algorithm is used for learning. In this work we model
local t-norms modifying the inner operation of convolution and replacing the linear combination
provided by matrix multiplication with fuzzy operators. We define a receptive field that applies
a triangular operation on a restricted area. As happen in convolution we have a kernel of size
 ×  ×  ×  where  and  are the spatial dimensions,  is the number of input
channels and  the number of output features maps. Kernel slides over the image with a
parametric step. Weights are initialized in range [
        <xref ref-type="bibr" rid="ref1">0, 1</xref>
        ] and constrained to be in the same interval
after the optimization step using a scaling operation based on minimum and maximum values:
 =
      </p>
      <p>
        − min()
max() − min()
(1)
The network structure is composed by an input layer and a fuzzification layer where the
membership function is just a scaling of the pixel value in range [
        <xref ref-type="bibr" rid="ref1">0 − 1</xref>
        ]. We compare the results
by using one or two hidden layers. Next there is a defuzzification operation that is composed
by a fully connected layer like in [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] and an output layer with a Categorical Crossentropy is
used for classification. Architectures have been tested with and without a threshold activation
function, a modification of leaky relu with a minimum boundary &gt; 0. Networks are compared
with equivalent CNN architectures.
      </p>
    </sec>
    <sec id="sec-3">
      <title>3. Experimental results</title>
      <p>
        FRNN has been applied for images classification and the MNIST [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] and CIFAR10 [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] datasets are
considered. Input images are scaled in range [
        <xref ref-type="bibr" rid="ref1">0 − 1</xref>
        ] as fuzzification step. The single hidden layer
architecture uses a feature map of size 8, in the two hidden layers setup there are respectively
feature maps of size 8 and 16. Weights are randomly initialized using a uniform distribution
in range [
        <xref ref-type="bibr" rid="ref1">0 − 1</xref>
        ] in order to define a random degree order. Further weights are constrained
to be in the same range after the backpropagation phase, because they have to define at any
moment a data membership degree for all channels. There is a soft constraint that re-scales
weights in the correct range without a hard clipping of the weights on boundaries, as in the
gradient clipping case. All layers have a kernel size of 3 on spatial dimensions. In table 1 it
is possible to check performance of CNN compared with fuzzy architectures. It is possible to
observe that performances on MNIST are comparable but on CIFAR10, CNN outperform FRNN.
However, observing the activations and the heatmaps (Fig. 1) of the models and some others
visualizations based on GradCAM [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ], Gradients*Inputs [10] and Integrated Gradients [11]
(Fig. 2, 3) it is possible to note that FRNN can explain more accurately the information used in
classification. GradCAM shows that conv2d and relu function cut-out important features of
the object retaining the non relevant one, instead, FRNN, also if unrelevant area are present,
is able to preserves the shape of the ships. It is more clear also observing the gradients. Both
techniques have some noise, but the fuzzy module is able to focus more on the image subject.
The same analysis is possible observing MNIST fig. 3 where gradients, the attention of the
network, are located following the object shape. This statement is not valid for convolutional
layer that shows a lot of noise and is unable to focus on the main subject.
      </p>
      <p>Model
Conv2D+ReLU
MaxMin
MaxMin (2L)</p>
    </sec>
    <sec id="sec-4">
      <title>4. Conclusions</title>
      <p>In this work a fuzzy relational neural network based model for extrapolating relevant information
from the image data has been introduced. From the preliminary results we observed that the
model permits to obtain a clearer indication on the classification processes.</p>
      <p>In the next future the authors will focus on further validations of the model from both
theoretical and practical point of views.
explanations from deep networks via gradient-based localization, in: Proceedings of the
IEEE international conference on computer vision, 2017, pp. 618–626.
[10] D. Baehrens, T. Schroeter, S. Harmeling, M. Kawanabe, K. Hansen, K.-R. Müller, How to
explain individual classification decisions, The Journal of Machine Learning Research 11
(2010) 1803–1831.
[11] Z. Qi, S. Khorram, F. Li, Visualizing deep networks by optimizing with integrated gradients.,
in: CVPR Workshops, volume 2, 2019.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>A.</given-names>
            <surname>Knapič</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Malhi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Saluja</surname>
          </string-name>
          ,
          <string-name>
            <surname>K.</surname>
          </string-name>
          <article-title>Främling, xplainable artificial intelligence for human decision-support system in medical domain</article-title>
          ,
          <source>arXiv</source>
          (
          <year>2021</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>J. M.</given-names>
            <surname>Mendel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. P.</given-names>
            <surname>Bonissone</surname>
          </string-name>
          ,
          <article-title>Critical thinking about explainable ai (xai) for rule-based fuzzy systems</article-title>
          ,
          <source>IEEE Transactions on Fuzzy Systems</source>
          <volume>14</volume>
          (
          <year>2019</year>
          )
          <fpage>69</fpage>
          -
          <lpage>81</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>C.</given-names>
            <surname>Mencar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Alonso</surname>
          </string-name>
          ,
          <article-title>Paving the way to explainable artificial intelligence with fuzzy modeling</article-title>
          , volume
          <volume>24</volume>
          ,
          <year>2018</year>
          , pp.
          <fpage>215</fpage>
          -
          <lpage>227</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>F.</given-names>
            <surname>Camastra</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Ciaramella</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Giovannelli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Lener</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Rastelli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Staiano</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Staiano</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Starace</surname>
          </string-name>
          ,
          <article-title>A fuzzy decision system for genetically modified plant environmental risk assessment using mamdani inference</article-title>
          ,
          <source>Expert Systems with Applications</source>
          <volume>42</volume>
          (
          <year>2015</year>
          )
          <fpage>1710</fpage>
          -
          <lpage>1716</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>A.</given-names>
            <surname>Ciaramella</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Tagliaferri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Pedrycz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Di</surname>
          </string-name>
          <string-name>
            <surname>Nola</surname>
          </string-name>
          ,
          <article-title>Fuzzy relational neural network</article-title>
          ,
          <source>International Journal of Approximate Reasoning</source>
          <volume>41</volume>
          (
          <year>2006</year>
          )
          <fpage>146</fpage>
          -
          <lpage>163</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>A.</given-names>
            <surname>Ciaramella</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Tagliaferri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Pedrycz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Di</surname>
          </string-name>
          <string-name>
            <surname>Nola</surname>
          </string-name>
          ,
          <article-title>Fuzzy relational neural network</article-title>
          ,
          <source>International Journal of Approximate Reasoning</source>
          <volume>41</volume>
          (
          <year>2006</year>
          )
          <fpage>146</fpage>
          -
          <lpage>163</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>Y.</given-names>
            <surname>LeCun</surname>
          </string-name>
          , C. Cortes,
          <article-title>MNIST handwritten digit database (</article-title>
          <year>2010</year>
          ). URL: http://yann.lecun. com/exdb/mnist/.
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>A.</given-names>
            <surname>Krizhevsky</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Hinton</surname>
          </string-name>
          , et al.,
          <article-title>Learning multiple layers of features from tiny images (</article-title>
          <year>2009</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>R. R.</given-names>
            <surname>Selvaraju</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Cogswell</surname>
          </string-name>
          ,
          <string-name>
            <surname>A. Das</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Vedantam</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Parikh</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Batra</surname>
          </string-name>
          , Grad-cam: Visual
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>