<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Image Recognition by a Second-Order Convolutional Neural Network with Dynamic Receptive Fields</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Oksana Mezentseva</string-name>
          <email>omezentceva@ncfu.ru</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Department of Information Systems &amp; Technologies, North-Caucasus Federal University.</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Adaptation of a Convolution</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>2, Kulakov Prospect</institution>
          ,
          <addr-line>Stavropol, Russian Federation</addr-line>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Maksim Brodnikov</institution>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Second-Order Neurons</institution>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2017</year>
      </pub-date>
      <abstract>
        <p>This paper presents the method description of convolutional neural networks parameters synthesis with dynamic receptive fields and secondorder neurons for convolutional layers. The experiments of pattern recognition which shown that the combination of second-order neurons and dynamic receptive fields allows to reduce the generalization error are described in the article.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Introduction</title>
      <p>with</p>
    </sec>
    <sec id="sec-2">
      <title>Dynamic Fields towards</title>
      <p>CNN idea with dynamic RFs is that if one changes the set of RFs for some layers then the same pattern can be
perceived in different ways by the network. Hence, one can extend a training set. It is known that the classical
form of a RF for a CNN is a square. It is offered to use a template for non-standard RF generation. The elements
of a template are indices indicating neighbors within two discrete steps from the element on a pixel array. If you
change all RFs lying on a feature map, additional information will impact on the configurable parameters. It
will lead to the extract of a better invariant (Fig. 1).</p>
      <p>High-order CNNs are expansion over signal direct distribution networks where polynomial neurons summators
are used. That type of neurons enables a network to extract information from the input signal more qualitatively
[6]. At the same time high-order functions (square, cube, trigonometric, etc.) are used instead of the usual
weighted summation. The main drawback of this approach is the need to introduce additional coefficient arrays,
but it is offset by higher quality features extracting (hence the generalization error reduces).</p>
      <p>The activation function of the standard neuron with two inputs is calculated according to the formula 1, the
activation function of the second-order neuron is calculated according to the formula 2.</p>
      <p>2
f (w1 ∗ x1 + w2 ∗ x2) = f (X wi ∗ xi)</p>
      <p>i=1
f (w1 ∗ x1 + w2 ∗ x2 + u1 ∗ x1 ∗ x1 + u2 ∗ x2 ∗ x2 + u3 ∗ x1 ∗ x1) = f (X xi ∗ wi + uj ∗ xj )
2
i,j
where w, u are vectors of adjustable coefficients, and x is an input vector, f(·) is the activation function of a
neuron.</p>
      <p>This approach does not touch activation functions of neurons and the principles of error calculation, so the
only difference in contrast to the standard back-propagation algorithm is the calculation of the weighted sum.</p>
      <p>We need to calculate two partial derivatives according to the formula 3 with the aim to know how much we
want to change the weights w and u.</p>
      <p>( ∂∂wvjji((nn)) = (Pim=0 wji(n) ∗ yi(n) + uji(n) ∗ yi2(n))0|wji(n) = Pim=0 yi(n)
∂∂uvjji((nn)) = (Pim=0 wji(n) ∗ yi(n) + uji(n) ∗ yi2(n))0|uji(n) = Pim=0 yi2(n)
(1)
(2)
(3)
where vj(n) is a weighted sum of the neuron j at iteration n, yi(n) is an input values for the neuron j.</p>
      <p>The algorithm [7] which changes RFs for neurons belonging to any combination of convolutional layers
(Clayers) before feeding a pattern was developed. The synthesis method of mathematical model parameters of CNN
with dynamic RFs and second-order neurons for C-layers is proposed. This method was developed on the basis
of the change algorithm of RFs and adaptation of the forward and back-propagation to neurons of second-order
(2, 3). This method adapts the back-propagation algorithm, it consists of the following steps:
1. In the process of training before entering the next pattern, we need to change neurons RFs at the desired
combination of C-layers using the algorithm of RFs form changing.</p>
      <p>2. During the forward propagation we need to calculate the output values of C-layer neurons according to the
http://ceur-ws.org
formula 4.</p>
      <p>(Cmi,n = ϕ(b + Pq∈Qi PkK=C0−1 PlK=C0−1 ([Xmq+k+Fi(·),n+l+Fj(·) ∗ Wkq,l] + [Xmq+k+Fi(·),n+l+Fj(·) ∗ Aqk,l]))
Smi,n = ϕ(b + u PKS−1 PKS−1 Cmi∗KS+k,n∗KS+l)</p>
      <p>k=0 l=0
where Cim,n is the neuron output disposed at i-th map of C-layer in position m, n, ϕ(·) = A ∗ tanh(B ∗ p)
when A = 1.7159, B = 2/3, b is a bias, Qi is indices set of the previous layer cards associated with Ci
card, KC is the size of square RF for the neuron Cim,n, Xqm+k,n+l is an input value to the neuron Cim,n, the
vectors W and A are custom weights for the second-order neurons of C-layer, Sim,n is the neuron output of
averaging layer; Fi(·) and Fj(·) are the Fi(RFm,n, k, l), Fj(RFm,n, k, l) (Functions which return the offsets
for the row and the column for RF-template belonging to neuron m, n at position k, l within the template).
indexk,l is a template element of at position k, l, indexk,l = 0..24.</p>
      <p>These functions are determined by the following formulas:
0; indexk,l ∈ {0, 4, 5, 16, 17}

1; indexk,l ∈ {6, 7, 8, 18, 19}

Fi(·) = 2; indexk,l ∈ {20, 21, 22, 23, 24}

−1; indexk,l ∈ {1, 2, 3, 14, 15}


−2; indexk,l ∈ {9, 10, 11, 12, 13}
0; indexk,l ∈ {0, 2, 7, 11, 22}

1; indexk,l ∈ {3, 5, 8, 12, 23}

Fj(·) = 2; indexk,l ∈ {13, 15, 17, 19, 24}

−1; indexk,l ∈ {1, 4, 6, 10, 21}


−2; indexk,l ∈ {9, 14, 16, 18, 20}
(4)
(6)
,
(5)
3. During the back propagation we need to receive the local gradient for C-layer according to the formula 6.
 ∂E
 ∂(Wkλ,l)q = PSizeC PSizeC δmλ,n ∗ ymλ−+1k+Fi(RFmλ,n,k,l),n+l+Fj(RFmλ,n,k,l)</p>
      <p>m=0 n=0
 ∂(Akλ,l)q = PSizeC PSizeC δm,n ∗ ymλ−+1k+Fi(RFmλ,n,k,l),n+l+Fj(RFmλ,n,k,l)
∂E λ</p>
      <p>m=0 n=0
where δmi,n is a residual was gathered for a neuron with coordinates m, n within the map of layer λ, q is
the part of the kernel of configurable features for which we receive the components of the gradient, SizeC
is the size of the C-layer card, Aqk,l, Wk,l is q-th part of the adjustable parameters, which is responsible for
q
interaction with the q-th card of the previous layer.</p>
      <p>CNN learning algorithm was developed on the basis of the proposed adaptation method of back propagation
algorithm to second-order neurons and dynamic RFs (Fig. 2).
3</p>
    </sec>
    <sec id="sec-3">
      <title>Experiments</title>
      <p>Experiments were carried out to evaluate the generalization capability of a CNN with second-order neurons and
dynamic RFs.</p>
      <p>Stereo dataset Small NORB [8] was used as a sample database for the invariant pattern recognition. The use
of classical CNN of LeNet-5 type gives the generalization error of 8.4% [9].</p>
      <p>
        The use of CNN with dynamic RFs and the general structure of LeNet-5 gives the generalization error of 4.3%
[
        <xref ref-type="bibr" rid="ref5">10</xref>
        ]. Table 1 shows the results of LeNet-5 with dynamic RFs and second-order neurons in various combinations
of the C-layer.
      </p>
      <p>Where Ci, Si in the Table 1 are the numbers of corresponding convolution and averaging layers, I is the input
layer, O is the output layer, number 2 means the second-order neurons in the corresponding layer. The value of
the generalization error was calculated as the average of 15 experiments.</p>
      <p>The table shows that not all combinations of the second-order neurons used in the C-layers provide a significant
reduction of the generalization error. The best results show the embodiment number 6 and number 9. The best</p>
      <p>C1</p>
      <p>S1</p>
      <p>C2</p>
      <p>S2</p>
      <p>C3
2
combination can only be found through exhaustive search of possible values. It limits the use of higher order
CNN.</p>
      <p>As a result of the proposed CNN with dynamic RFs and second-order neurons we were able to reduce the
generalization error by an average of 2% in comparison with the use of the CNN with simply dynamic RFs and
by 6.1% in comparison with the CNN with alternating C and S - layers. It is shown that the use of the artificial
neural network with the higher-order neurons leads to the decrease of the generalization error. But as a result
we have to carry out additional experiments in order to find the necessary combination of C-layers with the
modified neural activation functions.
[6] Lagunov, N.: Allocation and recognition of objects with the help of optimized algorithm of selective search
and a high-order convolutional neural network // Fundamental research. 2015. 5. pp. 511-516.
Nemkov, R.: The method of a mathematical model parameters synthesis for a convolution neural
network with an expanded training set. The modern problems science and education, 2015. 1. URL:
http://www.science-education.ru/125-19867/;
(NORB).</p>
      <p>URL:
Nemkov, R.: Dynamical Change of the Perceiving Properties of Neural Networks as Training with Noise
and Its Impact on Pattern Recognition / R. Nemkov // Young Scientists International Workshop on Trends
in Information Processing (YSIP) 2014. URL: http://ceur-ws.org/Vol-1145/paper4.pdf/</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          <string-name>
            <surname>Goodfellow</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bengio</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          :
          <article-title>Deep Learning (Adaptive Computation and Machine Learning series)</article-title>
          .
          <source>Adaptive Computation and Machine Learning series</source>
          .
          <year>2016</year>
          . P.
          <volume>800</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          <string-name>
            <surname>Mirza</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Courville</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bengio</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          :
          <source>Generalizable Features From Unsupervised Learning</source>
          ,
          <source>2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)</source>
          , pages
          <fpage>4945</fpage>
          -
          <lpage>4949</lpage>
          ,
          <year>2016</year>
          Nemkov,
          <string-name>
            <given-names>R.</given-names>
            ,
            <surname>Mezentseva</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            ,
            <surname>Mezintsev</surname>
          </string-name>
          ,
          <string-name>
            <surname>D.</surname>
          </string-name>
          :
          <article-title>Using of a Convolutional Neural Network with Changing Receptive Fields in the Tasks of Image Recognition /</article-title>
          <source>Proceedings of the First International Scientific Conference Intelligent Information Technologies for Industry (IITI16)</source>
          ,
          <source>Volume 451 of the series Advances in Intelligent Systems and Computing</source>
          ,
          <year>2016</year>
          , pp.
          <fpage>15</fpage>
          -
          <lpage>23</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          <source>The 4th International Scientific Conference: Applied Natural Sciences. - Novy Smokovec</source>
          ,
          <year>2013</year>
          . - pp.
          <fpage>284</fpage>
          -
          <lpage>289</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          <string-name>
            <surname>Nemkov</surname>
            ,
            <given-names>R. M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mezentseva</surname>
            <given-names>O. S.:</given-names>
          </string-name>
          <article-title>Dynamical change of the perceiving properties of convolutional neural networks and its impact on generalization</article-title>
          .
          <source>Neurocomputers: development and application</source>
          ,
          <year>2015</year>
          , no.
          <issue>2</issue>
          , pp.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [10]
          <string-name>
            <surname>Nemkov</surname>
            ,
            <given-names>R. M.</given-names>
          </string-name>
          :
          <article-title>Synthesis method of mathematical model parameters of the convolutional neural network with extended training set</article-title>
          . URL: http://www.science-education.
          <source>ru/125-19867/ (30.01</source>
          .
          <year>2016</year>
          ).
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>