=Paper= {{Paper |id=Vol-2853/paper17 |storemode=property |title=Method for Automatic Processing of Audit Content Based on Bidirectional Neural Network Mapping |pdfUrl=https://ceur-ws.org/Vol-2853/paper17.pdf |volume=Vol-2853 |authors=Tatiana Neskorodieva,Eugene Fedorov |dblpUrl=https://dblp.org/rec/conf/intelitsis/NeskorodievaF21 }} ==Method for Automatic Processing of Audit Content Based on Bidirectional Neural Network Mapping== https://ceur-ws.org/Vol-2853/paper17.pdf
Method for Automatic Processing of Audit Content Based on
Bidirectional Neural Network Mapping
Tatiana Neskorodievaa, Eugene Fedorova,b
a
    Vasyl' Stus Donetsk National University, 600-richchia str., 21, Vinnytsia, 21021, Ukraine
b
    Cherkasy State Technological University, Shevchenko blvd., 460, Cherkasy, 18006, Ukraine

                 Abstract
                 Currently, the analytical procedures used during the audit are based on data mining
                 techniques.
                 The object of the research is the process of the content auditing of the receipt of raw
                 materials for production and the manufactured products.
                 The aim of the work is to increase the effectiveness and efficiency of audit due to mapping
                 by full (bidirectional) counterpropagating neural network of content of the receipt of raw
                 materials for production and the manufactured products while automating procedures for
                 checking their compliance.
                 The vectors of feature for the objects of the sequences of the receipt of raw materials for
                 production and the manufactured products are generated, which are then used in the proposed
                 method. The created method, in contrast to the traditional one, provides for a batch mode,
                 which allows the method to increase the learning rate by an amount equal to the product of
                 the number of neurons in the hidden layer and the power of the training set, which is
                 critically important in the audit system for the implementation of multivariate intelligent
                 analysis, which involves enumerating various methods of forming subsets analysis.
                 The urgent task of increasing the audit efficiency was solved by automating the mapping of
                 audit indicators by full (bidirectional) counterpropagating neural network. A learning
                 algorithm based on π‘˜π‘˜-means has been created, intended for implementation on a GPU using
                 CUDA technology, which increases the speed of identifying parameters of a neural network
                 model.
                 The neural network with the proposed training method based on the π‘˜π‘˜-means rule can be used
                 to intellectualize the DSS audit. The prospects for further research are the application of the
                 proposed method by neural network mapping for a wide class of artificial intelligence tasks,
                 in particular, for creating a method for bidirectional mapping indicators of audit tasks.

                 Keywords 1
                 audit, mapping by neural network, full (bidirectional) counterpropagating neural network,
                 content of the receipt of raw materials for production and the manufactured products.

1. Introduction
    In the process of development of international and national economies and industry of IT in
particular, it is possible to distinguish the following basic tendencies: realization of digital
transformations, forming of digital economy, globalization of socio-economic processes and of IT
accompanying them [1]. These processes result in the origin of global, multilevel hierarchical
structures of heterogeneous, multivariable, multifunction connections, interactions and cooperation of
managing subjects (objects of audit), the large volumes of information about them have been
accumulated in the informative systems of account, management and audit.


IntelITSIS’2021: 2nd International Workshop on Intelligent Information Technologies and Systems of Information Security, March 24–26,
2021, Khmelnytskyi, Ukraine
EMAIL: t.neskorodieva@donnu.edu.ua (T. Neskorodieva); fedorovee75@ukr.net (E. Fedorov)
ORCID: 0000-0003-2474-7697 (T. Neskorodieva); 0000-0003-3841-7373 (E. Fedorov)
            Β© 2021 Copyright for this paper by its authors.
            Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0).
            CEUR Workshop Proceedings (CEUR-WS.org)
    Consequently, nowadays the scientific and technical issue of the modern information technologies
in financial and economic sphere of Ukraine is forming of the methodology of planning and creation
of the decision support systems (DSS) at the audit of enterprises in the conditions of application of IT
and with the use of information technologies on the basis of the automated analysis of the large
volumes of data about financial and economic activity and states of enterprises with the multi-level
hierarchical      structure     of    heterogeneous,     multivariable,     multifunction    connections,
intercommunications and cooperation of objects of audit with the purpose of expansion of functional
possibilities, increase of efficiency and universality of IT-audit.
    Currently, the analytical procedures used during the audit are based on data mining techniques [2-
4]. Automated DSS audit means the automatic forming of recommendable decisions, based on the
results of the automated analysis of data, that improves quality process of audit [5,6]. Unlike the
traditional approach, computer technologies of analysis of data in the system of audit accelerate and
promote the process accuracy of audit, that extremely critical in the conditions of plenty of associate
tasks on lower and middle levels, and also amounts of indexes and supervisions in every task [7,8].
    When developing a decision-making system in audit based on data mining technologies, three
methods have been created: classifying variables, forming analysis sets, mapping analysis sets.
    The peculiarity of the methodology for classifying indicators is that qualitatively different (by
semantic content) variables are classified: numerological, linguistic, quantitative, logical. The essence
of the second technique is determined by the qualitative meaning of the indicators. In accordance with
this, sets are formed with the corresponding semantic content: document numbers, the name of
indicators, quantitative estimates of the values of indicators, logical indicators.
    The third technique is subordinated to the mappings of formed sets of the same type on each other
in order to determine equivalence in the following senses: numerological, linguistic, quantitative,
logical.
    The most urgent problem is the mapping of quantitative indicators. The mapping of quantitative
indicators of the audit can be implemented through an ANN with associative memory. The main
ANNs with associative memory are presented in Table 1. The memory capacity was considered only
for ANNs with a binary or bipolar data type that perform reconstruction or classification. HAM stands
for hetero-associative memory, AAM stands for auto-associative memory.
    As follows from Table 1, most neural networks have one or more disadvantages:
    1. not used to reconstruct either the original or another sample;
    2. do not work with real data.
    3. do not have a high capacity of associative memory.
    The aim of the work is to increase the efficiency of automatic data analysis in the audit DSS by
means of a bidirectional (forward and reverse) neural network mapping of sets of audit indicators in
order to identify systematic misstatements that lead to misstatement of reporting. In the audit system,
the topical task of the middle level is the automation of the analysis of the conformity of the content
of the supply of raw materials for production and the manufactured products (by the periods of
quantization of the verification period).
    It is assumed that the audit indicators are noisy with Gaussian noise, which in turn simulates
random accounting errors (as opposed to systematic ones). It is also assumed that the residuals for the
quantization periods are distributed according to the normal law, the parameters of which can be
estimated from the accounting data. For the achievement of the aim it is necessary to solve the
following tasks:
    β€’     formalize the content of the audit process of the receipt of raw materials for production and
    the manufactured products;
    β€’     choose a neural network model for mapping audit indicators (which are noisy with Gaussian
    noise, which in turn simulates random accounting errors (as opposed to systematic ones, which
    lead to distortion of reporting));
    β€’     choose a criterion for evaluating the effectiveness of a neural network model;
    β€’     propose a method for training a neural network model in batch mode;
    β€’     propose an algorithm for training a neural network model in batch mode for implementation
    on a GPU;
    β€’     perform numerical studies.
Table 1
Basic ANNs with associative memory
                                          Memory        Memory          Data
                  ANN                                   capacity                             Purpose
                                           type                         type
  Forward-only                           HAM            -             Real         Reconstruction of other
  Counterpropagation Neural                                                        sample
  Network [9, 10]
  Full (bi-directional)                 AАМ,            -             Real         Reconstruction of the
  Counterpropagation Neural             HАМ                                        original or other sample
  Network [11]
  Sigmoid believe network [12,           AАМ            Medium        Binary       Reconstruction of the
  13]                                                                              original sample
  Helmholtz machine [14]                 AАМ            Medium        Binary       Reconstruction of the
                                                                                   original sample
  Self-Organizing map [15, 16]           AАМ            -             Real         Clustering

  Learning Vector Quantization           AАМ            -             Real         Clustering
  NN [17]
  Principal Component Analysis           HAM            -             Real         Dimension reduction
  NN [18]
  Independent Component                  HAM            -             Real         Dimension reduction
  Analysis NN [19, 20]
  Cerebellar Model Articulation          HAM            -             Real         Coding
  Controller [21]
  Recurrent correlation                  AАМ            High          Bipolar      Reconstruction of the
  associative memory [22]                                                          original sample
  Hopfield Neural Network [23]           AАМ            Low           Bipolar      Reconstruction of the
                                                                                   original sample
  Gauss machine [24]                     AАМ            Low           Bipolar      Reconstruction of the
                                                                                   original sample
  Bidirectional associative             AАМ,            Low           Bipolar      Reconstruction,
  memory [25]                           HAM                                        classification
  Brain State Model [26]                 AАМ            -             Real         Clustering
  Hamming neural network [27]            AАМ            High          Bipolar      Reconstruction of the
                                                                                   original sample
  Boltzmann machine [28,29,30]          AАМ,            Medium        Binary       Reconstruction of the
                                        HAM                                        original or other sample
  ART-2 [31]                             AАМ            -             Real          Clustering



2. MATERIALS AND METHODS
2.1. Formalize the content of the audit process of the receipt of raw
materials for production and the manufactured products
   Formalize the content of the supply of raw materials for production and the manufactured products
are formed on the basis of audit variables (Table 2). Elements of mapping sets – data of the receipt of
raw materials for production and the manufactured products by the periods of quantization of the
verification period. The vector of the receipt of raw materials features 𝒙𝒙𝑖𝑖 = (π‘₯π‘₯𝑖𝑖1 , . . . , π‘₯π‘₯𝑖𝑖𝑖𝑖 ) formed by
indicators of quantity of raw materials by type 𝑒𝑒, 𝑒𝑒 ∈ π‘ˆπ‘ˆ. The vector of manufactured products
features 𝑑𝑑𝑖𝑖 = (𝑑𝑑𝑖𝑖1 , . . . , 𝑑𝑑𝑖𝑖𝑖𝑖 ) is formed by indicators of the quantity of manufactured products by type
𝑝𝑝, 𝑝𝑝 ∈ 𝑃𝑃.

Table 2
Feature vectors for mapping raw material received - product produced
          Input vector elements 𝑋𝑋                         Output vector elements D
  designation               sense              designation                    sense
       𝑑𝑑         type of operation (receipt         π‘˜π‘˜         type of operation (production of
                     of raw materials for                       finished products (semi-finished
                         production)                                       products))
       𝑒𝑒           type of raw materials            𝑝𝑝                  type of product

         π‘ˆπ‘ˆ                  set of types of raw                           𝑃𝑃                   set of types of product
                                  materials
         𝑉𝑉𝑒𝑒           quantity of raw materials                       𝑉𝑉𝑝𝑝                     quantity of product


          (𝑐𝑐)
        π›₯π›₯𝑒𝑒                cost of raw material                           (𝑐𝑐)
                                                                       π›₯π›₯𝑒𝑒,𝑝𝑝          direct material costs for a product
                                                                                         type 𝑝𝑝 of a raw material type 𝑒𝑒
        π›₯π›₯(𝑐𝑐)          total cost of raw material                     π›₯π›₯𝑝𝑝
                                                                           (𝑐𝑐)         direct material costs for a product
                                                                                                      type 𝑝𝑝

   To assess the dimension of the features vector, an analysis was made of the nomenclature of
purchases of raw materials (components) of large machine-building enterprises. So, based on this
analysis, we can conclude that the sections of the nomenclature are on average from 8 to 12, the
number of groups in each section is from 2 to 10.
   We represent the implementation of the "generalized audit" in the form of a mapping (comparison)
of generalized quantitative features of the audited sets. The formation of generalized quantitative
features can be performed using ANN.

2.2.       Choosing a neural network model for mapping audit sets
   In the work, the Full (bidirectional) Counterpropagating Neural Network (FCPNN), which is a
non-recurrent static two-layer ANN, was chosen as a neural network. FCPNN output is linear.
   FCPNN advantages:
   1. unlike most ANNs are used to reconstruct another sample using auto-associative and hetero-
             associative memory.
   2. unlike bidirectional associative memory and the Boltzmann machine, it works with real data.
   3. unlike bidirectional associative memory and the Boltzmann machine, it has less
             computational complexity.
   FCPNN model performing mapping of each input sample 𝒙𝒙 = (π‘₯π‘₯1 , . . . , π‘₯π‘₯𝑁𝑁π‘₯π‘₯ ) to output sample π’˜π’˜
     (2)                (2)
=(𝑀𝑀𝑖𝑖 βˆ— 1 , . . . , 𝑀𝑀𝑖𝑖 βˆ— 𝑁𝑁𝑦𝑦 ), is represented as
                                                      π‘₯π‘₯             (1)
                 𝑖𝑖 βˆ— = π‘Žπ‘Žπ‘Žπ‘Žπ‘Žπ‘Ž π‘šπ‘šπ‘šπ‘šπ‘šπ‘š 𝑧𝑧𝑖𝑖 , 𝑧𝑧𝑖𝑖 = οΏ½βˆ‘π‘π‘                  2
                                                      π‘˜π‘˜=1(π‘₯π‘₯π‘˜π‘˜ βˆ’ π‘€π‘€π‘˜π‘˜π‘˜π‘˜ ) , 𝑖𝑖 ∈ 1, 𝑁𝑁
                                                                                        (1) ,                             (1)
                                𝑖𝑖
                (1)
   where π‘€π‘€π‘˜π‘˜π‘˜π‘˜ – connection weight from the π‘˜π‘˜-th element of the input sample to the 𝑖𝑖-th neuron,
      (2)
   𝑀𝑀𝑖𝑖 βˆ— 𝑗𝑗 – connection weight from the neuron-winner 𝑖𝑖 βˆ— to 𝑖𝑖-th element of output sample,
      (1)
   𝑁𝑁      – the number of neurons in the hidden layer.
    FCPNN model performing mapping of each output sample 𝒅𝒅 = (𝑑𝑑1 , . . . , 𝑑𝑑𝑁𝑁𝑑𝑑 ) to input sample
  (2)        (2)                (2)
𝒗𝒗𝑖𝑖 βˆ— = (𝑣𝑣𝑖𝑖 βˆ— 1 , . . . , π‘£π‘£π‘–π‘–βˆ— 𝑁𝑁π‘₯π‘₯ ), is represented as
                                                𝑑𝑑         (1)
    𝑖𝑖 βˆ— = π‘Žπ‘Žπ‘Žπ‘Žπ‘Žπ‘Ž π‘šπ‘šπ‘šπ‘šπ‘šπ‘š 𝑧𝑧𝑖𝑖 , 𝑧𝑧𝑖𝑖 = οΏ½βˆ‘π‘π‘                  2
                                         𝑠𝑠=1(𝑑𝑑𝑠𝑠 βˆ’ 𝑣𝑣𝑠𝑠𝑠𝑠 ) , 𝑖𝑖 ∈ 1, 𝑁𝑁
                                                                           (1) ,
                       𝑖𝑖
                (1)
   where 𝑣𝑣𝑠𝑠𝑠𝑠 – connection weight from the 𝑠𝑠-th element of the input sample to the 𝑖𝑖-th neuron of
hidden layer,
      (2)
   𝑣𝑣𝑖𝑖 βˆ— 𝑗𝑗 – connection weight from the neuron-winner 𝑖𝑖 βˆ— of hidden layer to 𝑗𝑗-th element of output
sample,
   𝑁𝑁 (1) – the number of neurons in the hidden layer.

2.3. Criterion choice for assessing the effectiveness of a neural network
model for mapping audit sets
   In this work for training model FCPNN was chosen target function, that indicates selection of the
                                                     (1)           (1)            (2)            (2)
vector      of      parameter      values   π‘Šπ‘Š = (𝑀𝑀11 , . . . , 𝑀𝑀𝑁𝑁π‘₯π‘₯ 𝑁𝑁(1) , 𝑀𝑀11 , . . . , 𝑀𝑀𝑁𝑁(1)𝑁𝑁𝑦𝑦 ), 𝑉𝑉 =
   (1)          (1)         (2)          (2)
(𝑣𝑣11 , . . . , 𝑣𝑣𝑁𝑁𝑑𝑑 𝑁𝑁(1) , 𝑣𝑣11 , . . . , 𝑣𝑣𝑁𝑁(1)𝑁𝑁π‘₯π‘₯ ) which deliver the minimum mean square error (difference
between the model sample and the test sample)
                                     1      1                  (2)     2         1       (2)          2
                              𝐹𝐹 = �𝑃𝑃𝑁𝑁𝑑𝑑 βˆ‘π‘ƒπ‘ƒπœ‡πœ‡=1 οΏ½π’˜π’˜πœ‡πœ‡π‘–π‘–βˆ— βˆ’ π’…π’…πœ‡πœ‡ οΏ½ + 𝑃𝑃𝑁𝑁π‘₯π‘₯ βˆ‘π‘ƒπ‘ƒπœ‡πœ‡=1 οΏ½π’—π’—πœ‡πœ‡π‘–π‘– βˆ— βˆ’ π’™π’™πœ‡πœ‡ οΏ½ οΏ½ β†’ π‘šπ‘šπ‘šπ‘šπ‘šπ‘š          (2)
                                     2                                                                        π‘Šπ‘Š,𝑉𝑉
            (2)
    where π’˜π’˜πœ‡πœ‡π‘–π‘– βˆ— – πœ‡πœ‡-th, output sample according to the model,
    π’…π’…πœ‡πœ‡ – πœ‡πœ‡-th test output sample,
      (2)
    π’—π’—πœ‡πœ‡π‘–π‘– βˆ— – πœ‡πœ‡-th, input sample according to the model,
    π’™π’™πœ‡πœ‡ – πœ‡πœ‡-th test input sample,
    𝑃𝑃 – training set power,
    𝑁𝑁 𝑑𝑑 – is length of the sample 𝒅𝒅,
    𝑁𝑁 π‘₯π‘₯ – is length of the sample 𝒙𝒙.

2.4.        Training method for neural network model in batch mode
     The disadvantage of FCPNN is that it does not have a batch learning mode, which leads to
reducing of the learning speed. For FCPNN was used concurrent training (combination of training
with and without a teacher). This work proposes training FCPNN in batch mode.
     First phase (training of the hidden layer) (steps 1-6).
                                                                                                       (1)
     The first phase allows you to calculate the weights of the hidden layer π‘€π‘€π‘˜π‘˜π‘˜π‘˜ and consists of the
following blocks (Fig. 1).
     1. Learning iteration number 𝑛𝑛 = 0, initialization by uniform distribution on the interval (0,1) or
                                         (1)         (1)
           [-0.5, 0.5] of weights 𝑀𝑀𝑖𝑖𝑖𝑖 (𝑛𝑛), 𝑣𝑣𝑠𝑠𝑠𝑠 (𝑛𝑛), 𝑖𝑖 ∈ 1, 𝑁𝑁 π‘₯π‘₯ , 𝑗𝑗 ∈ 1, 𝑁𝑁 (1), 𝑠𝑠 ∈ 1, 𝑁𝑁 𝑑𝑑 where 𝑁𝑁 π‘₯π‘₯ – is length
           of the sample π‘₯π‘₯ , 𝑁𝑁 𝑑𝑑 – is length of the sample 𝒅𝒅 and 𝑁𝑁 (1) – the number and the neurons in the
           hidden layer.
                                                 π‘₯π‘₯             𝑑𝑑
     Training set is {(π’™π’™πœ‡πœ‡ , π’…π’…πœ‡πœ‡ )|π’™π’™πœ‡πœ‡ ∈ 𝑅𝑅 𝑁𝑁 , π’…π’…πœ‡πœ‡ ∈ 𝑅𝑅 𝑁𝑁 }, πœ‡πœ‡ ∈ 1, 𝑃𝑃, where π’™π’™πœ‡πœ‡ – πœ‡πœ‡ -th training input vector,
π’…π’…πœ‡πœ‡ – πœ‡πœ‡ -th training output vector, 𝑃𝑃 – training set power.
     Initial shortest distance 𝑧𝑧̄ (0) =0.
     2. Calculating the distance to all hidden neurons.
     Distance π‘§π‘§πœ‡πœ‡πœ‡πœ‡ from Β΅-th input sample to each i-th neuron and from each Β΅-th output sample to each
i- th neuron of the hidden basis is determined by the formula:
               𝑁𝑁      π‘₯π‘₯             (1)      𝑁𝑁              𝑑𝑑          (1)
    π‘§π‘§πœ‡πœ‡πœ‡πœ‡ = οΏ½βˆ‘π‘˜π‘˜=1(π‘₯π‘₯πœ‡πœ‡πœ‡πœ‡ βˆ’ π‘€π‘€π‘˜π‘˜π‘˜π‘˜ (𝑛𝑛))2 + οΏ½βˆ‘π‘ π‘ =1(π‘‘π‘‘πœ‡πœ‡πœ‡πœ‡ βˆ’ 𝑣𝑣𝑠𝑠𝑠𝑠 (𝑛𝑛))2 , πœ‡πœ‡ ∈ 1, 𝑃𝑃, 𝑖𝑖 ∈ 1, 𝑁𝑁 (1) ,                  (3)
                 (1)
    where π‘€π‘€π‘˜π‘˜π‘˜π‘˜ (𝑛𝑛) – connection weight from k-th input sample to i-th neuron at time 𝑛𝑛,
     (1)
   𝑣𝑣𝑠𝑠𝑠𝑠 (𝑛𝑛) – pretrained connection weight from s-th element of output sample to i-th neuron of
hidden layer at time 𝑛𝑛.

                                1. Initializing the weights of the neurons of the hidden layer



                                2. Calculating the distance to all hidden neurons



                                3. Calculating the shortest distance and choosing the neuron
                                with the shortest distance


                                4. Setting the weights of the hidden layer neurons associated
                                with the neuron-winner and its neighbors



                                5. Calculating the average sum of the least distances




                                                                                                                yes
                                                   6. z (n + 1) βˆ’ z (n) > Ξ΅

                                                                           no
                                                                     7



   Figure 1. The sequence of steps in training method of FCPNN in batch mode (the first phase)

   3. Calculating the shortest distance and choosing the neuron with the shortest distance
   Calculating the shortest distance
                     π‘§π‘§πœ‡πœ‡ = π‘šπ‘šπ‘šπ‘šπ‘šπ‘šπ‘§π‘§πœ‡πœ‡πœ‡πœ‡ , πœ‡πœ‡ ∈ 1, 𝑃𝑃, 𝑖𝑖 ∈ 1, 𝑁𝑁 (1)                                                  (4)
                                    𝑖𝑖
   and choosing the neuron-winner π‘–π‘–πœ‡πœ‡βˆ— , for which the distance π‘§π‘§πœ‡πœ‡πœ‡πœ‡ is shortest
                         π‘–π‘–πœ‡πœ‡βˆ— = π‘Žπ‘Žπ‘Žπ‘Žπ‘Žπ‘Ž π‘šπ‘šπ‘šπ‘šπ‘šπ‘š π‘§π‘§πœ‡πœ‡πœ‡πœ‡ , πœ‡πœ‡ ∈ 1, 𝑃𝑃, 𝑖𝑖 ∈ 1, 𝑁𝑁 (1) .                                   (5)
                                          𝑖𝑖
   4. Setting the weights of the hidden layer neurons associated with the neuron-winner π‘–π‘–πœ‡πœ‡βˆ— and its
neighbors based on k-means rule
                           (1)                   βˆ‘π‘ƒπ‘ƒ          βˆ—
                                                  πœ‡πœ‡=1 β„Ž(𝑖𝑖,π‘–π‘–πœ‡πœ‡ )π‘₯π‘₯πœ‡πœ‡πœ‡πœ‡
                         π‘€π‘€π‘˜π‘˜π‘˜π‘˜ (𝑛𝑛 + 1) =          βˆ‘π‘ƒπ‘ƒ          βˆ—         , π‘˜π‘˜ ∈ 1, 𝑁𝑁 π‘₯π‘₯ , 𝑖𝑖 ∈ 1, 𝑁𝑁 (1) ,          (6)
                                                     πœ‡πœ‡=1 β„Ž(𝑖𝑖,π‘–π‘–πœ‡πœ‡ )
                          (1)                  βˆ‘π‘ƒπ‘ƒ          βˆ—
                                                πœ‡πœ‡=1 β„Ž(𝑖𝑖,π‘–π‘–πœ‡πœ‡ )π‘‘π‘‘πœ‡πœ‡πœ‡πœ‡
                        𝑣𝑣𝑠𝑠𝑠𝑠 (𝑛𝑛 + 1) =        βˆ‘π‘ƒπ‘ƒ          βˆ—          , 𝑠𝑠 ∈ 1, 𝑁𝑁 𝑑𝑑 , 𝑖𝑖 ∈ 1, 𝑁𝑁 (1) ,           (6’)
                                                  πœ‡πœ‡=1 β„Ž(𝑖𝑖,π‘–π‘–πœ‡πœ‡ )
   where β„Ž(𝑖𝑖, 𝑖𝑖 βˆ— ) – rectangular topological neighborhood function,
                    1, 𝑖𝑖 = 𝑖𝑖 βˆ—
   β„Ž(𝑖𝑖, 𝑖𝑖 βˆ— ) = οΏ½              .
                    0, 𝑖𝑖 β‰  𝑖𝑖 βˆ—
   5. Calculating the average sum of the shortest distances
                                                            1
                                              𝑧𝑧̄ (𝑛𝑛 + 1) = βˆ‘π‘ƒπ‘ƒπœ‡πœ‡=1 π‘§π‘§πœ‡πœ‡ .                                             (7)
                                                            𝑃𝑃
   6. Checking the termination condition
   If |𝑧𝑧̄ (𝑛𝑛 + 1) βˆ’ 𝑧𝑧̄ (𝑛𝑛)| ≀ πœ€πœ€, the finish, else 𝑛𝑛 = 𝑛𝑛 + 1, go to step 2.
   Second phase (training the output layer) (steps 7-12). The second phase allows you to calculate the
                                     (2)        (2)
weights of the output layer 𝑀𝑀𝑖𝑖𝑖𝑖 and 𝑣𝑣𝑖𝑖𝑖𝑖 and consists of the following blocks (Fig. 2).
   7. Learning iteration number 𝑛𝑛 = 0, initialization by uniform distribution on the interval (0,1) or [-
                               (2)        (2)
0.5, 0.5] of weights 𝑀𝑀𝑖𝑖𝑖𝑖 (𝑛𝑛), 𝑣𝑣𝑖𝑖𝑖𝑖 (𝑛𝑛), 𝑖𝑖 ∈ 1, 𝑁𝑁 (1) , 𝑗𝑗 ∈ 1, 𝑁𝑁 𝑑𝑑 , π‘žπ‘ž ∈ 1, 𝑁𝑁 π‘₯π‘₯ , where 𝑁𝑁 π‘₯π‘₯ – length of the
input sample 𝒙𝒙, 𝑁𝑁 𝑑𝑑 – length of the output sample 𝒅𝒅, 𝑁𝑁 (1) – the number and the neurons in the hidden
layer.
                                                π‘₯π‘₯             𝑦𝑦
    Training set is {(π’™π’™πœ‡πœ‡ , π’…π’…πœ‡πœ‡ )|π’™π’™πœ‡πœ‡ ∈ 𝑅𝑅 𝑁𝑁 , π’…π’…πœ‡πœ‡ ∈ 𝑅𝑅 𝑁𝑁 }, πœ‡πœ‡ ∈ 1, 𝑃𝑃, where π’™π’™πœ‡πœ‡ – πœ‡πœ‡-th training input vector, π’…π’…πœ‡πœ‡
– πœ‡πœ‡ -th training output vector, 𝑃𝑃 – training set power.
    Initial shortest distance 𝑧𝑧̄ (0) =0.

                                                               6


                    7. Initialization of the neuron weights of the output layer




                     8. Calculating the distance to all hidden neurons


                     9. Calculating the shortest distance and choosing the neuron
                     with the shortest distance


                      10. Calculating the distance to all output neurons


                      11. Setting the weights of the hidden layer neurons associated
                      with the neuron-winner and its neighbors


                     12. Calculating the average sum of the least distances z ( n + 1)



                                                                                                   yes
                                               13. z ( n + 1) βˆ’ z ( n) > Ξ΅

                                                                     no
                                              14. Output the weights


   Figure 2. Sequence of procedures for the FCPNN training method in batch mode (second phase)

   8. Calculating the distance to all hidden neurons
   Sum of distances π‘§π‘§πœ‡πœ‡πœ‡πœ‡ from Β΅-th input sample in input layer from each i-th neuron of the hidden
layer and from each Β΅-th output sample in the input layer to each i-th neuron of the hidden layer is
determined by the formula:
                             π‘₯π‘₯              (1)                𝑑𝑑               (1)
               π‘§π‘§πœ‡πœ‡πœ‡πœ‡ = οΏ½βˆ‘π‘π‘                    2    𝑁𝑁
                          π‘˜π‘˜=1(π‘₯π‘₯πœ‡πœ‡πœ‡πœ‡ βˆ’ π‘€π‘€π‘˜π‘˜π‘˜π‘˜ ) + οΏ½βˆ‘π‘ π‘ =1(π‘‘π‘‘πœ‡πœ‡πœ‡πœ‡ βˆ’ 𝑣𝑣𝑠𝑠𝑠𝑠 ) , πœ‡πœ‡ ∈ 1, 𝑃𝑃, 𝑖𝑖 ∈ 1, 𝑁𝑁
                                                                           2                         (1) ,               (8)
            (1)
   where π‘€π‘€π‘˜π‘˜π‘˜π‘˜ – pretrained connection weight from k-th element of input sample to i-th neuron at
time 𝑛𝑛,
     (1)
   𝑣𝑣𝑠𝑠𝑠𝑠 – pretrained connection weight from s- th element of input sample to i- th neuron of hidden
layer at time 𝑛𝑛.
   9. Calculating the shortest distance and choosing the neuron with the shortest distance.
   Calculating the shortest distance
                              π‘§π‘§πœ‡πœ‡ = π‘šπ‘šπ‘šπ‘šπ‘šπ‘šπ‘§π‘§πœ‡πœ‡πœ‡πœ‡ , πœ‡πœ‡ ∈ 1, 𝑃𝑃, 𝑖𝑖 ∈ 1, 𝑁𝑁 (1)                     (9)
                                              𝑖𝑖
   and choosing the neuron-winner π‘–π‘–πœ‡πœ‡βˆ— , for which the distance π‘§π‘§πœ‡πœ‡πœ‡πœ‡ is shortest.
                                  π‘–π‘–πœ‡πœ‡βˆ— = π‘Žπ‘Žπ‘Žπ‘Žπ‘Žπ‘Ž π‘šπ‘šπ‘šπ‘šπ‘šπ‘š π‘§π‘§πœ‡πœ‡πœ‡πœ‡ , πœ‡πœ‡ ∈ 1, 𝑃𝑃, 𝑖𝑖 ∈ 1, 𝑁𝑁 (1) .                          (10)
                                                   𝑖𝑖
   10. Calculating the distance to all output neurons.
   Distance π‘§π‘§πœ‡πœ‡ from the neuron-winner π‘–π‘–πœ‡πœ‡βˆ— to Β΅-th input and output sample in output layer is
determined by the formula:
                                   𝑑𝑑              (2)                           π‘₯π‘₯              (2)
                      π‘§π‘§πœ‡πœ‡ = οΏ½βˆ‘π‘π‘                           2    𝑁𝑁
                               𝑗𝑗=1(π‘‘π‘‘πœ‡πœ‡πœ‡πœ‡ βˆ’ 𝑀𝑀𝑖𝑖 βˆ— 𝑗𝑗 (𝑛𝑛)) + οΏ½βˆ‘π‘Ÿπ‘Ÿ=1(π‘₯π‘₯πœ‡πœ‡πœ‡πœ‡ βˆ’ 𝑣𝑣𝑖𝑖 βˆ— π‘Ÿπ‘Ÿ (𝑛𝑛)) , πœ‡πœ‡ ∈ 1, 𝑃𝑃,
                                                                                              2                     (11)
             (2)
   where π‘€π‘€π‘–π‘–πœ‡πœ‡βˆ— 𝑗𝑗 (𝑛𝑛) – weight of connection from the winner neuron π‘–π‘–πœ‡πœ‡βˆ— of hidden layer to j-th element of
the input sample in output layer at time 𝑛𝑛,
   11. Setting the weights of the output layer neurons associated with the neuron-winner π‘–π‘–πœ‡πœ‡βˆ— and its
neighbors based on k-means rule
                                  (2)                βˆ‘π‘ƒπ‘ƒ          βˆ—
                                                      πœ‡πœ‡=1 β„Ž(𝑖𝑖,π‘–π‘–πœ‡πœ‡ )π‘‘π‘‘πœ‡πœ‡πœ‡πœ‡
                                𝑀𝑀𝑖𝑖𝑖𝑖 (𝑛𝑛 + 1) =        βˆ‘π‘ƒπ‘ƒ          βˆ—        , 𝑖𝑖 ∈ 1, 𝑁𝑁 (1) , 𝑗𝑗 ∈ 1, 𝑁𝑁 𝑑𝑑 ,   (12)
                                                          πœ‡πœ‡=1 β„Ž(𝑖𝑖,π‘–π‘–πœ‡πœ‡ )
                                 (2)               βˆ‘π‘ƒπ‘ƒ          βˆ—
                                                    πœ‡πœ‡=1 β„Ž(𝑖𝑖,π‘–π‘–πœ‡πœ‡ )π‘₯π‘₯πœ‡πœ‡πœ‡πœ‡
                               𝑣𝑣𝑖𝑖𝑖𝑖 (𝑛𝑛 + 1) =     βˆ‘π‘ƒπ‘ƒ          βˆ—          , 𝑖𝑖 ∈ 1, 𝑁𝑁 (1) , 𝑠𝑠 ∈ 1, 𝑁𝑁 π‘₯π‘₯ ,     (12’)
                                                      πœ‡πœ‡=1 β„Ž(𝑖𝑖,π‘–π‘–πœ‡πœ‡ )
   where β„Ž(𝑖𝑖, 𝑖𝑖 βˆ— ) – rectangular topological neighborhood function,
                    1, 𝑖𝑖 = 𝑖𝑖 βˆ—
   β„Ž(𝑖𝑖, 𝑖𝑖 βˆ— ) = οΏ½              .
                    0, 𝑖𝑖 β‰  𝑖𝑖 βˆ—
   12. Calculating the average sum of the shortest distances
                                                  1
                                   𝑧𝑧̄ (𝑛𝑛 + 1) = βˆ‘π‘ƒπ‘ƒπœ‡πœ‡=1 π‘§π‘§πœ‡πœ‡ .                                                    (13)
                                                  𝑃𝑃
   13. Checking the termination condition
   If |𝑧𝑧̄ (𝑛𝑛 + 1) βˆ’ 𝑧𝑧̄ (𝑛𝑛)| ≀ πœ€πœ€, the finish, else 𝑛𝑛 = 𝑛𝑛 + 1, go to step 8.

2.5. Algorithm for training neuron network model in batch mode for
implementation on GPU
          For the proposed method of training FCPNN on audit data example, examines the algorithm for
implementation on a GPU with usage of CUDA parallel processing technology.
          The first phase (training the hidden layer). The first phase based on formulas (1)-(7) is shown in
Fig. 3. This block diagram functions as follows.
          Step 1 – Operator enters lengths 𝑁𝑁 π‘₯π‘₯ of the sample 𝒙𝒙, the lengths 𝑁𝑁 𝑑𝑑 of the sample 𝒅𝒅, the
number and the neurons in the hidden layer 𝑁𝑁 (1) , power of the training set 𝑃𝑃, training set
                            π‘₯π‘₯              𝑑𝑑
{(π’™π’™πœ‡πœ‡ , π’…π’…πœ‡πœ‡ )|π’™π’™πœ‡πœ‡ ∈ 𝑅𝑅 𝑁𝑁 , π’…π’…πœ‡πœ‡ ∈ 𝑅𝑅 𝑁𝑁 }, πœ‡πœ‡ ∈ 1, 𝑃𝑃.
          Step 2 – Initialization by uniform distribution over the interval (0,1) or [-0.5, 0.5] of weights
     (1)         (1)
𝑀𝑀𝑖𝑖𝑖𝑖 (𝑛𝑛), 𝑣𝑣𝑠𝑠𝑠𝑠 (𝑛𝑛), 𝑖𝑖 ∈ 1, 𝑁𝑁 π‘₯π‘₯ , 𝑠𝑠 ∈ 1, 𝑁𝑁 𝑑𝑑 , 𝑗𝑗 ∈ 1, 𝑁𝑁 (1) .
          Step 3 – Calculation of distances to all hidden neurons of the ANN, using 𝑃𝑃 β‹… 𝑁𝑁 (1) threads on
GPU, which are grouped into P blocks. Each thread calculates the distance from Β΅-th input sample to
each i-th neuron π‘§π‘§πœ‡πœ‡πœ‡πœ‡ .
      Step 4 – Computation based on shortest distance reduction and determining the neurons with the
shortest distance using 𝑃𝑃 β‹… 𝑁𝑁 (1) threads on GPU, which are grouped into P blocks. The result of the
work of each block is a neuron-winner π‘–π‘–πœ‡πœ‡βˆ— with the smallest distance π‘§π‘§πœ‡πœ‡ .
      Step 5 – Setting the weights of the output layer neurons associated with the neuron- winner π‘–π‘–πœ‡πœ‡βˆ— and
its neighbors based on reduction using 𝑁𝑁 π‘₯π‘₯ β‹… 𝑁𝑁 (1) β‹… 𝑃𝑃 threads on GPU, which are grouped into 𝑁𝑁 π‘₯π‘₯ β‹…
                                                                            (1)
𝑁𝑁 (1) blocks. The result of the work of each block is the weight π‘€π‘€π‘˜π‘˜π‘˜π‘˜ (𝑛𝑛 + 1).
      Step 6 – Setting the weights of the output layer neurons associated with the neuron- winner π‘–π‘–πœ‡πœ‡βˆ— and
its neighbors based on reduction using 𝑁𝑁 𝑑𝑑 β‹… 𝑁𝑁 (1) β‹… 𝑃𝑃 threads on GPU, which are grouped into 𝑁𝑁 𝑑𝑑 β‹…
                                                                           (1)
𝑁𝑁 (1) blocks. The result of the work of each block is the weight 𝑣𝑣𝑠𝑠𝑠𝑠 (𝑛𝑛 + 1).
      Step 7 – Calculation based on reduction of the average sum of the shortest distances using 𝑃𝑃
threads on GPU, which are grouped into 1 block. The result of the block is the average sum of the
smallest distances 𝑧𝑧̄ (𝑛𝑛 + 1).
      Step 8 – If average sum of smallest distances of neighboring iterations are close, |𝑧𝑧̄ (𝑛𝑛 + 1) βˆ’
𝑧𝑧̄ (𝑛𝑛)| ≀ πœ€πœ€, then finish, else – increasing number of iteration 𝑛𝑛 = 𝑛𝑛 + 1, go to step 3.
                                       1


                                       2


                                       3


                                       4


                                       5


                                       6


                                       7


                                       8
                                  +                  9


   Figure 3. Block diagram of the FCPNN learning algorithm in batch mode (first phase)

                                      (1)             (1)
      Step 9 – Recording of weight π‘€π‘€π‘˜π‘˜π‘˜π‘˜ (𝑛𝑛 + 1) and 𝑣𝑣𝑠𝑠𝑠𝑠 (𝑛𝑛 + 1) in the database.
      Second phase (training of output layer). The second phase based on formulas (8)-(13) is showed in
Figure. 4. This flowchart operates as follows.
      Step 1 – Operator enters lengths 𝑁𝑁 π‘₯π‘₯ of the sample 𝒙𝒙, the lengths 𝑁𝑁 𝑑𝑑 of the sample 𝒅𝒅 the number
and the neurons in the hidden layer 𝑁𝑁 (1) , power of the training set 𝑃𝑃, training set
                            π‘₯π‘₯             𝑑𝑑
οΏ½οΏ½π’™π’™πœ‡πœ‡ , π’…π’…πœ‡πœ‡ οΏ½οΏ½π’™π’™πœ‡πœ‡ ∈ 𝑅𝑅 𝑁𝑁 , π’…π’…πœ‡πœ‡ ∈ 𝑅𝑅 𝑁𝑁 οΏ½ , πœ‡πœ‡ ∈ 1, 𝑃𝑃.
      Step 2 – Initialization by uniform distribution over the interval (0,1) or [-0.5, 0.5] of weights
     (2)         (2)
𝑀𝑀𝑖𝑖𝑖𝑖 (𝑛𝑛), 𝑣𝑣𝑖𝑖𝑖𝑖 (𝑛𝑛), 𝑖𝑖 ∈ 1, 𝑁𝑁 (1) , 𝑗𝑗 ∈ 1, 𝑁𝑁 𝑑𝑑 , π‘žπ‘ž ∈ 1, 𝑁𝑁 π‘₯π‘₯ .
      Step 3 – Calculation of distances to all hidden neurons of the ANN, using 𝑃𝑃 β‹… 𝑁𝑁 (1) threads on
GPU, which are grouped into P blocks. Each thread calculates the distance from Β΅-th input sample to
each 𝑖𝑖-th neuron π‘§π‘§πœ‡πœ‡πœ‡πœ‡ .
      Step 4 – Computation based on shortest distance reduction and determining the neurons with the
shortest distance using 𝑃𝑃 β‹… 𝑁𝑁 (1) threads on GPU, which are grouped into 𝑃𝑃 blocks. The result of the
work of each block is a neuron-winner π‘–π‘–πœ‡πœ‡βˆ— with the smallest distance π‘§π‘§πœ‡πœ‡ .
      Step 5 – Calculating distances from the neuron-winner π‘–π‘–πœ‡πœ‡βˆ— to Β΅-th output sample using 𝑃𝑃 β‹… 𝑁𝑁 (𝑑𝑑)
threads on GPU, which are grouped into 𝑃𝑃 blocks. Each thread calculates the distance from the
neuron-winner π‘–π‘–πœ‡πœ‡βˆ— to Β΅-th output sample π‘§π‘§πœ‡πœ‡ .
      Step 6 – Setting the weights of the output layer neurons associated with the neuron-winner π‘–π‘–πœ‡πœ‡βˆ— and
its neighbors based on reduction using 𝑁𝑁 (1) β‹… 𝑁𝑁 𝑑𝑑 β‹… 𝑃𝑃 threads on GPU, which are grouped into 𝑁𝑁 (1) β‹…
                                                                               (2)
𝑁𝑁 𝑑𝑑 blocks. The result of the work of each block is the weight 𝑀𝑀𝑖𝑖𝑖𝑖 (𝑛𝑛 + 1).
      Step 7 – Setting the weights of the output layer neurons associated with the neuron-winner π‘–π‘–πœ‡πœ‡βˆ— and
its neighbors based on reduction using 𝑁𝑁 (1) β‹… 𝑁𝑁 π‘₯π‘₯ β‹… 𝑃𝑃 threads on GPU, which are grouped into 𝑁𝑁 (1) β‹…
                                                                                        (2)
𝑁𝑁 π‘₯π‘₯ block. The result of the block is the average sum of the smallest distances 𝑣𝑣𝑖𝑖𝑖𝑖 (𝑛𝑛 + 1).
      Step 8 – Calculation based on reduction of the average sum of the shortest distances using 𝑃𝑃
threads on GPU, which are grouped into 1 block. The result of the block is the average sum of the
smallest distances 𝑧𝑧̄ (𝑛𝑛 + 1).
      Step 9 – If average sum of smallest distances of neighboring iterations are close, |𝑧𝑧̄ (𝑛𝑛 + 1) βˆ’
𝑧𝑧̄ (𝑛𝑛)| ≀ πœ€πœ€, then finish, else – increasing number of iteration 𝑛𝑛 = 𝑛𝑛 + 1, go to step 3.
                                                   (2)                     (2)
      Step 10 – Recording of weight 𝑀𝑀𝑖𝑖𝑖𝑖 (𝑛𝑛 + 1) and 𝑣𝑣𝑖𝑖𝑖𝑖 (𝑛𝑛 + 1), in the database.
                                      1


                                      2


                                      3


                                      4


                                      5


                                      6


                                      7


                                      8


                                      9
                                 +                        10


   Figure 4. Block diagram of the FCPNN training algorithm in batch mode (second phase)

2.6.    Numerical research
    The results of the comparison of the proposed method using GPU and the traditional FCPNN
training method are presented in Table 3.
Table 3
Comparison of the computational complexity of the proposed and traditional training methods of
FCPNN
                                                                   Method
    Feature
                                     proposed                                    traditional
 Computational
                             𝑂𝑂(𝑛𝑛1π‘šπ‘šπ‘šπ‘šπ‘šπ‘š + 𝑛𝑛2π‘šπ‘šπ‘šπ‘šπ‘šπ‘š )        𝑂𝑂(𝑃𝑃𝑁𝑁(1) 𝑛𝑛1π‘šπ‘šπ‘šπ‘šπ‘šπ‘š + (𝑃𝑃𝑁𝑁(1) + 𝑃𝑃)𝑛𝑛2π‘šπ‘šπ‘šπ‘šπ‘šπ‘š )
 complexity

    Evaluation of computational complexity of the proposed method using the GPU, and the
traditional method of teaching FOCPNN were based on the number of calculation distances,
computing of which is the most consuming part of method. Moreover, 𝑛𝑛1π‘šπ‘šπ‘šπ‘šπ‘šπ‘š – the maximum number
of iterations of the first training phase, 𝑛𝑛2π‘šπ‘šπ‘šπ‘šπ‘šπ‘š – the maximum number of iterations of the second
training phase, 𝑁𝑁 (1) – the number of neurons in the hidden layer, 𝑃𝑃 – the power of the training set.

2.7.    Discussion
   The traditional FCPNN learning method does not provide support for batch mode, which increases
computational complexity (Table 3). Proposed method eliminates this flaw and allows for
approximate increase of learning rate in 𝑃𝑃𝑁𝑁 (1) . By reducing the computational complexity, it is
possible to increase the accuracy of the method by decreasing the parameter Ξ΅ and increasing the
number of neurons in the hidden and output layers.

2.8.    Conclusion
   1. The urgent task of increasing the effectiveness of audit in the context of large volumes of
analyzed data and limited verification time was solved by automating the formation of generalized
features of audit sets and their mapping by means of a full bidirectional counterpropagating neural
network.
    2. For increased learning rate of full bidirectional counterpropagating neural network, was
developed a method based on the π‘˜π‘˜-means rule for training in batch mode. The proposed method
provides: approximately increase learning rate in 𝑃𝑃𝑁𝑁 (1), where 𝑁𝑁 (1) is the number of neurons in the
hidden layer and 𝑃𝑃 is the power of the learning set.
    3. Created a learning algorithm based on π‘˜π‘˜-means, intended for implementation on a GPU using
CUDA technology.
    4. The proposed method of training based on the π‘˜π‘˜-means rule can be used to intellectualize the
DSS audit.
    Prospects for further research is the study of the proposed method for a wide class of artificial
intelligence tasks, as well as the creation of a method for bidirectional mapping audit features to solve
audit problems.

3. References
[1] The World Bank: World Development Report 2016: Digital Dividends, 2016. URL:
     https://www.worldbank.org/en/publication/wdr2016.
[2] M. Schultz, and M. Tropmann-Frick. Autoencoder Neural Networks versus External Auditors:
     Detecting Unusual Journal Entries in Financial Statement Audits. Proceedings of the 53rd
     Hawaii International Conference on System Sciences (HICSS-2020), Maui, Hawaii, USA. 2021,
     pp. 5421-5430. doi:10.24251/hicss.2020.666.
[3] J. Nonnenmacher, F. Kruse, G. Schumann, G. Marx. Using Autoencoders for Data-Driven
     Analysis in Internal Auditing. In Proceedings of the 54th Hawaii International Conference on
     System Sciences, Maui, Hawaii, USA, 2021, pp. 5748-5757. doi: 10.24251/hicss.2021.697.
[4] Y. Bodyanskiy, O. Boiko, Y. Zaychenko, G. Hamidov and A. Zelikman. The Hybrid GMDH-
     Neo-fuzzy Neural Network in Forecasting Problems in Financial Sphere. Proceedings of 2nd
     International Conference on System Analysis & Intelligent Computing (SAIC), Kyiv, Ukraine,
     IEEE, 2020, pp. 1-6, doi: 10.1109/SAIC51296.2020.9239152.
[5] T. NeskorodΡ–eva, E. Fedorov, I. Izonin. Forecast Method for Audit Data Analysis by Modified
     Liquid State Machine. Proceedings of the 1st International Workshop on Intelligent Information
     Technologies & Systems of Information Security (IntelITSIS 2020), Khmelnytskyi, Ukraine, 10-
     12 June, 2020: proceedings. – CEUR-WS vol. 2623, 2020, pp. 25-35.
[6] T. NeskorodΡ–eva, E. Fedorov. Method for Automatic Analysis of Compliance of Expenses Data
     and the Enterprise Income by Neural Network Model of Forecast. Proceedings of the 2nd
     International Workshop on Modern Machine Learning Technologies and Data Science
     (MoMLeT&DS-2020), Lviv-Shatsk, Ukraine, 2-3 June, 2020: proceedings. – CEUR-WS,
     Volume I: Main Conference. vol. 2631, 2020, pp. 145-158.
[7] A.V. Barmak, Y.V. Krak, E.A. Manziuk, V.S. Kasianiuk. Information technology separating
     hyperplanes synthesis for linear classifiers. Journal of Automation and Information Sciences, vol.
     51(5) (2019) 54-64. doi: 10.1615/JAutomatInfScien.v51.i5.50
[8] T.V. Prokopenko, O. Grigor. Development of the comprehensive method to manage risks in
     projects related to information technologies. Eastern-European Journal of Enterprise
     Technologies vol. 2, 2018, pp. 37-43. doi:10.15587/1729-4061.2018.128140.
[9] U.P. Singh, S. Jain, A. Tiwari, R.K. Singh. Gradient evolution-based counter propagation
     network for approximation of noncanonical system. Soft Computing, 23 13, (2019) 4955–4967.
     doi:10.1007/s00500-018-3160-7.
[10] A. Asokan, K. Gunavathi, R. Anitha. Classification of Melakartha ragas using neural networks.
     Proceedings of the International Conference on Innovations in Information, Embedded and
     Communication Systems (ICIIECS-2017), 17-18 March 2017, Coimbatore, India. IEEE. 2017,
     pp.1-6. doi:10.1109/ICIIECS.2017.8276040.
[11] Sonika; A. Pratap; M. Chauhan; A. Dixit. New technique for detecting fraudulent transactions
     using hybrid network consisting of full-counter propagation network and probabilistic network.
     2016 International Conference on Computing, Communication and Automation (ICCCA), 29-30
     April 2016, Greater Noida, India, IEEE. 2016, pp. 29-30. doi:10.1109/CCAA.2016.7813713.
[12] P.M. Baggenstoss. Applications of Projected Belief Networks (PBN). 27th European Signal
     Processing Conference (EUSIPCO). 2-6 Sept. 2019, A Coruna, Spain, IEEE. 2019 pp. 1-5
     doi:10.23919/EUSIPCO.2019.8902708.
[13] P. M. Baggenstoss, On the Duality Between Belief Networks and Feed-Forward Neural
     Networks, in IEEE Transactions on Neural Networks and Learning Systems, vol. 30 (1). 2019,
     pp. 190-200, doi: 10.1109/TNNLS.2018.2836662.
[14] P. Sountsov, P. Miller. Spiking neuron network Helmholtz machine. Frontiers in Computational
     Neuroscience. Frontiers Media SA. vol. 9, 2015. doi:10.3389/fncom.2015.00046.
[15] T Kohonen Self-organizing and associative memory, 3rd edn. Springer, New York, (2012).
[16] T Kohonen. Essentials of the self-organizing map. Neural Networks Part of special issue:
     Twenty-fifth      Anniversay       Commemorative       Issue.     vol.    37,   2013,      52-65.
     doi:10.1016/j.neunet.2012.09.018.
[17] H. de Vries, R. Memisevic, A. Courville. Deep Learning Vector Quantization. ESANN 2016
     proceedings, European Symposium on Artificial Neural Networks, Computational Intelligence
     and Machine Learning. Bruges (Belgium), April 2016 (2016) 27-29. URL:
     http://www.i6doc.com/en/
[18] J.-W. Cho, H.-M. Park, Independent vector analysis followed by hmm-based feature
     enhancement for robust speech recognition, Sig. Process. 120 (2016) 200–208.
     doi:10.1016/j.sigpro.2015.09.002.
[19] R. Chai, G.R. Naik, T.N. Nguyen, S.H. Ling, Y. Tran, A. Craig, H.T. Nguyen, Driver fatigue
     classification with independent component by entropy rate bound minimization analysis in an
     eeg-based system, IEEE J. Biomed. Health Inf. IEEE vol. 21 (3), 2017, pp. 715–724.
     doi:10.1109/JBHI.2016.2532354.
[20] A. Tharwat, Principal component analysis-a tutorial, Int. J. Appl. Pattern Recognit. Inderscience
     Publishers; vol. 3 (3), 2016, pp. 197–240. doi:10.1504/ijapr.2016.10000630.
[21] M. I. Achmad, H. Adinugroho and A. Susanto, "Cerebellar model articulation controller
     (CMAC) for sequential images coding," 2014 The 1st International Conference on Information
     Technology, Computer, and Electrical Engineering, Semarang, Indonesia, 2014, pp. 160-166,
     doi: 10.1109/ICITACEE.2014.7065734.
[22] R.A. Lobo, M.E. Valle. Ensemble of Binary Classifiers Combined Using Recurrent Correlation
     Associative Memories. In: Cerri R., Prati R.C. (eds) Intelligent Systems (BRACIS 2020). Lecture
     Notes in Computer Science, vol 12320. Springer, Cham. doi:10.1007/978-3-030-61380-8_30.
[23] Kobayashi M. Quaternionic Hopfield neural networks with twin-multistate activation function.
     Neurocomputing 267, 2017, pp. 304–310. doi:10.1016/j.neucom.2017.06.013.
[24] Du K.L., Swamy M.N.S. Neural Networks and Statistical Learning. Springer-Verlag London,
     (2014). doi: 10.1007/978-1-4471-5571-3.
[25] E. Javidmanesh. Global Stability and Bifurcation in Delayed Bidirectional Associative Memory
     Neural Networks With an Arbitrary Number of Neurons. Journal of Dynamic Systems,
     Measurement, and Control. ASME International. 139(8) (2017). doi:10.1115/1.4036229.
[26] Y. Park. Optimal and robust design of brain-state-in-a-box neural associative memories. Neural
     Networks. Elsevier BV. 23(2), 2010, pp. 210–8. doi: 10.1016/j.neunet.2009.10.008
[27] O.I. Khristodulo, A.A. Makhmutova, T.V. Sazonova Use algorithm Based at Hamming Neural
     Network Method for Natural Objects Classification. Procedia Computer Science. Elsevier BV.
     103, 2017, pp. 388–395. doi: 10.1016/j.procs.2017.01.126.
[28] A. Fischer, C. Igel. Training restricted Boltzmann machines: An introduction. Pattern
     Recognition. Elsevier BV, 47(1) (2014) 25–39. doi:10.1016/j.patcog.2013.05.025.
[29] Q. Wang, X. Gao, K. Wan, F. Li, Z. Hu. A Novel Restricted Boltzmann Machine Training
     Algorithm with Fast Gibbs Sampling Policy. Mathematical Problems in Engineering. Hindawi
     Limited (2020) 1–19. doi:10.1155/2020/4206457T.
[30] A. Fischer, C. Igel. Training Restricted Boltzmann Machines: An Introduction Pattern
     Recognition, 47 (2014) 25-39. doi.10.1016/j.patcog.2013.05.025.
[31] Barszcz, A. Bielecki, M. WΓ³jcik, M. Bielecka. ART-2 Artificial Neural Networks Applications
     for Classification of Vibration Signals and Operational States of Wind Turbines for Intelligent
     Monitoring. Advances in Condition Monitoring of Machinery in Non-Stationary Operations.
     Springer Berlin Heidelberg. 2013, pp. 679–88. doi:10.1155/2020/4206457.