<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Training feedforward neural networks using hybrid particle swarm optimization, Multi-Verse Optimization</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>1st Rabab Bousmaha</string-name>
          <email>Rabab.bousmaha@gmail.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>2nd Reda Mohamed Hamou</string-name>
          <email>hamoureda@yahoo.fr</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>3rd Amine Abdelmalek</string-name>
          <email>amineabd1@yahoo.fr</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>GeCoDe Laboratory, Department of Computer Science, University of Saida</institution>
          ,
          <addr-line>Saida</addr-line>
          ,
          <country country="DZ">Algeria</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>-The learning process of artificial neural networks is an important and complex task in the supervised learning field. The main difficulty of training a neural network is the process of fine-tuning the best set of control parameters in terms of weight and bias. This paper presents a new training method based on hybrid particle swarm optimization with Multi-Verse Optimization (PMVO) to train the feedforward neural networks. The hybrid algorithm is utilized to search better in solution space which proves its efficiency in reducing the problems of trapping in local minima. The performance of the proposed approach was compared with five evolutionary techniques and the standard momentum backpropagation and adaptive learning rate. The comparison was benchmarked and evaluated using six bio-medical datasets. The results of the comparative study show that PMVO outperformed other training methods in most datasets and can be an alternative to other training methods. Index Terms-Particle swarm optimization, Multi-Verse Optimization, Training feedforward neural networks, Real world Datasets</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>I. INTRODUCTION</title>
      <p>
        Artificial neural network (ANN) is one of the most
important data mining techniques. It has been successfully applied
to many fields. The feedforward multilayer perceptron (MLP)
is one of the best-known neural networks. The multilayer
perceptron (MLP) consists of three layers composed of neurons
organized into input, output and hidden layers. The success
of an MLP generally depends on the training process that is
determined by training algorithms. The objective of the
training algorithms is to find the best connection between weights
and biases that minimize the classification error. Training
algorithms can be classified into two classes: gradient-based
and stochastic search methods. Backpropagation (BP) and its
variants are gradient-based methods and considered as one of
the most popular techniques used to train the MLP neural
network. Gradient-based methods have many drawbacks, such
as the slow convergence, the high dependency on the initial
value of weights and biases and the tendency to be trapped
in local minima [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ].To address these problems, stochastic
search methods, such as metaheuristics have been proposed as
alternative methods for training feedforward neural network.
Metaheuristics have many advantages: they apply to any type
of ANN with any activation function [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ], are particularly useful
for dealing with large complex problems that generate many
local optima [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. Genetic algorithm (GA) and Particle
Swarm Optimization (PSO) considered as the most
wellknown nature inspired MLP trainers. Montana and Davis
proposed one of the earliest works on training the feedforward
neural network (FFNN) with GA [22]. They showed that GA
outperform BP when solving real problems.Slowik and Bialko
[23] employed Differential Evolution (DE) for training MLP
and showed that it has promising performance compared to
BP and Levenberg-Marquardt methods.
      </p>
      <p>
        Others metaheuristics algorithms have been applied in
training feedforward MLP, such as the modified BAT [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ],
MultiVerse Optimization MVO [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ], Whale Optimization Algorithm
(WOA) [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ], Grey Wolf Optimizer (GWO) [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ],
Biogeography Based on Optimizer (BBO) [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ], Moth-Flame
Optimization (MFO) [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] and Improved Monarch Butterfly
Optimization (IMBO) [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ]. Furthermore, several hybrid algorithms have
been proposed to train a neural network. Tarkhaneh and Shen
[
        <xref ref-type="bibr" rid="ref13">13</xref>
        ] suggested a hybrid approach to neural network training
by combining PSO, Mantegna Levy flight and neighbor search
(LPSONS). The comparison experiments showed that the
proposed algorithm can find optimal results. Khan et al [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ]
introduced a new method based on two algorithms, accelerated
particle swarm optimization (APSO) and cuckoo search (CS),
named HACPSO. The comparison results demonstrated that
the proposed algorithm outperforms other algorithms in term
of accuracy, MSE and standard deviation. This paper presents
a new training approach based on hybrid particle swarm
optimization (PSO) with Multi-Verse Optimization (MVO), called
PMVO, to train the feedforward neural network (FFNN). Six
datasets were solved by the proposed trainer. Moreover, the
application of the trainer was investigated in bio-medical. The
performance of PMVO was compared with five well-known
trainer metaheuristics algorithms in the literature: PSO [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ],
MFO [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ], MVO [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ], WOA [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ], HACPSO [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ].
      </p>
      <p>
        II. ARTIFICIAL NEURAL NETWORKS (ANNS)
An artificial neural network (ANN) is a computational
model based on the structure and functions of the biological
brain and nervous system. The feedforward neural network
(FFNN) is one of the most popular types of artificial neural
network [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ].FFNN has three interconnected layers. The first
layer consists of input neurons. These neurons send the data
to the second layer, called the hidden layer, which sends the
output neurons to the third layer. In FFNN, the information
travels in one direction, from the input layer to the output
layer. The node or the artificial neuron multiplies each of these
inputs by weight, as shown in (1):
      </p>
      <p>
        Sj =
n
X wi;j Ii + j
i=1
where, n is the total number of neuron inputs, W ij is the
connection weight connecting Ij to neuron j and j is a bias
weight [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. Then, the node or the artificial neuron adds the
multiplications and sends the sum to a transfer function, for
example, Sigmoid function presented in (2):
f (x) =
      </p>
      <p>1
1 + e x
n
yj = fj (X wi;j Ii + j )</p>
      <p>
        i=1
The output of the neuron j can be described as follows (3):
After building the neural network, the set of network weights
are adjusted to approximate the desired results. This process
is carried out by applying a training algorithm to adapt the
weights until error criteria are met [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ].
      </p>
      <p>III. PARTICLE SWARM OPTIMIZATION (PSO)
in 1995 Russell Eberhart and James Kennedy have invented
the particle swarm optimization which is a population-based
stochastic optimization technique inspired by birds flocking
around food sources. like each other evolutionary
computational algorithms. In PSO, each individual is a bird in the
search space. We call it a particle. All of the particles have
fitness values which are evaluated by the fitness function to
be optimized and flies in the space with a velocity which is
dynamically adjusted according to its own flying experience
[16].</p>
      <p>
        Vit;j+1 = Vit;j W + C1R1(P bestt Xt) + C2R2(Gbestt Xt)
(4)
Xt+1 = Xt + V t+1 i = 1; 2:::N P )And(j = 1; 2:::N G)
(5)
where P bestt and Gbestt denote the best particle position
and best group position and w is inertia weight w =
(1)
(2)
(3)
wmax ( (wmax wmin) iteration ),C1; C2 two positive constants
maxiteration
, R1,R2 are random numbers in the interval of [
        <xref ref-type="bibr" rid="ref1">0, 1</xref>
        ], Vit;j+1 is
the velocity of jth member of ith particle at iteration number
(t) and (t+1). The new position values Xt+1 are obtained by
adding the velocity updates determined by the formula given
in Equation (5).
      </p>
    </sec>
    <sec id="sec-2">
      <title>IV. MULTI-VERSE OPTIMIZATION(MVO)</title>
      <p>Multiverse optimization proposed by Syed Ali mirjalili in
2015 [17] As Inspired by the concepts of white holes, black
holes, and wormholes in the multi-verse theory and big bang
theory. In this algorithm, the models of these three concepts are
developed to perform exploration and exploitation and local
search. The fitness function for each search agent is indicated
by the inflation rate, and each object and each universe in the
search agent represent a candidate solution and a variable in
the candidate solution.</p>
      <p>In this algorithm, the larger universes tend to send objectives
to smaller universes. A large universe is defined based on
inflation rate in the multi-verse theory. The following rules
are applied to the universes of the MVO:</p>
      <p>If the inflation rate rate is higher, the probability of having
a white hole is higher.</p>
      <p>If the inflation rate rate is higher, the probability of having
black holes is lower.</p>
      <p>Universes having higher inflation rate rate send the
objects through white holes.</p>
      <p>Universes having lower inflation rate rate tend to receive
more objects through black holes.</p>
      <p>The objects of all universes may be replaced by the
objects of the universe with the greater inflation rate.
The mathematical model of this algorithm is as follows:
8 xj + T DR + ((ubj
xij = &lt; xj T DR + ((ubj</p>
      <p>
        :
(6)
lbj) r4 + lbj);if r3 &lt; 00::55 ;if r2 &lt; W EP
lbj) r4 + lbj);if r3
xij; if r2 W EP
Where xj indicates the jth variable in the bests universe,
lbi indicates the lower bound in jth variable, ubi shows the
upper bound in jth variable, r2 ,r3 , r4 are random numbers
j
in the interval of [
        <xref ref-type="bibr" rid="ref1">0, 1</xref>
        ], T DP=W EP are coefficients, and xi
indicates the jth parameter in ith universe.
      </p>
    </sec>
    <sec id="sec-3">
      <title>V. HYBRID PSO-MVO:</title>
      <p>Hybrid PSO-MVO is sequential combination of PSO and
MVO. The algorithm merges the best strength of both PSO
in exploitation and MVO in exploration towards the optimum
solution when the universe value of MVO replace the Pbest
value of PSO [20] [21]. In this paper we propose a novel
training algorithm based on this algorithm for the first time in
the following section.The equation can be written as follows:
Vit;j+1 = Vit;j W + C1R1(U niversest
Xt) + C2R2(Gbestt
(7)</p>
      <p>Xt)</p>
    </sec>
    <sec id="sec-4">
      <title>Step 1: Initialize the PSO values</title>
      <p>Step 2: Evaluate the fitness function of each particle
Step 3: Determine Gbest from the Pbest value
Step 4: updated velocity and position values of each particule
Step 5: verify the solution whether it is feasible or not
Step 6: steps 2 to 5 were repeated until the maximum number
of iterations was reached.</p>
      <p>Step 9: Use the optimal solutions of PSO as boundary to
MVO algorithm
Step 10: Initialize the MVO values
Step 11: Evaluate the inflation rate of the universe (fitness
function)
Step 12: Update the position of the universes
Step 13: if the convergence criterion is reached; get the
results
Step 14: if the convergence criterion is not reached; continue
the process from step 11-14</p>
    </sec>
    <sec id="sec-5">
      <title>VI. PMVO FOR TRAINING MLP</title>
      <p>
        This section presents the proposed approach based on the
PMVO to train the MLP network named PMVO.Two
important points are taken into consideration: the fitness function
and the representation of the PMVO solutions. In this work,
the PMVO algorithm was applied for training MLP network
with a single hidden layer and each PMVO solution (weights
and biases) was formed by three parts: the connection weights
between the input layer and the hidden layer, the weights
between the hidden layer and the output layer, and the bias
weights.The length of each solution vector is given by equation
(8), where n is the number of input features and m is the
number of neurons in the hidden layer [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ].
      </p>
      <p>IndividualLength = (n
m) + (2
m) + 1
(8)</p>
      <p>
        PMVO solutions are implemented as real number vectors
when each vector belongs to the interval [
        <xref ref-type="bibr" rid="ref1">-1, 1</xref>
        ]. The mean
square error (MSE) was used to measure the fitness value of
PMVO solutions.MSE was calculated based on the difference
between the estimated and actual values of the neural network
using the training datasets, as shown in equation (9), where n
is the number of samples in the training dataset y and y^ are
respectively the actual and predicted values:
n
M SE = 1 X(y
n
i=1
y^)
2
(9)
      </p>
    </sec>
    <sec id="sec-6">
      <title>VII. EXPERIMENTS AND RESULTS</title>
      <p>This section presents the evaluation of the proposed PMVO
for training MLP networks on six well-known datasets, which
were selected from University Of California Irvine machine
learning (UCI)1 and Kaggle2 dataset repositories. Table I
shows the classification of these datasets in terms of features
number, classes, training and testing samples. The comparison</p>
    </sec>
    <sec id="sec-7">
      <title>1http://archive.ics.uci.edu/ml/</title>
      <p>
        2https://www.kaggle.com/datasets
of PMVO was carried out with five approaches used to train
feedforward neural network in the literature: PSO [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ], MFO
[
        <xref ref-type="bibr" rid="ref11">11</xref>
        ], MVO [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ], WOA [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ], HACPSO [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ] In addition, the
proposed algorithm was compared with standard momentum
Back-Propagation and adaptive learning rate and (BP), which
are gradient-based algorithms.
      </p>
    </sec>
    <sec id="sec-8">
      <title>VIII. EXPERIMENTAL SETUP</title>
      <p>The proposed trainer and other algorithms were
implemented with Python language and a personal computer with
Intel(R) Core(TM) CPU 1.60 GHz 2.30 GHz, 64 Bits Windows
7 operating system and 4 GB (RAM).</p>
      <p>
        The metaheuristics are sensitive to the value of their
parameters, which requires a careful initialization. Therefore, the
control parameters recommended in the literature were used
[
        <xref ref-type="bibr" rid="ref15">15</xref>
        ] [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] and summarized in Table II. All datasets were
divided into 66% for training and 34% for testing. Moreover,
all features were mapped to the interval [
        <xref ref-type="bibr" rid="ref1">0, 1</xref>
        ] to eliminate
the effect of features that have different scales. Min-max
normalization is applied to perform a linear transformation
on the original data, were v0 is the normalized value of v in
the range [minA; maxA] as given in (10).
      </p>
      <p>v0 =</p>
      <p>
        vi
maxA
minA
minA
(10)
In the literature, there is no standard method for selecting the
number of hidden neurons. In this work, the method proposed
in [18] [19] [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] was used; the number of neurons in the hidden
layer equals to 2N + 1 , where N , is a number of features in
the dataset.
      </p>
    </sec>
    <sec id="sec-9">
      <title>IX. RESULTS</title>
      <p>All algorithms were tested ten times for every dataset and
the population size and the maximum number of generations
were set to 50 and 200, respectively.</p>
      <p>Table IV shows the statistical results: average, best, worst
and standard deviation of classification accuracy. The results
of PMVO outperformed other approaches in breast cancer,
blood, liver, vertebral with an average accuracy of 0.962,
0.766, 0.752, 0.839. In addition, PMVO was ranked second
in diabetes and Parkinson datasets with an average accuracy
of 0.783, 0.842 respectively. Moreover, it can also be seen that
the PMVO has a smaller Std which indicates that PMVO is
stable. Table V shows the average, best and worst MSE with
standard deviation, obtained for each algorithm. As a result,
it can be noted that PMVO outperforms other techniques in
four datasets: breast cancer, blood, liver and vertebral with an
average MSE of 0.032, 0.168, 0.176, 0.131, respectively. In
addition, it can be also noticed that PMVO has small standard
deviation value for all datasets which proves the efficiency and
robustness of this algorithm.</p>
      <p>Figures 1 2 3 4 5 and 6 show the convergence curves of
all metaheuristics training algorithms based on the average
values of MSE.The convergence curves show that PMVO has
the lowest value of MSE for four datasets: breast cancer,
blood, liver and vertebral. Moreover, PMVO has the fastest
convergence speed in liver, vertebral, blood and European
datasets. For diabetes dataset, PMVO provides a very close
performance compared to MVO algorithm. These results show
that PMVO has a faster convergence and a better optimization
than other metaheuristic algorithms.</p>
      <p>TableIII shows the average ranks obtained by each
optimization technique in the Friedman test. The comparative shows
that the proposed algorithm outperforms other algorithms.</p>
      <p>In this paper, we have proposed a new training approach
based on Particle swarm optimization, Multi-Verse
Optimization to train the feedforward neural network. The training
method took into account the capabilities of the PMVO in
terms of high exploration and exploitation to locate the optimal
values for weights and biases of FFNN. The approach is
proposed to minimize the training error and to increase the
accuracy. The approach is benchmarked and evaluated using
six standard bio-medical datasets.</p>
      <p>The comparison between the proposed algorithm and PSO,
MFO, MVO, WOA, HACPSO and standard BP with
momentum term and adaptive learning rate shows the superiority of
the PMVO algorithm with high accuracy and small MSE in
most of the datasets compared with other training algorithms.
Moreover, the small value of standard deviation shows that
our trainer is robust and stable. Finally, from the experiment,
we can conclude that PMVO can give good results and can
be an alternative to other training methods.</p>
      <p>In future works, we focus on how to extend this work to
solve a more real world problem and we test the performance
of PMVO to train other types of neural networks.
[16] J. Kennedy, ; R.Eberhart, Particle Swarm Optimization, Proceedings of
IEEE International Conference on Neural Networks. ICNN.1995.488968
IV. pp. 19421948 , 1995.
[17] S. Mirjalili, S. M. Mirjalili, and A. Hatamlou, Multi-Verse Optimizer:
a nature-inspired algorithm for global optimization, Neural Computing
and Applications, vol. 27, no. 2, pp. 495513, 2015.
[18] S. Mirjalili, S. M. Mirjalili, and A. Lewis, Let a biogeography-based
optimizer train your Multi-Layer Perceptron, Information Sciences, vol.
269, pp. 188209, 2014.
[19] Mirjalili, S. (2015). S. Mirjalili, How effective is the Grey Wolf
optimizer in training multi-layer perceptrons, Applied Intelligence, vol.
43, no. 1, pp. 150161, 2015.
[20] Sagarika, T.R.Jyothsna, Tunning of PSO algorithm for single machine
and multi machine power system using STATCOM controller,
international journal of engineering and technology, vol 2, issue 4, pp.175-182,
2015.
[21] K.Karthikeyan, P.K.Dhal, Transient stability enhancement by optimal
location and tuning of STATCOM using PSO, Procedia technology,
2015.
[22] DJ.Montana , L.Davis Training feedforward neural networks using
genetic algorithms. In: Proceedings of the 11th International Joint
Conference on Artificial Intelligence - Volume 1, Morgan Kaufmann
Publishers Inc., San Francisco, CA, USA, IJCAI89, pp 762767, 1989.
[23] A.Slowik, M.Bialko Training of artificial neural networks using
differential evolution algorithm. In: Conference on Human System
Interactions,IEEE, pp 6065,2008.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>J.-R.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , J. Zhang, T.
          <string-name>
            <surname>-M. Lok</surname>
            , and
            <given-names>M. R.</given-names>
          </string-name>
          <string-name>
            <surname>Lyu</surname>
          </string-name>
          ,
          <article-title>A hybrid particle swarm optimizationback-propagation algorithm for feedforward neural network training</article-title>
          ,
          <source>Applied Mathematics and Computation</source>
          , vol.
          <volume>185</volume>
          , no.
          <issue>2</issue>
          , pp.
          <fpage>10261037</fpage>
          ,
          <year>2007</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>S.</given-names>
            <surname>Kiranyaz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Ince</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Yildirim</surname>
          </string-name>
          , and
          <string-name>
            <given-names>M.</given-names>
            <surname>Gabbouj</surname>
          </string-name>
          ,
          <article-title>Evolutionary artificial neural networks by multi-dimensional particle swarm optimization</article-title>
          ,
          <source>Neural Networks</source>
          , vol.
          <volume>22</volume>
          , no.
          <issue>10</issue>
          , pp.
          <fpage>14481462</fpage>
          ,
          <year>2009</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>T.</given-names>
            <surname>Kenter</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Borisov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C. V.</given-names>
            <surname>Gysel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Dehghani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. D.</given-names>
            <surname>Rijke</surname>
          </string-name>
          , and
          <string-name>
            <given-names>B.</given-names>
            <surname>Mitra</surname>
          </string-name>
          ,
          <article-title>Neural Networks for Information Retrieval</article-title>
          ,
          <source>Proceedings of the Eleventh ACM International Conference on Web Search and Data Mining - WSDM 18</source>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>L.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Huang</surname>
          </string-name>
          , and
          <string-name>
            <given-names>S.</given-names>
            <surname>Lazebnik</surname>
          </string-name>
          ,
          <article-title>Learning Two-Branch Neural Networks for Image-Text Matching Tasks</article-title>
          ,
          <source>IEEE Transactions on Pattern Analysis and Machine Intelligence</source>
          , vol.
          <volume>41</volume>
          , no.
          <issue>2</issue>
          , pp.
          <fpage>394407</fpage>
          ,
          <string-name>
            <surname>Jan</surname>
          </string-name>
          .
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>N. S.</given-names>
            <surname>Jaddi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Abdullah</surname>
          </string-name>
          ,
          <article-title>and</article-title>
          <string-name>
            <given-names>A. R.</given-names>
            <surname>Hamdan</surname>
          </string-name>
          ,
          <article-title>Optimization of neural network model using modified bat-inspired algorithm</article-title>
          ,
          <source>Applied Soft Computing</source>
          , vol.
          <volume>37</volume>
          , pp.
          <fpage>7186</fpage>
          ,
          <year>2015</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>H.</given-names>
            <surname>Faris</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Aljarah</surname>
          </string-name>
          , and
          <string-name>
            <given-names>S.</given-names>
            <surname>Mirjalili</surname>
          </string-name>
          ,
          <article-title>Training feedforward neural networks using multi-verse optimizer for binary classification problems</article-title>
          ,
          <source>Applied Intelligence</source>
          , vol.
          <volume>45</volume>
          , no.
          <issue>2</issue>
          , pp.
          <fpage>322332</fpage>
          , May
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>I.</given-names>
            <surname>Aljarah</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Faris</surname>
          </string-name>
          , and
          <string-name>
            <given-names>S.</given-names>
            <surname>Mirjalili</surname>
          </string-name>
          ,
          <article-title>Optimizing connection weights in neural networks using the whale optimization algorithm</article-title>
          ,
          <source>Soft Computing</source>
          , vol.
          <volume>22</volume>
          , no.
          <issue>1</issue>
          , pp.
          <fpage>115</fpage>
          ,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>M. F.</given-names>
            <surname>Hassanin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. M.</given-names>
            <surname>Shoeb</surname>
          </string-name>
          ,
          <article-title>and</article-title>
          <string-name>
            <given-names>A. E.</given-names>
            <surname>Hassanien</surname>
          </string-name>
          ,
          <article-title>Grey wolf optimizerbased back-propagation neural network algorithm,</article-title>
          <year>2016</year>
          12th International Computer Engineering Conference (ICENCO),
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>H.</given-names>
            <surname>Faris</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Mirjalili</surname>
          </string-name>
          ,
          <string-name>
            <surname>and I. Aljarah</surname>
          </string-name>
          ,
          <article-title>Automatic selection of hidden neurons and weights in neural networks using grey wolf optimizer based on a hybrid encoding scheme</article-title>
          ,
          <source>International Journal of Machine Learning and Cybernetics</source>
          , vol.
          <volume>10</volume>
          , no.
          <issue>10</issue>
          , pp.
          <fpage>29012920</fpage>
          ,
          <string-name>
            <surname>Jun</surname>
          </string-name>
          .
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>I.</given-names>
            <surname>Aljarah</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Faris</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Mirjalili</surname>
          </string-name>
          , and
          <string-name>
            <given-names>N.</given-names>
            <surname>Al-Madi</surname>
          </string-name>
          ,
          <article-title>Training radial basis function networks using biogeography-based optimizer</article-title>
          ,
          <source>Neural Computing and Applications</source>
          , vol.
          <volume>29</volume>
          , no.
          <issue>7</issue>
          , pp.
          <fpage>529553</fpage>
          ,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>H.</given-names>
            <surname>Faris</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Aljarah</surname>
          </string-name>
          , and
          <string-name>
            <given-names>S.</given-names>
            <surname>Mirjalili</surname>
          </string-name>
          ,
          <string-name>
            <surname>Evolving Radial Basis Function Networks Using MothFlame Optimizer</surname>
          </string-name>
          ,
          <source>Handbook of Neural Computation</source>
          , pp.
          <fpage>537550</fpage>
          ,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>H.</given-names>
            <surname>Faris</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Aljarah</surname>
          </string-name>
          , and
          <string-name>
            <given-names>S.</given-names>
            <surname>Mirjalili</surname>
          </string-name>
          ,
          <article-title>Improved monarch butterfly optimization for unconstrained global search and neural network training</article-title>
          ,
          <source>Applied Intelligence</source>
          , vol.
          <volume>48</volume>
          , no.
          <issue>2</issue>
          , pp.
          <fpage>445464</fpage>
          ,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>O.</given-names>
            <surname>Tarkhaneh</surname>
          </string-name>
          and
          <string-name>
            <given-names>H.</given-names>
            <surname>Shen</surname>
          </string-name>
          ,
          <article-title>Training of feedforward neural networks for data classification using hybrid particle swarm optimization, Mantegna Lvy flight and neighborhood search</article-title>
          ,
          <source>Heliyon</source>
          , vol.
          <volume>5</volume>
          , no.
          <issue>4</issue>
          ,
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>A.</given-names>
            <surname>Khan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Shah</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Imran</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Khan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. I.</given-names>
            <surname>Bangash</surname>
          </string-name>
          , and
          <string-name>
            <given-names>K.</given-names>
            <surname>Shah</surname>
          </string-name>
          ,
          <article-title>An alternative approach to neural network training based on hybrid bio meta-heuristic algorithm</article-title>
          ,
          <source>Journal of Ambient Intelligence and Humanized Computing</source>
          , vol.
          <volume>10</volume>
          , no.
          <issue>10</issue>
          , pp.
          <fpage>38213830</fpage>
          ,
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>R.</given-names>
            <surname>Mendes</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Cortez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Rocha</surname>
          </string-name>
          , and
          <string-name>
            <given-names>J.</given-names>
            <surname>Neves</surname>
          </string-name>
          ,
          <article-title>Particle swarms for feedforward neural network training</article-title>
          ,
          <source>Proceedings of the 2002 International Joint Conference on Neural Networks. IJCNN02 (Cat. No.02CH37290)</source>
          ,
          <year>2002</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>