<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Breast Cancer Classification Using Seahorse Swarm Optimization</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Tanya Dixit</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Arunima Jaiswal</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Manaswini De</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Department of Computer Science and Engineering, Indira Gandhi Delhi Technical University for Women</institution>
          ,
          <addr-line>Kashmere Gate, Delhi - 110006</addr-line>
        </aff>
      </contrib-group>
      <abstract>
        <p>Globally, breast cancer ranks as the most widespread form of cancer among women. Machine Learning and Deep Learning approaches provide more effective means for detecting and managing this condition when compared to conventional detection methods. Advanced deep learning techniques, including LSTM (Long Short-Term Memory), Gate Recurrent Units (GRU), and Deep Belief Networks (DBN) have been used for classification of cancer. In this paper, the publicly available breast cancer dataset namely Wisconsin dataset is employed to investigate the efficacies of these deep learning techniques for classification of breast cancer. Further, the network architecture parameters are tuned for achieving better results using one of the latest swarm intelligence technique namely Sea Horse Optimization. Success rate of 95.61%, 96.49% and 98.24% respectively have been achieved by the proposed SHO-LSTM, SHO-GRU and SHO-DBN models when applied to the Wisconsin dataset.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Breast Cancer Classification</kwd>
        <kwd>Long Short-term Memory Network</kwd>
        <kwd>Deep Belief Network</kwd>
        <kwd>Gated Recurrent Unit</kwd>
        <kwd>Sea horse Optimization1</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Breast cancer happens to be the most common cancer type among women globally.
According to the Global Cancer Observatory (GLOBOCAN 2020) report [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], India ranked
third worldwide in terms of cancer cases and it has been predicted that the number of
cases would be increasing drastically in future. Therefore, studies on prediction of breast
cancer have great importance of prevention and control of this disease. Various studies
have been carried out to highlight the present scenario, challenges and cancer awareness
among women in India [
        <xref ref-type="bibr" rid="ref2 ref3 ref4">2-4</xref>
        ]. Leveraging Machine Learning (ML) and Deep Learning (DL)
techniques for breast cancer classification not only enhances the faster detection at an
early stage with improved accuracy but also contributes to better patient care by reducing
the subjectivity of human interpretation of medical information. Authors of this paper
have earlier studied the impact of reduction of features on breast cancer classification data
by employing classical and quantum machine learning algorithms using the Wisconsin
dataset [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. Further, the effect tuning of hyper-parameters on deep learning models such
as Long Short-Term Memory (LSTM) networks, Gate Recurrent Units (GRU) networks and
Deep Belief Networks (DBN) has been investigated in this paper.
      </p>
      <p>
        Various deep learning techniques e.g., MLP, RNN, LSTM, GRU and DBN play crucial roles
in breast cancer detection by capturing context and features using sequential data. LSTM
and GRU are two variants of Recurrent Neural Networks (RNN) that use gating mechanism
to selectively update information overtime, with GRU being simpler and faster than LSTM
[
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. DBNs, a form of artificial neural networks, are adept at unsupervised feature learning.
They consist of layers of Restricted Boltzmann Machines (RBMs), which are generative
models capable of learning valuable features from raw input data without supervision [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ].
      </p>
      <p>
        In Machine Learning, selection of optimal values of various parameters of a model is
termed as hyper-parameter tuning. Hyper-parameters are basically the configuration
settings that control the learning process of a model, such as learning rate, the number of
neural network layers, model architecture, batch size, activation functions, etc. specific to
the model employed. Various strategies such as grid search and random search,
metaheuristic optimization, swarm optimization, Bayesian Optimization, etc. are generally used
as optimization technique in hyper-parameter tuning. Among these approaches, swarm
intelligence plays a significant role in achieving optimal solutions and reducing detection
time for efficient control of disease. Sea-horse optimization (SHO) is a new swarm
intelligence-based meta-heuristic approach rooted in swarm intelligence, drawing
inspiration from the captivating behaviors observed in sea horses in their natural
environment [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]. In this study, the detection of breast cancer using deep learning models
namely LSTM, GRU and DBN has been attempted and the results of the optimized
SHOLSTM, SHO-GRU and SHO-DBN are compared when applied to the Wisconsin dataset for
breast cancer classification.
      </p>
      <p>The structure of this paper is organized as follows: Section 2 offers a concise literature
review covering diverse techniques employed for breast cancer detection using ML, DL
and hyper-parameter optimization. Section 3 delves into details regarding the dataset
employed in this study. Methodology is elucidated in Section 4, followed by a discussion of
the results in Section 5. Lastly, conclusions based on the findings are drawn in Section 6.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related Work</title>
      <p>A brief overview of a few relevant research works carried out in the field for detection
of breast cancer using various techniques has been presented in this section.</p>
      <p>
        The authors of [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ], proposed a stacked GRU-LSTM-BRNN deep learning model using
Recurrent Neural Networks (RNNs) to classify patients’ health records as benign or
malignant for breast cancer diagnosis using the Wisconsin Breast Cancer Dataset. The
paper compares the three baseline modes: RNN, stacked LSTM and stacked GRU using
metrics like accuracy, MSE and Cohen-Kappa score. The results reported by the proposed
model outperform the baseline models on all metrics, achieving 97.34% accuracy, 0.97
F1score, 0.03 MSE and 0.94 Cohen-Kappa score.
      </p>
      <p>
        In [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ], the researchers proposed a swarm intelligence-based meta-heuristic technique
called the Sea-horse optimizer (SHO). SHO replicates diverse movement patterns and the
probabilistic predation mechanism observed in sea horses. The performance of SHO has
been assessed across 23 established functions and the CEC2014 benchmark function.
Experimental findings showcase SHO as a proficient optimizer capable of effectively
addressing constraint problems. The authors of [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] introduced a novel Chaotic Sea Horse
Optimization with DL models (CSHODL-PDC) for classification of pneumonia on CXR
images. CSHODL-PDC makes use of NASNet Large model and Fuzzy Deep Neural Network.
The proposed model achieved the maximum accuracy of 99.22%, 98.96% precision and
recall of 99.22%.
      </p>
      <p>
        The researchers of [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] proposed a Particle Swarm Optimization (PSO) optimized
Multilayer Perceptron Neural Network (MLP). This model is compared with other machine
learning models like K-Nearest Neighbors, Decision Tree and Naïve Bayes and shows
higher accuracy, sensitivity and specificity. In [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ], a stacked GRU (SGRU) for Deep
Transfer Learning is used for breast cancer classification along Chaotic Sparrow Search
Algorithm (CSSA) for hyper-parameter optimization. The proposed model achieves the
accuracy of 98.61% compared to existing models when applied on a benchmark image
dataset.
      </p>
      <p>
        The paper [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ] aims to overcome the limitations of Back Propagation Learning
Algorithm for RNNs, such as slow convergence, local minima and long term dependencies
by using the R programming language to implement proposed models and compare them
with standard RNN and LSTM. It employs four distinct meta-heuristic algorithms –
Harmony Search, Ant Lion Optimization, Sine Cosine and Grey Wolf Optimizer – to train
the LSTM model for classification tasks using real and medical time-series datasets,
including the Breast Cancer Wisconsin and Epileptic Seizure Recognition datasets. The
proposed models have been reported to achieve higher accuracy rates than the standard
ones on both data sets.
      </p>
      <p>
        The study outline in [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ] presents an arithmetic optimization algorithm combined with
deep-learning-based histopathological breast cancer classification (AOADL-HBCC), which
comprises four sequential steps: noise elimination and contrast enhancement, feature
extraction utilizing AOA and SqueezeNet, feature selection employing DBN, and
classification utilizing the Adamax Optimizer. The results show that the proposed model
achieves the highest accuracy of 96.77% on the 100x dataset and 96.4% on 200x dataset,
outperforming the other models. The study in [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] introduces a hybrid model that
individually trains Random Forest (RF), MLP, and DBN on the Wisconsin Breast Cancer
dataset. These models are then integrated using a weighted average method to achieve
final classification. The proposed model achieves 96.5% accuracy against individual
accuracies of 93.9%, 91.3% and 97.5% respectively.
      </p>
      <p>The paper [14] proposes an Enhanced Sea Horse Optimization (ESHO) combined with
sine-cosine and Tent Choatic Mapping to adaptively tune the ResNet-50 parameters and
optimize its performance on two agricultural image datasets: jade fungus and corn
diseases. The ResNet-50 model optimized by ESHO achieves an accuracy of 96.7% for corn
disease image recognition and 96.4% for jade fungus image recognition. In [15], a novel
hybrid PSO-SHO algorithm combining the advantages of PSO and SHO is proposed. The
proposed approach strives to minimize the real power losses and the voltage deviation of
the power system by optimizing the generator voltages, transformer tap settings and
reactive power compensators.</p>
    </sec>
    <sec id="sec-3">
      <title>3. Data</title>
      <p>The dataset that is employed is the Wisconsin Breast Cancer Dataset (WBC) [16]. The
WBC dataset consists of 569 samples of breast cancer patients with a distribution of 357
benign and 212 malignant cases. 30 characteristic features are quantified from digitized
images of breast masses. They are numerical measures of attributes like cell nuclei, means,
standard errors and worst values of texture, perimeter, radius, smoothness, compactness,
area, concavity, symmetry, concave points and fractal dimension. The classifying labels are
denoted by 0 (benign) and 1 (malignant) under the attribute ‘diagnosis’.</p>
      <sec id="sec-3-1">
        <title>3.1. Data Pre-Processing</title>
        <p>The data preprocessing involves cleaning, scaling features, and encoding labels for the
target variables. Subsequently, it is portioned into training and testing sets, with 80%
allocated for training data and 20% for testing data. Further, reshaping is used of the
training and testing feature arrays to add an additional dimension representing number of
channels. LSTMs and GRUs expect input data in a specific 3D shape, typically represented
as (batch_size, timesteps, and features). For deep learning models, particularly those using
RNNs, input data should be reshaped to meet the input requirements of the model [17].</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Proposed Method</title>
      <p>In this paper, Gate Recurrent Unit Networks (GRU), Long Short-Term Memory (LSTM)
and Deep Belief Networks (DBN) have been used to predict breast cancer. Further, we
have designed these models with optimized parameters for our classification problem. In
order to achieve this, we have used the Seahorse Optimization algorithm to optimize the
hyper-parameters and achieve better accuracy. The focus of optimization is on various
parameters involved in these neural networks, like learning rate, filter numbers, neurons,
epochs etc. The fitness function optimizes based on the classification accuracy. The figure
2 demonstrates a flowchart of our proposed method.</p>
      <p>The process begins with by selecting a deep learning mode, such as GRU, LSTM, or DBN.
The input data is preprocessed. Seahorse Optimizer’s (SHO) parameters like epochs,
population size are initialized along with the Objective function that optimizes based on
the model’s accuracy. SHO then generates hyper-parameters for the chosen model, and the
respective model is trained with these hyper-parameters. These steps are repeated till the
maximum number of iterations is completed. Overtime, SHO produces the highest global
accuracy, and we subsequently assess the model’s performance using various traditional
performance metrics based on the parameters associated with this accuracy. The models
and optimizer’s involved are discussed in detail in the coming sections.</p>
      <sec id="sec-4-1">
        <title>4.1. Long Short-Term Memory (LSTM)</title>
        <p>Long Short-Term Memory (LSTM) is a deep learning model that addresses the
challenges of learning long-term dependencies [18], which traditional recurrent neural
networks struggle with. They are specialized neural networks designed to handle
sequential data and learn long-term dependencies. LSTMs extend the basic RNN cell [19].
The basic RNN cell takes input at each step, computes a hidden state based on this input
and its previous hidden state. The output from the cell helps in training and prediction.
Instead of this simple cell, an LSTM cell contains three key components: (1) Input Gate:
This gate controls how the amount of new information enters the cell, (2) Forget Gate:
Determines which information from the previous hidden state needs to be forgotten, (3)
Output Gate: Sets the output based on the current input and hidden state. Additional to the
hidden state (similar to RNN’s hidden state), LSTMs have a cell state which is essentially a
separate memory component that stores long-term information. Cell state is updated with
the combined information of input, forget and output gates.</p>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. Gate Recurrent Units (GRU)</title>
        <p>
          Gated Recurrent Units (GRUs) stem from RNNs, incorporating a gating mechanism
initially introduced in [20]. Similar to LSTMs, GRUs employ gating mechanism to
selectively incorporate or forget certain features, albeit lacking an output gate, thereby
resulting in fewer parameters compared to LSTMs. GRU also finds use in processing
sequential data such as text, speech and time-series data. GRU can control the flow of
information from previous activation state while computing the new activation state [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ].
Compared to LSTM, GRU has a superior convergence rate as it has lesser number of
parameters and can outperform LSTM models [18]. The GRU comprises two gating
mechanism: the reset gate and the update gate. The reset gate regulates the extent to
which the previous hidden state is forgotten, while the update gate determines the
amount of new input required to update the hidden state.
        </p>
      </sec>
      <sec id="sec-4-3">
        <title>4.3. Deep Belief Networks (DBN)</title>
        <p>
          Deep Belief Network (DBN) is a generative graphical model, composed of multiple
layers of latent variables or more popularly known as “hidden units”. They have
connections between the layers but no between units within each layer [19]. Consider
visualizing these systems as intricate, multi-layer networks with each layer processing
information from the preceding one, progressively constructing a sophisticated
comprehension of the entire dataset. DBNs are built by layering simple, unsupervised
networks such as Restricted Boltzmann Machines (RBMs) or auto encoders. The
configuration of the output layer in a DBN is contingent upon the specific task at hand. For
instance, in a classification task involving k classes, the DBN would utilize k SoftMax units,
each dedicated to one class [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ]. The hidden layer of each sub-network serves as the visible
layer for the next one in sequence. The training process involves contrastive divergence
applied layer by layer, from the lowest visible layer, which serves as the training set [21].
        </p>
      </sec>
      <sec id="sec-4-4">
        <title>4.4. Hyper-parameter Optimization of Deep Learning Models</title>
        <p>Hyper-parameters are external configuration variables set by programmers to operate
model training. They are parameters that define the details of learning process. Examples
of hyper-parameter optimization encompass various instances such as learning rate
(which regulates the magnitude of steps taken during gradient descent optimization),
batch size (which dictates the quantity of training examples utilized in each iteration of
gradient descent), the number of hidden layers, activation functions (like ReLU, sigmoid,
tanh), dropout rate (indicating the fraction of neurons within dense layers, among others.
Hyper-parameter optimization (HPO) is the process of selecting optimal values for a
machine/deep learning model’s hyper-parameters. HPO can be seen as the last step of
model design and the initial step of neural network training [22]. It finds a tuple of
hyperparameters that gives an optimal model with enhanced accuracy or prediction. Over the
years many techniques have been employed for hyper-parameter optimization such as
Grid Search, Random Search, Bayesian Optimization, Genetic Algorithms and even swarm
based techniques like Particle Swarm Optimization and Ant Colony Optimization [23]. For
the optimization of DL models proposed in this study we have taken the hyper-parameters
as represented in table 1, along with the new Seahorse Optimization Algorithm.</p>
      </sec>
      <sec id="sec-4-5">
        <title>4.5. Seahorse Optimization Algorithm</title>
        <p>
          Introduced in 2022, the Seahorse Optimization (SHO) algorithm represents a novel
swarm-based meta-heuristic optimization approach [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ]. SHO replicates the natural
movement, hunting and breeding patterns observed in seahorses. Seahorse movement
behavior encompasses two scenarios: (1) Spiral Movement of the hippocampus in
conjunction with the ocean vortex and (2) Brownian motion of the hippocampus amidst
the waves. Further, predatory behavior consists of the following two situations: success
and failure. The breeding behavior of seahorse is described by random mating and
offspring inherit traits of both parents. An equal mix of male and female seahorses is taken
in the population. To enhance the balance of the SHO algorithm, global strategies are
applied to motion behavior and local strategies are applied to predation behavior.
        </p>
        <p>The following equations (1) and (2) denote the spiral and Brownian movement
behaviors of seahorses respectively. SHO utilizes Lévy flight to emulate the spiral
movement observed in seahorses, which aids in preventing SHO from becoming trapped
in local optima. In the spiral movement, the three dimensional vector of coordinates
(x, y, z) is denoted by x, y and z. Regarding the Brownian equationl denotes the constant
coefficient and βt represents the coefficient for the motion random walk.</p>
        <p>Equation (3) is the mathematical representation of predation, withr2representing a
random number generated by SHO to distinguish between success and failure scenarios. If
r2exceeds 0.1, the predation by the seahorse is deemed successful; otherwise, it results in
failure. α denotes the step size of the seahorse’s movement in pursuit of prey.
 2 ( + 1) =  ( ) = {  × ( 
−</p>
        <p>
          ×  1 ( )) + (1 −  ) ×   ,
(1 −  ) × ( 1 ( ) − 
×  
) +  ×  1
( ),
seahorses. Xs2ort is fitness values of the population in ascending order. (6) denotes the
mathematical equation representing an offspring. r3 is a random number between [
          <xref ref-type="bibr" rid="ref1">0, 1</xref>
          ].
 ℎ =  
2
        </p>
        <p>(1:  ⁄2)</p>
        <p>ℎ =  2 ( ⁄2 + 1:  )



=  3</p>
        <p>ℎ + (1 −  3)  ℎ
 
 ( + 1) =   ( ) +</p>
        <p>(λ)((  ( ) −   ( ) ×  ×  ×  +   ( ))</p>
        <p>X1new(t + 1) = Xi(t) + rand× l × βt × (Xi(t) − βt × Xelite)</p>
        <p>The algorithm starts with a randomly generated population of seahorses, with each
seahorse representing a potential solution. It uses a normal distribution to decide among
the two movement behaviors of seahorses. It mimics the high success rate of seahorses in
hunting to enhance exploitation capabilities. It draws inspiration from the breeding
behavior of seahorses to generate new solutions, hoping to improve upon the current best
solution. If the hunt is successful, the seahorse (problem) moves towards the prey (best
solution) otherwise search space exploitation continues. The algorithm followed by SHO is
explained in the following flowchart:</p>
        <p>The SHO algorithm maintains a balance between exploration (diversification) and
exploitation (intensification) to mitigate the risk of getting stuck in local optima and to
efficiently locate the global optima. This algorithm has been applied to various engineering
design problems, and shows promising results.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Results</title>
      <p>Various Performance metrics namely accuracy, F1-score, recall, precision and
specificity have been used for comparison of results between standard LSTM, GRU and
DBN and their SHO optimized counterparts and these parameters are given in the
following equations (7)-(11).</p>
      <p>1 



=  + 
+ 
+  +   100
=   +</p>
      <p>= 
2 × 
=</p>
      <p>Accuracy is the ratio of correctly classified instances to the aggregate instances. TP
(True Positive), TN (True Negative), FP (False Positive) and FN (False Negative) are the
number of correctly and incorrectly classified positive and negative cases. Sensitivity or
recall is the fraction of positive cases that the classifier correctly identifies, whereas
specificity is the fraction of negative cases that a classifier correctly identifies. Precision is
the proportion of true positives out of all predicted positives. F1-score is a measure of
performance that combines precision and recall as it is the harmonic mean of the two
metrics.</p>
      <p>The results obtained from DL models with and without SHO optimization are presented
in the following Table 2 and Table 3 respectively.</p>
      <p>Before optimization, the LSTM and GRU models achieved accuracies of 92.10% and
93.85%, respectively, while the DBN model achieved an accuracy of 92.11%. However
after the hyper-parameter optimization with SHO, the performances of all models
significantly improved. The SHO-LSTM achieved an accuracy of 95.61%, SHO-GRU
demonstrated an accuracy of 96.49% and SHO-DBN acquired an accuracy of 98.25%.</p>
      <p>Furthermore, the SHO-DBN model consistently outperformed the other models across
all metrics, demonstrating its effectiveness in breast cancer detection. Notably, the
SHOGRU model achieved perfect recall (100%), indicating its ability to correctly identify all
positive cases of breast cancer. GRU demonstrated better performance than LSTM across
all metrics with or without optimization, indicating its greater efficiency due to lesser
number of model parameters. The graphical comparisons of performance metrics for the
un-optimized and optimized DL models are presented in Figure 4(a), 4(b), and 4(c).</p>
    </sec>
    <sec id="sec-6">
      <title>6. Conclusion</title>
      <p>Machine learning and deep learning techniques offer more efficient ways to detect and
manage breast cancer compared to traditional methods. Advanced DL models, namely
LSTM, GRU and DBN have been successfully applied to Wisconsin dataset for breast cancer
classification. Additionally, network architecture parameters are fine-tuned using the
state-of-the-art swarm intelligence technique known as Sea Horse Optimization. In this
study, we explored the effectiveness of these deep learning techniques for achieving
improved performances when applied to two different breast cancer datasets. Specifically,
the SHO-DBN model emerges as the most effective model for this task, with the highest
accuracy, precision, specificity and F1-Score. These findings underscore the potential of
optimized deep learning models as valuable tools in the early detection and management
of breast cancer.
Histopathological Breast Cancer Diagnosis”, Cancers, vol. 15, p. 885, 2023, doi:
10.3390/cancers15030885.
[14] Z. Li, S. Qu, Y. Xu, X. Hao, N. Lin, “Enhanced Sea Horse Optimization Algorithm for
Hyperparameter Optimization of Agricultural Image Revolution”, Mathematics, vol. 12, no.
3, p. 368, 2024, doi:10.3390/math12030368.
[15] H. M. Hasanien, I. Alsaleh, M. Tostado-Véliz, M. Zhang, A. Alateeq, F. Jurado, A. Alassaf,
“Hybrid particle swarm and sea horse optimization algorithm-based optimal reactive
power dispatch of power systems comprising electric vehicles”, Energy, vol. 286, pp.
129583, 2024, doi:10.1016/j.energy.2023.129583.
[16] “Diagnostic Wisconsin Breast Cancer Database”, UCI Machine Learning Repository.</p>
      <p>Available: https://archive.ics.uci.edu/dataset/17/breast+cancer+wisconsin+diagnostic.
[17] M. Chhetri, S. Kumar, P. P. Roy, B.-G. Kim, Deep BLSTM-GRU “Model for Monthly Rainfall
Prediction: A Case Study of Simtokha”, Bhutan, Remote Sensing, vol. 12, no. 19, p. 3174,
2020, doi:10.3390/rs12193174.
[18] A. Sherstinsky, “Fundamentals of recurrent neural network (RNN) and long short-term
memory (LSTM) network”, Physica D: Nonlinear Phenomena, vol. 404, p. 132306,
2020, doi: 10.1016/j.physd.2019.132306.
[19] G. E. Hinton, “Deep belief networks”. In E. N. Zalta (Ed.), Scholarpedia, vol. 4.5, p. 5947,
2009, doi:10.4249/scholarpedia.5947. Available:
http://www.scholarpedia.org/article/Deep_ belief_networks.
[20] J. Chung, C. Gulcehre, K. Cho, Y. Bengio, “Empirical evaluation of gated recurrent neural
networks on sequence modeling”, arXiv preprint, arXiv: 1412.3555v1, 2014, doi:
10.48550/arXiv.1412.3555.
[21] Y. Bengio, P. Lamblin, D. Popovici, H. Larochelle, “Greedy layer-wise training of deep
networks”. In B. Schölkopf, J. Platt, T. Hoffman (Eds.), Advances in Neural Information
Processing Systems, MIT Press, vol. 19, pp. 153-160, 2006. Available:
https://deeplearning.cs.cmu.edu/ pdfs/Bengio_et_al_NIPS_2006.pdf.
[22] T. Yu, H. Zhu, “Hyper-Parameter Optimization: A Review of Algorithms and Applications”,
2020. Available: https://arxiv.org/abs/2003.05689.
[23] W.-C. Yeh, Y.-P. Lin, Y.-C. Liang, C.-M. Lai, X.-Z. Gao, “Simplified Swarm Optimisation for the
Hyperparameters of a Convolutional Neural Network”, 2023. Available:
https://arxiv.org/ftp/arxiv/papers/2103/2103.03995.pdf.
[24] H. Alahmer, A. Alahmer, M. I. Alamayreh, M. Alrbai, R. Al-Rbaihat, A. Al-Manea, R.</p>
      <p>Alkhazaleh, “Optimal water addition in emulsion diesel fuel using machine learning and
sea-horse optimizer to minimize exhaust pollutants from diesel engine”, Atmosphere, vol.
14, p. 449, 2023, doi: 10.3390/atmos14030449.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>H.</given-names>
            <surname>Sung</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Ferlay</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. L.</given-names>
            <surname>Siegel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Laversanne</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Soerjomataram</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Jemal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Bray</surname>
          </string-name>
          , “Global cancer statistics
          <year>2020</year>
          :
          <article-title>GLOBOCAN estimates of incidence and mortality worldwide for 36 cancers in 185 countries”</article-title>
          ,
          <source>CA: A Cancer Journal for Clinicians</source>
          , vol.
          <volume>71</volume>
          , no.
          <issue>3</issue>
          , pp.
          <fpage>209</fpage>
          -
          <lpage>249</lpage>
          ,
          <year>2021</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>A A.</given-names>
            <surname>Gupta</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Shridhar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. K.</given-names>
            <surname>Dhillon</surname>
          </string-name>
          , “
          <article-title>A review of breast cancer awareness among women in India: Cancer literate or awareness deficit?”</article-title>
          ,
          <source>European J. Cancer</source>
          , vol.
          <volume>51</volume>
          , no.
          <issue>14</issue>
          , pp.
          <fpage>2058</fpage>
          -
          <lpage>2066</lpage>
          ,
          <year>2015</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>R.</given-names>
            <surname>Mehrotra</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Yadav</surname>
          </string-name>
          ,
          <article-title>Breast cancer in India: “Present scenario and the challenges ahead”</article-title>
          ,
          <source>World journal of clinical oncology</source>
          , vol.
          <volume>13</volume>
          .3, pp.
          <fpage>209</fpage>
          -
          <lpage>218</lpage>
          ,
          <year>2022</year>
          , doi:10.5306/wjco.v13.
          <year>i3</year>
          .
          <fpage>209</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>K.</given-names>
            <surname>Sathishkumar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Chaturvedi</surname>
          </string-name>
          ,
          <string-name>
            <surname>P. Das</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Stephen</surname>
          </string-name>
          , P. Mathur, “
          <article-title>Cancer incidence estimates for 2022 &amp; projection for 2025: Result from National Cancer Registry Programme”</article-title>
          ,
          <source>The Indian journal of medical research</source>
          , vol.
          <volume>156</volume>
          , no.
          <issue>4</issue>
          &amp;
          <issue>5</issue>
          , pp.
          <fpage>598</fpage>
          -
          <lpage>607</lpage>
          ,
          <year>2022</year>
          , doi:10.4103/ijmr.ijmr_
          <year>1821</year>
          _
          <fpage>22</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <surname>M. De</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Jaiswal</surname>
          </string-name>
          , T. Dixit, “
          <article-title>Comparative analysis of classical and quantum machine learning for breast cancer classification”</article-title>
          ,
          <source>International Conference on Computational Intelligence and Mathematical Applications</source>
          ,
          <year>2023</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>S.</given-names>
            <surname>Dutta</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. K.</given-names>
            <surname>Mandal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T. H.</given-names>
            <surname>Kim</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. K.</given-names>
            <surname>Bandyopadhyay</surname>
          </string-name>
          , “
          <article-title>Breast Cancer Prediction Using Stacked GRU-LSTM-BRNN”</article-title>
          ,
          <source>Applied Computer Systems</source>
          , vol.
          <volume>25</volume>
          , no.
          <issue>2</issue>
          , pp.
          <fpage>163</fpage>
          -
          <lpage>171</lpage>
          ,
          <year>2020</year>
          , doi: 10.2478/acss-2020-0018.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>S.</given-names>
            <surname>Yamani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z. H.</given-names>
            <surname>Choudhury</surname>
          </string-name>
          , “
          <article-title>Integrating Random Forest, MLP and DBN in a Hybrid Ensemble Model for Accurate Breast Cancer Detection”</article-title>
          ,
          <source>International Journal of Innovative Science and Research Technology</source>
          , vol
          <volume>8</volume>
          , no.
          <issue>7</issue>
          , pp.
          <fpage>1556</fpage>
          -
          <lpage>1564</lpage>
          ,
          <year>2023</year>
          , doi:10.31219 /osf.io/sdjqf.
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>S.</given-names>
            <surname>Zhao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , S. Ma, M. Wang, “
          <article-title>Sea-horse optimizer: a novel nature-inspired metaheuristic for global optimization problems”</article-title>
          ,
          <source>Applied Intelligence</source>
          , vol.
          <volume>53</volume>
          , no.
          <issue>10</issue>
          , pp.
          <fpage>11833</fpage>
          -
          <lpage>11860</lpage>
          ,
          <year>2023</year>
          , doi:10.1007/s10489-022-03994-3.
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>V.</given-names>
            <surname>Parthasarathy</surname>
          </string-name>
          , S. Saravanan, “
          <article-title>Chaotic Sea Horse Optimization with Deep Learning Model for lung disease pneumonia detection and classification on chest X-ray images”</article-title>
          ,
          <source>Multimedia Tools and Applications</source>
          ,
          <year>2024</year>
          , doi: 10.1007/s11042-024-18301-0.
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>M.</given-names>
            <surname>Alimardani</surname>
          </string-name>
          , M. Almasi, “
          <article-title>Investigating the application of particle swarm optimization algorithm in the neural network to increase the accuracy of breast cancer prediction”</article-title>
          ,
          <source>International Journal of Computer Trends and Technology</source>
          , vol.
          <volume>68</volume>
          , no.
          <issue>4</issue>
          , pp.
          <fpage>65</fpage>
          -
          <lpage>72</lpage>
          ,
          <year>2020</year>
          , doi: 10.14445/22312803/IJCTT-V68I4P112.
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>K.</given-names>
            <surname>Shankar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. K.</given-names>
            <surname>Dutta</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Kumar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G. P.</given-names>
            <surname>Joshi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>I. C.</given-names>
            <surname>Doo</surname>
          </string-name>
          , “
          <article-title>Chaotic Sparrow Search Algorithm with Deep Transfer Learning Enabled Breast Cancer Classification on Histopathological Image”</article-title>
          ,
          <source>Cancers</source>
          , vol.
          <volume>14</volume>
          , no.
          <issue>11</issue>
          ,
          <year>2022</year>
          , p.
          <volume>2770</volume>
          ,
          <issue>doi</issue>
          : 10.3390/cancers14112770.
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>T. A.</given-names>
            <surname>Rashid</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Fattah</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. K.</given-names>
            <surname>Awla</surname>
          </string-name>
          , “
          <article-title>Using accuracy measure for improving the training of LSTM with metaheuristic algorithms”</article-title>
          ,
          <source>Procedia Computer Science</source>
          , vol.
          <volume>140</volume>
          , pp.
          <fpage>324</fpage>
          -
          <lpage>333</lpage>
          ,
          <year>2018</year>
          , doi:10.1016/j.procs.
          <year>2018</year>
          .
          <volume>10</volume>
          .307.
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>M.</given-names>
            <surname>Obayya</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Alkhalaf</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Alrowais</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Alshahrani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Alzahrani</surname>
          </string-name>
          ,
          <string-name>
            <surname>A</surname>
          </string-name>
          . Alzahrani, “
          <article-title>Hyperparameter Optimizer with Deep Learning-Based Decision-Support Systems for</article-title>
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>