<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Evaluation of Initialization Methods for the Archerfish Hunting Optimizer: A Comparative Study</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Aridj Ferhat</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Farouq Zitouni</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Abdelhadiimane Limane</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Rihab Lakbichi</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Saad Harous</string-name>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Department of Computer Science and Information Technology, Laboratory of Articfiial Intelligence and Information Technology, Kasdi Merbah University - Ouargla</institution>
          ,
          <country country="DZ">Algeria</country>
        </aff>
      </contrib-group>
      <kwd-group>
        <kwd>eol&gt;Metaheuristic Algorithms</kwd>
        <kwd>Population Initialisation</kwd>
        <kwd>Probability Distributions</kwd>
        <kwd>Low-Discrepancy Sequences</kwd>
        <kwd>Archerfish Hunting Optimizer</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>Optimization is a fundamental concept in various fields such as artificial intelligence [ 1], computer
science [2], machine learning [3], and operations research [4]. It involves finding the best possible
solution, either a minimum or maximum, to a given objective function. Traditional optimization
techniques, including linear programming [5], quadratic programming [6], convex optimization [7], and
interior-point methods [8], have been widely used to address these challenges. However, the growing
complexity and scale of real-world problems have led researchers to explore more flexible and robust
approaches, such as Metaheuristic Algorithms (MAs), which have gained significant attention in recent
years.</p>
      <p>MAs are high-level problem-independent techniques that explore the search space to find
nearoptimal solutions within a reasonable computational time. They are inspired by natural processes
and phenomena, including biological evolution, animal behaviours, and physical principles. Examples
of widely used MAs include the genetic algorithm [9], inspired by the process of natural selection,
the particle swarm optimization [10], which mimics the social behaviour of bird flocking and fish
schooling, the ant colony optimization [11], based on the foraging behaviour of ants, and the solar
system algorithm [12], which mimics the orbiting behaviour of celestial objects found in the solar
system. These MAs have demonstrated considerable success across various applications due to their
adaptability and robustness in exploring complex and large search spaces.</p>
      <p>One of the key factors influencing the performance of population-based MAs is the initialization of
the population. The way the initial population is generated can significantly impact the algorithm’s
convergence speed, accuracy, and ability to explore the solution space efectively. Traditionally,
random number generators, which use uniform probability distributions, have been the most commonly
employed method for population initialization. However, the limitations of random initialization, such
as poor diversity and uneven coverage of the search space, have driven researchers to investigate
alternative initialization strategies [13]. Recent studies have explored various approaches to improve
population initialization, such as using diferent probability distributions (e.g., Beta, Rayleigh,
Exponential) [14], quasi-random sequences (e.g., Halton, Sobol, Faure) [15], and chaos-based methods [16].
Additionally, hybrid techniques combining MAs with other optimization methods have been proposed
to enhance the initialization process and improve the overall performance [17]. These methods aim
to ensure better diversity and coverage of the search space, which are critical for avoiding premature
convergence and ensuring that MAs can explore a wide range of possible solutions.</p>
      <p>In this paper, we examine the impact of nine distinct initialization methods on the performance of
AHO [18]. These methods include six probability distributions: Beta, Normal, Logarithmic Normal,
Exponential, Rayleigh, and Weibull distributions [14], as well as three low-discrepancy sequences:
Halton, Faure, and Sobol [15]. In addition, we analyze how varying population sizes and iteration
counts afect the optimizer’s eficiency. The findings reveal that AHO’s performance is significantly
influenced by the choice of initialization scheme. Based on these findings, we propose optimal population
sizes, iteration counts, and initialization methods for enhancing AHO’s efectiveness across diferent
benchmark functions.</p>
      <p>The paper is structured as follows: Section 2 provides a brief overview of related work in the
literature. Section 3 discusses background information on pseudo-random number generators, probability
distributions, and the Archerfish Hunting Optimizer used in our experiments, which are presented in
Section 4. The paper concludes in Section 5 with suggestions for future work to expand on the findings
of this study.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related Work</title>
      <p>Initialization is critical in implementing MAs, significantly influencing their eficiency and performance
in finding optimal solutions to complex optimization problems. Traditional approaches often rely
on random number generators to create initial populations. However, the limitations of random
initialization, such as uneven coverage of the search space, leading to premature convergence, and
poor performance in high-dimensional problems, are well-documented in the literature, prompting
researchers to explore alternative probability distributions to enhance algorithm performance [19, 13].</p>
      <p>To overcome these limitations, numerous studies have investigated the efects of diferent probabilistic
distributions as alternative initialization methods. For instance, in a comprehensive study, conducted in
[14], 22 diferent initialization methods based on distributions such as Beta, Rayleigh, and Exponential
were compared across five MAs: diferential evolution [ 20], particle swarm optimization [21], cuckoo
search [22], artificial bee colony [ 23], and genetic algorithm [9]. The results indicated that some of these
distributions could significantly improve the accuracy and convergence of these optimizers compared
to traditional random initialization.</p>
      <p>Another notable work explored the significance of initialization methods on the performance of
population-based MAs [24]. The authors examined the impact of population size, the number of
iterations, and eleven diferent initialization techniques on ten widely used optimizers, including the
bat algorithm [25], grey wolf optimizer [26], and whale optimization algorithm [27]. The study revealed
that specific algorithms were sensitive to the choice of initialization methods, which influenced their
convergence and accuracy across various test functions. Furthermore, the research underscored that
the optimal population size and the number of iterations varied depending on the specific algorithm
and problem, highlighting the importance of tailored initialization strategies to boost algorithmic
performance.</p>
      <p>In another work [15], researchers delved into the application of randomized low-discrepancy
sequences, specifically the Halton, Sobol, and Faure sequences, for initializing particle swarms in the
particle swarm optimization algorithm [28]. The study compared its performance using these
lowdiscrepancy sequences with the traditional uniform initialization method that employs a pseudo-random
number generator. The findings suggested that the use of randomized low-discrepancy sequences could
enhance the algorithm’s performance on benchmark problems. Moreover, the study indicated that
diferent low-discrepancy sequences might have varying efects on the algorithm’s performance, suggesting
that the choice of sequence could influence optimization outcomes.</p>
      <p>Finally, an article provided a comprehensive review of various initialization methods used in MAs to
enhance their performance [29]. It categorized initialization schemes into several groups, including
random numbers, quasi-random sequences, chaos theory, and probability distributions. The authors
conducted experiments comparing ten diferent initialization methods, evaluating their impact on the
performance of three MAs: bat algorithm [25], grey wolf optimizer [26], and butterfly optimization
algorithm [30]. The experiments emphasized the importance of balancing population size, the diversity
of the initial population, and the number of iterations as crucial factors in achieving optimal solutions
in complex optimization problems.</p>
    </sec>
    <sec id="sec-3">
      <title>3. Background</title>
      <p>This section provides an overview of AHO and the various initialization methods adopted in this study.
These foundational concepts are critical to understanding how these initialization techniques influence
AHO’s performance.</p>
      <sec id="sec-3-1">
        <title>3.1. Archerfish Hunting Optimizer</title>
        <p>AHO [18] is an MA inspired by the hunting strategies of archerfish, which capture prey by either
shooting water droplets or leaping out of water. AHO translates these behaviours into an optimization
algorithm where solutions are refined through iterations, focusing on eficient search space exploration
and the avoidance of local optima. The algorithm operates with two key parameters: the attractiveness
rate () and the swapping angle ( ).</p>
        <p>The algorithm begins by initializing a population of candidate solutions x within the search space
boundaries using Equation 1. The symbols r represent -dimensional vectors drawn from the uniform
distribution. The vectors ub and lb denote the upper and lower bounds of the search space, respectively.</p>
        <p>x = r ∘ (ub − lb) + lb ,  ∈ {1, . . . ,  }</p>
        <p>The generation of new solutions during the search process is governed by three key phases:
exploration, exploitation, and local optima avoidance. The exploration phase uses Equation 2; the exploitation
phase renfies solutions by employing Equation 3; and to avoid getting trapped in local optima, Equation
4 is utilized. The symbol  is a random number uniformly distributed between 0 and 1. The symbol L
is a vector derived from the Lévy flight distribution. The symbol  is a small number. The symbol  is a
random number drawn from {0, 1}.</p>
        <p>xnew = xbest + exp (︀ −‖ x1 − x‖2 × (x1 − x)</p>
        <p>2)︀
xnew = xbest + exp (︀ −‖ x2 − x‖2 × (x2 − x)</p>
        <p>
          2)︀
xnew = x +  × L
x1 = x +  × 1 + v1 , x2 = x +  × 1 + v2
v() = ︂{ 0 ,  ∈ {1, . . . , } ∖ { (
          <xref ref-type="bibr" rid="ref1">1</xref>
          )}
1  × sin(2 ×  0) ,  =  (
          <xref ref-type="bibr" rid="ref1">1</xref>
          )
        </p>
        <p>
          ⎧ 0 ,  ∈ {1, . . . , } ∖ { (
          <xref ref-type="bibr" rid="ref1">1</xref>
          ),  (
          <xref ref-type="bibr" rid="ref2">2</xref>
          )}
v2() = ⎨  × sin(2 ×  0) ,  =  (
          <xref ref-type="bibr" rid="ref1">1</xref>
          )
⎩  × sin2( 0) ,  =  (
          <xref ref-type="bibr" rid="ref2">2</xref>
          )
 0 = (− 1) ×  × 
(
          <xref ref-type="bibr" rid="ref1">1</xref>
          )
(
          <xref ref-type="bibr" rid="ref2">2</xref>
          )
(
          <xref ref-type="bibr" rid="ref3">3</xref>
          )
(
          <xref ref-type="bibr" rid="ref4">4</xref>
          )
        </p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Probability Distributions</title>
        <p>Usually, distributions can be used as initial population schemes, where each afects MAs’ search
behaviour diferently. In the following sections, some common distributions are described.</p>
        <sec id="sec-3-2-1">
          <title>3.2.1. Beta Distribution</title>
          <p>
            It is defined over (
            <xref ref-type="bibr" rid="ref1">0, 1</xref>
            ) with shape parameters  and . It can be written as  ∼
density function (PDF) is given by Equation 5, where the symbol Γ represents the Gamma function.
Be(, ). Its probability
Also, its expected value and variance are provided in Equation 6.
︂(
︂(
          </p>
        </sec>
        <sec id="sec-3-2-2">
          <title>3.2.3. Normal Distribution</title>
          <p>Equation 9, and its mean and variance are described in Equation 10.</p>
          <p>It is characterized by mean  and variance  2. It can be written as  ∼  (,  2). Its PDF is given by</p>
        </sec>
        <sec id="sec-3-2-3">
          <title>3.2.4. Logarithmic Normal Distribution</title>
          <p>12.
ln  ∼  (, 
It is defined as the logarithm of the variable following a normal distribution. It can be written as
2). Its PDF is given by Equation 11, and its mean and variance are specified in Equation
 (; , ) =
Γ()Γ()
Γ( + ) − 1(1 − )− 1
{︃ [] = +</p>
          <p />
          <p>Var() = (+)(++1)
 () =
{︃ [] = +</p>
          <p>2
Var() = (− )2</p>
          <p>12
 (; , 
2) = √
1
2 2
exp −
( −  )2 )︂</p>
          <p>2 2
︂{ [] =</p>
          <p>Var() =  2
 (; ,</p>
          <p>2) =
{︃
[] = exp  +  22 )︁</p>
          <p>︁(
Var() = (︀ exp( 2)</p>
          <p>1
√
 2 2
exp −
(ln  −  )2 )︂</p>
          <p>2 2
− 1︀) exp (︀ 2 +  2)︀</p>
        </sec>
        <sec id="sec-3-2-4">
          <title>3.2.2. Uniform Distribution</title>
          <p>
            It is defined over [, ]. It can be written as  ∼
mean and variance are calculated using Equation 8.
 (, ). Its PDF is described by Equation 7, and its
(
            <xref ref-type="bibr" rid="ref5">5</xref>
            )
(
            <xref ref-type="bibr" rid="ref6">6</xref>
            )
(
            <xref ref-type="bibr" rid="ref7">7</xref>
            )
(
            <xref ref-type="bibr" rid="ref8">8</xref>
            )
(
            <xref ref-type="bibr" rid="ref9">9</xref>
            )
(
            <xref ref-type="bibr" rid="ref10">10</xref>
            )
(
            <xref ref-type="bibr" rid="ref11">11</xref>
            )
(
            <xref ref-type="bibr" rid="ref12">12</xref>
            )
          </p>
        </sec>
        <sec id="sec-3-2-5">
          <title>3.2.5. Exponential Distribution</title>
          <p>provided in Equation 14 ( &gt;</p>
          <p>0).</p>
          <p>It can be written as  ∼ exp( ). Its PDF is given by Equation 13, and its mean and variance are
 (;  ) =
{︃ exp(−  ) for  ≥ 0
0</p>
          <p>for  &lt; 0
︂{ [] = 1</p>
          <p>Var() =  12</p>
        </sec>
        <sec id="sec-3-2-6">
          <title>3.2.6. Rayleigh Distribution</title>
          <p>calculated using Equation 16.</p>
          <p>It can be written as  ∼ Rayleigh( ). Its PDF is given by Equation 15, and its mean and variance are
The Halton Sequence [34, 31] extends the Van der Corput sequence to multiple dimensions. The Van
der Corput sequence in a given base  is defined by representing a non-negative integer  in base-
form, as shown in Equation 19.
, where  are the digits in base , and  is the largest integer such that 
sequence maps  to a fraction by reversing the digits of the base- expansion, according to Equation 20.
≤ . The Van der Corput
 = 0 + 1 + 22 + · · ·</p>
          <p>+ 
 () =
0 +

21 + · · ·
+

+1
18.</p>
          <p>It generalizes several distributions with scale parameter  and shape parameter . It can be written as
 ∼  (,  ) . Its PDF is given by Equation 17, and its mean and variance are given in Equation
︁(
{︃  (︀  )︀ − 1 exp −


︀(  )︀ )︁

for  ≥ 0
for  &lt; 0
 (; ,  ) =</p>
          <p>0
{︃ [] =  Γ (︀ 1 + 1 )︀</p>
          <p>Var() =  2 [︁Γ (︀ 1 + 2 )︀ − Γ (︀ 1 + 1 )︀ 2]︁
3.2.7. Weibull Distribution
 (;  ) =</p>
        </sec>
      </sec>
      <sec id="sec-3-3">
        <title>3.3. Low-Discrepancy Sequences</title>
        <p>
          Quasi-random number generators are known to generate sequences that have low discrepancy, which
makes them powerful tools for improving optimization algorithms. Low-discrepancy sequences, such
as Van der Corput [31], Sobol [31], Faure [32], and Halton [33], are particularly useful in uniformly
covering the search space, which enhances the initialization process in MAs. This section will focus on
three common low-discrepancy sequences – Halton, Faure, and Sobol – which have shown significant
promise in initialization schemes for MAs.
3.3.1. Randomized Halton Sequence
(
          <xref ref-type="bibr" rid="ref13">13</xref>
          )
(
          <xref ref-type="bibr" rid="ref14">14</xref>
          )
(
          <xref ref-type="bibr" rid="ref15">15</xref>
          )
(16)
(17)
(18)
(19)
(20)
        </p>
        <p>For higher dimensions, the Halton sequence uses diferent prime numbers 1, 2, . . . ,  as bases, as
described in Equation 21. To address the deterministic nature of the Halton sequence, small perturbations
such as Gaussian noise are introduced, which ensures variability between multiple algorithm runs.</p>
        <p>x = ( 1 (),  2 (), . . . ,   ())
, where each   () is the radical inverse function in base .</p>
        <sec id="sec-3-3-1">
          <title>3.3.2. Randomized Faure Sequence</title>
          <p>The Faure sequence [34, 32] is a permutation of the Halton sequence with a single base , chosen as
the smallest prime number greater than or equal to the number of dimensions. The representation
of a point  is given by Equation 22. The Faure sequence improves stratification by applying a
specific permutation to the coeficients. The recursive permutation for the Faure sequence is outlined
in Equation 23.</p>
          <p>= − 1 + 1− 2 + · · ·
0</p>
          <p>+ − (+1)
 ≡
∑︁ (︂ ︂)
≥ 


mod</p>
          <p>We also apply randomization to the Faure sequence by introducing small Gaussian noise to the
coordinates. This randomization ensures that the Faure sequence can be used in randomized MAs,
providing better performance compared to pseudo-random numbers.
(21)
(22)
(23)
(24)</p>
        </sec>
        <sec id="sec-3-3-2">
          <title>3.3.3. Randomized Sobol Sequence</title>
          <p>The Sobol sequence [34, 33] is another popular low-discrepancy sequence based on linear recurrence
relations over a finite field. The -th element of the Sobol sequence in dimension , denoted by (), is
computed using the recurrence relation defined in Equation 24.</p>
          <p>() = 1()1 ⊕ 2()2 ⊕ ... ⊕ ()
() are known as direction numbers, and the symbol ⊕ denotes the bit-wise XOR operation.
, where</p>
          <p>We also introduce randomization by adding Gaussian noise to the generated points to enhance
exploratory capabilities while maintaining the low discrepancy properties of the Sobol sequence.</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Experiments, and Discussion</title>
      <p>
        To examine how diferent initialization methods influence the performance of AHO, we conducted
a series of experiments. The key objective is to assess how varying the population size (NP) and
the maximum number of iterations (T) afects the algorithm’s eficiency, with the ultimate goal of
determining the optimal values for NP and T that produce the best outcomes. The total number of
function evaluations (NFEs) is consistently set at 600,000, while the number of iterations is adjusted
based on the chosen population size. Consequently, larger NP values corresponded to fewer iterations,
while smaller NP values allowed for more iterations. The NP and T combinations tested include
1 = (
        <xref ref-type="bibr" rid="ref10">10, 60000</xref>
        ), 2 = (30, 20000), 3 = (60, 10000), and 4 = (100, 6000), with the goal of
identifying the combination that delivers the best performance for AHO.
      </p>
      <p>The experiments were conducted on a machine equipped with an Intel® Core™ i7-8750H CPU,
operating at a base frequency of 2.20 GHz with a boost frequency of 2.21 GHz. The system is equipped
with 32 GB of RAM and runs Windows 11. All computational analyses were performed using Matlab
R2023a. Each test function was evaluated with 30 independent runs to reduce variance.</p>
      <p>We evaluated the performance of AHO across functions 1, 2, 6, and 9 in dimensions 10 and 20
(Not all twelve benchmark functions were included due to space constraints). These functions were
specifically chosen to represent diferent categories of benchmark functions – unimodal, multimodal,
hybrid, and composition. Table 1 provides a detailed summary of the Best, Worst, Mean, Standard
Deviation (STD), and mean execution time (ET) values obtained from 30 independent runs. The STD was
computed based on the optimal solutions of the test functions, as defined in the CEC 2022 benchmark 1.
Finally, the table presents the statistical analysis performed using Friedman’s test, showing that the
best performance for both the 10-dimensional and 20-dimensional cases was achieved when NP = 30
and T = 20000.</p>
      <p>After determining the optimal values for NP and T, we assess the efectiveness of various probability
distributions and low-discrepancy sequences for initializing the population in AHO. Specifically, we
tested six probability distributions – Beta, Normal, Logarithmic Normal, Exponential, Rayleigh, and
Weibull – as well as three low-discrepancy sequences: Halton, Faure, and Sobol. Each distribution and
sequence is used to generate the initial populations, and the performance is compared to the original
algorithm, which utilizes a uniform distribution for initialization. We provide an overview of the
initialization methods and their parameter settings in Table 2. Additionally, the results derived from the
comparison of these methods are systematically presented in Tables 3 and 4.</p>
      <p>We give an in-depth analysis of the performance of AHO on the CEC 2022 benchmark functions,
using various initialization schemes, across two dimensions:  = 10 and  = 20. First, the discussion
will be focused on the type of test functions (unimodal, basic, hybrid, and composition). Then, an
analysis of the Friedman statistical test will be provided. Finally, a general overview of some key
observations is given.</p>
      <p>For the unimodal function 1 at dimension  = 10, initialization schemes such as 2, 3, 4,
5, and 6 exhibited very poor performance, with significantly high mean values and large standard
deviations, reflecting instability in the performance. This contrasts with their performance at  = 20,
where the results were more consistent across all initialization methods. At  = 20, the best, worst,
and mean values were identical at 300 for all methods, indicating that this unimodal function does
not efectively distinguish between diferent initialization techniques when applied in the context of
AHO. This suggests that for unimodal functions, the choice of initialization method becomes more
1https://github.com/P-N-Suganthan/2022-SO-BO
Overview of the used probability distributions and low-discrepancy sequences</p>
      <p>Initialization Method</p>
      <p>Parameter
Random (0)
Beta (1)
Normal Distribution (2)
Logarithmic Normal (3)
Exponential (4)
Rayleigh (5)
Weibull (6)
Sobol (1)
Halton (2)
Faure (3)
(, )
( ,  )
(,  )
(,  )
–


–
–
–</p>
      <p>Value
–
critical at lower dimensions, where instability in certain methods can negatively impact performance.
However, at higher dimensions, this sensitivity diminishes, leading to more uniform outcomes across
various methods. When analyzing the basic functions (2 to 5) at  = 10, the results showed more
variability. Initialization methods like 2, 3, 4, and 5 produced unstable results, characterized by
high standard deviation. However, the 2 and 3 were the most stable, showing the lowest standard
deviations across these functions. The 1 also demonstrated competitive performance, closely matching
the results of 0. At  = 20, the results followed the same general pattern, though the average values
and standard deviations increased slightly, with 3 showing the best performance across three of the
four functions, while 5 remained inconsistent. For the hybrid functions (6 to 8), the majority
of methods displayed similar performance, but 2 was particularly unstable and performed poorly
across both dimensions. At  = 20, methods like 2, 3, 4, 5, and 6 delivered the worst results,
especially on function F6. Meanwhile, 2, 3, and 1 were more stable and provided better results. 1
and 3 remained the top-performing methods, confirming their efectiveness in hybrid functions. These
ifndings indicate that for hybrid functions, the careful selection of an initialization method is crucial.
The probability distributions, with the exception of 1, were generally inefective for this category
of functions. In contrast, low-discrepancy sequences, consistently performed well, suggesting that
they are better suited for hybrid optimization problems. In the composition functions (9 to 12), 2
consistently outperformed other methods at both  = 10 and  = 20. This performance highlights
the efectiveness of low-discrepancy sequences in eficiently navigating complex search spaces, which
is essential for composition functions characterized by their intricate landscapes and multiple optima.</p>
      <p>The Friedman test (last columns of Tables 3 and 4), used to rank the initialization methods across
all functions, further highlighted the statistical superiority of 2 and 1 in both dimensions. These
methods consistently demonstrated stability and eficiency, yielding optimal results with low standard
deviations. 1 also outperformed 0 at  = 10, while 3 showed better performance than 0 at
 = 20, though it did not perform as well at  = 10. Most initialization methods outperformed 0 at
 = 20, with 2, 3, and 1 being the most efective. At  = 10, only three methods outperformed
0, with 2, 1, and 1 leading the way.</p>
      <p>Finally, the results from the Friedman test rankings of execution times for diferent initialization
methods across the dimensions  = 10 and  = 20, as presented in Table 5, highlight significant
diferences in computational eficiency. At</p>
      <p>= 10, methods 0 and 1 ranked among the slowest with
scores of 6.5000, indicating high computational overhead, while methods 3 and 4 exhibited relatively
better execution times, although they still performed poorly overall. In contrast, low-discrepancy
sequence methods such as 1, 2, and 3 displayed more balanced execution times, with 2 achieving a
score of 5.0833. At  = 20, the eficiency landscape shifted, with
1 ranking the lowest at 6.8333, while
2 and 0 also exhibited poor performance. Notably, methods 4, 5, and 6 provided the shortest
execution times but did not perform well on this dimension. Overall, low-discrepancy sequences
consistently outperformed probabilistic methods in execution time, reinforcing their suitability for
optimization tasks where both solution quality and computational eficiency are critical.</p>
    </sec>
    <sec id="sec-5">
      <title>5. Conclusion and Perspectives</title>
      <p>The research presented in this paper contributes to the growing body of knowledge on initialization
schemes for MAs, an area that has yet to be fully explored in the optimization literature. While
numerous MAs have been developed and refined over the years, much less attention has been given
to the initialization phase, which significantly impacts the overall performance of these algorithms.
Most researchers continue to rely on traditional random number generators for initialization, despite
well-documented limitations. This study confirms the importance of exploring alternative initialization
methods.</p>
      <p>Our findings demonstrated the efectiveness of various initialization schemes, including probabilistic
distributions and low-discrepancy sequences, in influencing the performance of the AHO algorithm.
Specifically, the experimental results on the CEC 2022 benchmark functions revealed that AHO’s
sensitivity to initialization schemes varies depending on the problem at hand. Some functions were
more responsive to the choice of initialization, while others were relatively insensitive. Notably, the
results showed that low-discrepancy sequences were particularly efective and exhibited greater stability,
especially in higher dimensions. They consistently achieved the best results across most functions,
underscoring their suitability as a robust initialization strategy, particularly in more complex and
higher-dimensional optimization problems. This suggested that problem-specific characteristics, along
with dimensionality, play a key role in determining the most appropriate initialization strategy.</p>
      <p>Moving forward, we intend to conduct an in-depth study on the parameter initialization of MAs more
broadly. In particular, we aim to explore innovative methods for automatic parameter initialization,
which could further improve the adaptability and performance of these algorithms across a wide
range of optimization problems. By incorporating novel approaches, such as hybrid techniques and
dynamic parameter tuning, we seek to develop frameworks that allow metaheuristics to adjust their
key parameters in response to the problem landscape. Additionally, we will focus on the exploration of
hyper-parameter optimization strategies to systematically identify the most suitable configurations for
enhancing both convergence speed and solution quality.</p>
    </sec>
    <sec id="sec-6">
      <title>Declaration on Generative AI</title>
      <p>The authors have not employed any Generative AI tools.
sequences: the comparative results, in: 2007 IEEE congress on evolutionary computation, IEEE,
2007, pp. 1985–1992.
[16] A. H. Gandomi, X.-S. Yang, S. Talatahari, A. H. Alavi, Firefly algorithm with chaos, Communications
in Nonlinear Science and Numerical Simulation 18 (2013) 89–98.
[17] M. Lozano, A. Duarte, F. Gortázar, R. Martí, A hybrid metaheuristic for the cyclic antibandwidth
problem, Knowledge-Based Systems 54 (2013) 103–113.
[18] F. Zitouni, S. Harous, A. Belkeram, L. E. B. Hammou, The archerfish hunting optimizer: A novel
metaheuristic algorithm for global optimization, Arabian Journal for Science and Engineering 47
(2022) 2513–2553.
[19] R. Brits, A. P. Engelbrecht, F. van den Bergh, A niching particle swarm optimizer, in: Proceedings of
the 4th Asia-Pacific conference on simulated evolution and learning, volume 2, 2002, pp. 692–696.
[20] R. Storn, K. Price, Diferential evolution–a simple and eficient heuristic for global optimization
over continuous spaces, Journal of global optimization 11 (1997) 341–359.
[21] Y. Shi, R. Eberhart, A modified particle swarm optimizer, in: 1998 IEEE international conference
on evolutionary computation proceedings. IEEE world congress on computational intelligence
(Cat. No. 98TH8360), IEEE, 1998, pp. 69–73.
[22] X.-S. Yang, S. Deb, Cuckoo search via lévy flights, in: 2009 World congress on nature &amp; biologically
inspired computing (NaBIC), Ieee, 2009, pp. 210–214.
[23] D. Karaboga, B. Basturk, Artificial bee colony (abc) optimization algorithm for solving constrained
optimization problems, in: International fuzzy systems association world congress, Springer, 2007,
pp. 789–798.
[24] J. O. Agushaka, A. E. Ezugwu, L. Abualigah, S. K. Alharbi, H. A. E.-W. Khalifa, Eficient initialization
methods for population-based metaheuristic algorithms: A comparative study, Archives of
Computational Methods in Engineering 30 (2023) 1727–1787.
[25] X.-S. Yang, Nature-inspired metaheuristic algorithms, Luniver press, 2010.
[26] S. Mirjalili, S. M. Mirjalili, A. Lewis, Grey wolf optimizer, Advances in engineering software 69
(2014) 46–61.
[27] S. Mirjalili, A. Lewis, The whale optimization algorithm, Advances in engineering software 95
(2016) 51–67.
[28] A. P. Engelbrecht, Fundamentals of computational swarm intelligence, John Wiley &amp; Sons, Inc.,
2006.
[29] J. O. Agushaka, A. E. Ezugwu, Initialisation approaches for population-based metaheuristic
algorithms: a comprehensive review, Applied Sciences 12 (2022) 896.
[30] S. Arora, S. Singh, Butterfly optimization algorithm: a novel approach for global optimization,</p>
      <p>Soft computing 23 (2019) 715–734.
[31] X. Wang, F. J. Hickernell, Randomized halton sequences, Mathematical and Computer Modelling
32 (2000) 887–899.
[32] E. I. Atanassov, Eficient cpu-specific algorithm for generating the generalized faure sequences, in:</p>
      <p>International Conference on Large-Scale Scientific Computing, Springer, 2003, pp. 121–127.
[33] P. Bratley, B. L. Fox, Algorithm 659: Implementing sobol’s quasirandom sequence generator, ACM</p>
      <p>Transactions on Mathematical Software (TOMS) 14 (1988) 88–100.
[34] J. E. Gentle, Random number generation and Monte Carlo methods, volume 381, Springer, 2003.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>S. J.</given-names>
            <surname>Russell</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Norvig</surname>
          </string-name>
          ,
          <article-title>Artificial intelligence: a modern approach</article-title>
          , Pearson,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>C. H.</given-names>
            <surname>Papadimitriou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Steiglitz</surname>
          </string-name>
          ,
          <article-title>Combinatorial optimization: algorithms and complexity</article-title>
          ,
          <source>Courier Corporation</source>
          ,
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>C. M.</given-names>
            <surname>Bishop</surname>
          </string-name>
          ,
          <source>Pattern recognition and machine learning</source>
          ,
          <source>Springer google schola 2</source>
          (
          <year>2006</year>
          )
          <fpage>1122</fpage>
          -
          <lpage>1128</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>F. S.</given-names>
            <surname>Hillier</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G. J.</given-names>
            <surname>Lieberman</surname>
          </string-name>
          , Introduction to operations research,
          <source>McGraw-Hill</source>
          ,
          <year>2015</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>D.</given-names>
            <surname>Bertsimas</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. N.</given-names>
            <surname>Tsitsiklis</surname>
          </string-name>
          , Introduction to linear optimization, volume
          <volume>6</volume>
          ,
          <string-name>
            <given-names>Athena</given-names>
            <surname>Scientific</surname>
          </string-name>
          <string-name>
            <surname>Belmont</surname>
          </string-name>
          , MA,
          <year>1997</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>S. J.</given-names>
            <surname>Wright</surname>
          </string-name>
          , Numerical optimization,
          <year>2006</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>S.</given-names>
            <surname>Boyd</surname>
          </string-name>
          , L. Vandenberghe, Convex optimization, Cambridge university press,
          <year>2004</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>S. J.</given-names>
            <surname>Wright</surname>
          </string-name>
          ,
          <article-title>Primal-dual interior-point methods</article-title>
          ,
          <source>SIAM</source>
          ,
          <year>1997</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <surname>J. H. Holland,</surname>
          </string-name>
          <article-title>Adaptation in natural and artificial systems: an introductory analysis with applications to biology, control, and artificial intelligence</article-title>
          , MIT press,
          <year>1992</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>J.</given-names>
            <surname>Kennedy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Eberhart</surname>
          </string-name>
          ,
          <article-title>Particle swarm optimization</article-title>
          ,
          <source>in: Proceedings of ICNN'95-international conference on neural networks</source>
          , volume
          <volume>4</volume>
          , ieee,
          <year>1995</year>
          , pp.
          <fpage>1942</fpage>
          -
          <lpage>1948</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>A.</given-names>
            <surname>Colorni</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Dorigo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Maniezzo</surname>
          </string-name>
          , et al.,
          <article-title>Distributed optimization by ant colonies</article-title>
          ,
          <source>in: Proceedings of the first European conference on artificial life</source>
          , volume
          <volume>142</volume>
          , Paris, France,
          <year>1991</year>
          , pp.
          <fpage>134</fpage>
          -
          <lpage>142</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>F.</given-names>
            <surname>Zitouni</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Harous</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Maamri</surname>
          </string-name>
          ,
          <article-title>The solar system algorithm: a novel metaheuristic method for global optimization</article-title>
          ,
          <source>IEEE Access 9</source>
          (
          <year>2020</year>
          )
          <fpage>4542</fpage>
          -
          <lpage>4565</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>E.</given-names>
            <surname>Cantú-Paz</surname>
          </string-name>
          ,
          <article-title>On random numbers and the performance of genetic algorithms</article-title>
          ,
          <source>Science Direct Working Paper No S1574-034X (04)</source>
          (
          <year>2002</year>
          )
          <fpage>70205</fpage>
          -
          <lpage>7</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>Q.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.-Y.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.-S.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <article-title>Influence of initialization on the performance of metaheuristic optimizers</article-title>
          ,
          <source>Applied Soft Computing</source>
          <volume>91</volume>
          (
          <year>2020</year>
          )
          <fpage>106193</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>N. Q.</given-names>
            <surname>Uy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N. X.</given-names>
            <surname>Hoai</surname>
          </string-name>
          ,
          <string-name>
            <surname>R. I. McKay</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. M.</given-names>
            <surname>Tuan</surname>
          </string-name>
          ,
          <article-title>Initialising pso with randomised low-discrepancy</article-title>
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>