<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Efficiency Improvement of the Algorithm Based on an Artificial Immune System Modeling Applied to Continuous and Combinatorial Problems</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Timur Zheldak</string-name>
          <email>zheldak.t.a@nmu.one</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Illia Ziborov</string-name>
          <email>Ziborov.Il.K@nmu.one</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Vladyslav Lyman</string-name>
          <email>Lyman.V.I@nmu.one</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Andrii Zhuk</string-name>
          <email>Zhuk.A.V@nmu.one</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Dnipro University of Technology</institution>
          ,
          <addr-line>av. Dmytra Yavornytskoho, 19, Dnipro, 49005</addr-line>
          ,
          <country country="UA">Ukraine</country>
        </aff>
      </contrib-group>
      <fpage>82</fpage>
      <lpage>95</lpage>
      <abstract>
        <p>One of the heuristic optimization methods that implements the simulation of an artificial immune system is discussed in this article. The list of optimization problems, their classification and known approaches of solving those ones are reviewed on the whole. The research of the main settings of operators in the proposed algorithm is applied to the different dimensions test problems. For the first time, an attempt of finding a universal approach of solving both combinatorial and continuous problems, applying the proposed algorithm, was implemented based on the common approach to the algorithm operators. The most efficient settings, based on the minimum solution search time criteria, are recommended for agent generation search, parental individuals selection, mutation, crossover, local search and population compression operators. Artificial immune systems, AIS, algorithm, optimization, heuristic method, mutation, Saaty</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>selection, combinatorial problems</p>
    </sec>
    <sec id="sec-2">
      <title>1. Introduction</title>
      <p>Optimization problem solving in terms of the multicomponent industrial production is considered
to be a daily practice. Decision makers are to incorporate several criterias, which are generally ranged
by their significance, and to solve the problems in the
multidimensional space. One of the
mechanisms that allow solving this kind of complex mathematical problems is evolutionary
decisionmaking technologies.</p>
      <p>A feature of evolutionary technologies is the heuristic (usually not due to mathematical
establishment) nature of the decision-making algorithm, which can be improved, using new, more
effective in practice heuristics. Many of them belong to analogical methods - a tool of applied systems
analysis, which is used for study and simulation of complex systems analogously to biological or
inorganic nature.</p>
      <p>
        One of described heuristics is an artificial immune systems simulation algorithm, which allows
finding a solution of the conditional or unconditional optimization problem, using a combination of
evolutionary operators that simulate the human immune system. A detailed analysis of artificial
immune system simulation algorithms applied to optimization problems solving is given in [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] and
[
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. At the moment, algorithms following the principles of artificial immune systems simulation are
actively used to solve applied problems of technological control and optimization processes in
industry. The purpose of this study is to improve the efficiency of the optimization algorithm based on
artificial immune system simulation by increasing the efficiency of its individual operators: agent
generation, parental individuals selection,
mutation, crossover, local search and
population
compression.
      </p>
      <p>2021 Copyright for this paper by its authors.</p>
    </sec>
    <sec id="sec-3">
      <title>2. The current state of the research subject 2.1.</title>
    </sec>
    <sec id="sec-4">
      <title>Algorithm of modelling artificial immune systems</title>
      <p>
        Modeling artificial immune system (AIS) is a method of computable intelligence that uses the
analogy with one of the natural systems, namely the human defense system. As artificial neural
networks, artificial immune systems can learn new information, recall previously learned information,
and perform pattern recognition in a highly decentralized manner. In [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] author provides an overview
of the immune system approach from the computational viewpoint.
      </p>
      <p>
        In research [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ], authors classified, compare, and generalize the general nowadays techniques and
algorithms based on an idea of modeling AIS. Also considered theoretical aspects, methods, and
algorithms three of the main, and total techniques of AIS, namely algorithm of negative choice, the
algorithm of an immune network, and algorithm of clonal selection. The primary focus for analysis of
their similarities and differences, data types, learning methods, separate operators that using, and the
sphere of implementation of AIS.
      </p>
      <p>
        Authors [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] say that the emergence of nature-inspired algorithms (NIA) is a significant milestone
in the computational intelligence community. As one of the NIAs, the AIS shows its effectiveness,
implicit parallelism, flexibility, and applicability when solving various engineering problems.
However, the authors notice that AIS also stays in problem with evolution premature, local
localization of minima, and at slow convergence owing to stochastic search dynamics.
      </p>
      <p>
        Many efforts have been made by various researchers to increase the search effectiveness of AIS on
the different sides, namely: helping population diversity, adaptive parameters control, etc. Authors [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]
concentrate on the most effective evolution operators - mutation operator using "multi-learning". With
him depending on the result, giving from previous steps, for the heirs in every generation perhaps
using differentiated by force and principles mutation. In this way, authors trying to improve the
working dynamics of an algorithm, which, when trying to solve quite a big problem, quickly found a
suboptimal solution but need a long time to clarify.
      </p>
      <p>
        Instead, at work [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ], suggested break problem with large dimension on a few small. The latter is
significantly improves the working time of AIS when solving a combinatorial problem. For a problem
where variables are long binary vectors, the authors proposed to break him on a few small vectors,
and perform a local search of bits, that more than others affects the general solution. Then global
solutions search as a combination of local influences signs. The proposed break large dimension of
search on “subset” considerably minimizes the time of calculation of fitness function and the total
time of the working algorithm. As the authors say, experimental results on an extensive diversity of
synthetic datasets and UCI show that our proposed method achieves better performance as a modern
global algorithm for solving the combinatory problem.
      </p>
      <p>
        Previously with AIS, namely using an approach of clonal selection, resolve a combinatory problem
of optimal choice a carrier of the plurality outdoor advertising offers, available in the formulation of
the knapsack problem [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. The proposed realization of the algorithm has been investigated on test
plural with previously knowing global optimum for a various number of a potential solution. Also on
research was shown, that the AIS method instead of classical genetic algorithm (GA) reliably finds
the exact solution for a problem where dimension less than 1000 and find the same solution quickly in
absolutely time and more quality investigate area next to the optimum. It is noted that modeling
methods of an artificial immune system may be used to solve a large range of combinatorial search
and constrained optimization problems.
      </p>
      <p>
        Also, an algorithm founded on clonal selection was used in [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] for solving some tasks in real
space. Noting random character of mutation, that present in AIS based on clonal selection main search
operators, authors propose hybrid algorithm clonal selection learning. There are two mechanisms
(Baldwinian learning and orthogonal learning) that have opposite directions. Baldwinian learning is
used for investigation (global search), orthogonal - for investigating areas of current decision (local
improvement). At the same time, computational complexity increases slightly (global optimum found
slowly), but increase reliability (percentage of launches, where randomly find global optimum).
      </p>
      <p>
        Another approach also based on learning while working propose authors [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ], that named them
algorithm “Two-Population Co-evolutionary Algorithm with Dynamic Learning Strategy for
ManyObjective Optimization”. This is an AIS algorithm used only as a common search pattern, where
settings change in the process according to achieve the result.
      </p>
      <p>
        Similar principles use one recent research [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ], where researched self-study another famous
heuristic method — genetic algorithm (GA). In research shows that change characters of mutation and
use local purposeful search can give more productivity on some testing problems. In research, authors
prove that the best operators for solving problems in multidimensional real space it is rank selection,
discretion crossover, uneven mutation, and operator of local search that uses Newtonian approach
(based on first- and second-order derivatives).
      </p>
      <p>
        The idea of involvement local search to the modeling artificial immune system algorithm based on
the clonal selection use one of author's researches [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]. Authors propose, by analogy from island GA,
to break the population of cells on a subset, evolution in which occurs independently. After some
periods, individuals can exchange genetic information with individuals from other generations. This
gain opportunity investigates all areas of search. As in various versions of AIS, in [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] propose to save
previously find solutions for future generation using them new individuals, when in some time search
lost effectuality.
      </p>
      <p>Summing up, mark the following features of modern algorithms that use AIS for modeling for
solving optimization problems:
 applying different selection operators significantly improves search speed and effectivity;
 a new generation must create partially using crossover, not through only cloning;
 must use different mutation operators, and radius (size of surroundings) should be determined
during the working of the algorithm (learn);
 not enough attention to compressed population operators, that determine the diversity of the
new generation.
2.2.</p>
    </sec>
    <sec id="sec-5">
      <title>The HINO-SF algorithm</title>
      <p>
        Majority mentioned ideas is incorporated in the Hybrid Immune Network Optimization algorithm
with Saaty selection and Fibonacci search [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ], named as HINO-SF by authors. It consists of the
following steps:
1. The population of antibodies is generated randomly. A counter of generations is set to t=0.
2. If preassigned maximum number of generations is achieved then transition to step 11 should be
made; otherwise - transition to step 3.
      </p>
      <p>3. Depending on the cells' adaptability assessment of the current generation, the number of clones
is assigned for each cell of the current generation - it is the selection operator.</p>
      <p>4. Cloning of cells in an amount is determined by the selection operator.
5. Recombination of cloned individuals by using the operator of probability crossover.
6. Calculation the dynamic probability of mutation for clones and implementation of adaptive
mutation operator.</p>
      <p>7. Execution of the one-dimensional local search for each clone with a certain probability. The
method of golden ratio (Fibonacci) for a random coordinate is used in continuous problems and hill
climbing method is used in combinatorial problems instead – it is local search operator.</p>
      <p>8. Compression of the general population of parents and clones using the appropriate to a specified
level Np – it is compression operator. Memorizing the best of solution, if it has not occurred before.</p>
      <p>9. Destruction of cells, whose age has reached the set value   , from the current generation.
Their place in the main population is occupied by cells, randomly generated uniformly in the search
area.</p>
      <p>10. The counter of generations increments  =  + 1. Go to the step 2.
11. The output of the current generation as a set of suboptimal solutions.
12. Stopping the algorithm.</p>
      <p>Block diagram of the algorithm is shown in Figure 1. This algorithm research, its limitations, and
optimal configuration of its operators are dedicated to the study.
2.3.</p>
    </sec>
    <sec id="sec-6">
      <title>Problems solving with the AIS</title>
      <p>
        The foresaid Hybrid Immune Network Optimization algorithm was investigated both for solving
problems in an infinite space and for solving combinatorial kind of problems. The last one is
described in detail, following [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ]. It is the problem of orders execution planning, implementing by a
metallurgical enterprise, which typically belongs to the combinatorial optimization problem class of
the scheduling theory. At the same time, the order execution scheduling problem is characterized with
the features of another well-known combinatorial problem - the construction of a travel salesman's
route (Traveling salesman problem, TSP), particularly, open Loop Traveling salesman problem.
      </p>
      <p>
        The other one combinatorial applied optimization problem is the steel amount minimization
problem which provides the steel order size limit of the only one cast volume [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ]. The developed
model specifies the possibility pre-evaluate the billet optimal size, based on the necessary cutting
along the final product length, appropriate for the certain billet form of section, and ingot weight
limits at the top and bottom.
      </p>
      <p>
        In addition to reviewed applied problems, several test sets of various sizes and complexity were
used, which describe the knapsack problem, the minimum vertex cover problem, and the set partition
problem. All the problems are formulated according to the classic statements. All while in both case
of problem [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] and [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] the transition from constrained optimization to an unconstrained problem,
which is solved by the algorithm formulated in Section 1.2, was implemented, using the penalty
functions method. Basically, solutions, which were not satisfied with the constraints, were acceptable,
but worse than those, which were satisfied with the same ones.
      </p>
      <p>
        The test problems of Rastrigin, Ackley, Rosenbrock, and others considered in [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] were used in an
infinite space - a total of 30 test functions. Each of functions was tested in two-, ten-, twenty five- and
one hundred- dimensional space. Under these circumstances the authors evaluated both the accuracy
of the obtained solution (deviation from the known global optimum) for the allotted time, and the time
spent on achieving acceptable accuracy of the calculation.
      </p>
      <p>An additional factor that was studied in solving both combinatorial and continuous problems was
the number of calls to the objective function. This factor is significant in case of the machine costs,
especially in solving combinatorial problems, where the calculation of the objective function is based
on matrix multiplication.</p>
    </sec>
    <sec id="sec-7">
      <title>3. The study content</title>
      <p>This issue is devoted to solving a number of applied and test problems using the algorithm
described. The settings of algorithm operators that provide better convergence and search speed for
the known optimal solutions are found.
3.1.</p>
    </sec>
    <sec id="sec-8">
      <title>Uniformity of new search agents generation</title>
      <p>In any population-based global optimization algorithm like partial swarm optimization (PSO),
evolution strategy (ES), GA, or AIS and among others a pseudo-random number generator is used to
sample the first   individuals of the initial generation from a uniform probability distribution
 (0; 1). If its generation power N is smaller than 103 – 104, it results in too nonuniform mesh.</p>
      <p>
        The same authors suggest using quasi-random numbers, which have better homogeneity than
pseudo-random numbers [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ]. Quasi-random numbers share many similarities with pseudo-random
numbers, but they are deterministically chosen based on low-discrepancy sequences.
      </p>
      <p>The Halton sequence and the Hammersley set define two deterministic ways of sampling a
Ddimensional space so that the consecutive points are as far apart as possible from each other. In
essence, the Halton sequence and the Hammersley set are both a generalization of the Van der Corput
sequence for multidimensional cases.</p>
      <p>
        Sobol sequences are another widely used quasi-random number generator, which was invented by
Ilya M. Sobol back in 1967. This quasi-random number generator uses a base of 2 to generate
 (0; 1), to then perform a special re-ordering of the master sequence for each one of the dimensions
of the sampled hyper-space [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ].
      </p>
      <p>The Figure 2 shows compared 1000 points in 2-dimension space from pseudo-random sample (top
left), from the Halton sequence (top right), from the Hammersley set (bottom left) and from the Sobol
sequence (bottom right). Quasi-random sequences cover the space more evenly than pseudo-random.</p>
      <p>
        We applied the described generators in the algorithm shown in Figure 3 to the set covering
problem scpcyc07 from OR-Library (selection the minimum number of subsets from the 672 available
to cover the set of 448 graph vertices. Each subset contains exactly 4 vertices) [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ] and Rastrigin
problem at multidimensional continuous space. At the same time built-in MATLAB system classical
generators were used for comparison. Results of comparison at first 10000 iterations are illustrated in
Table 1 and Figure 3.
      </p>
      <sec id="sec-8-1">
        <title>Generator</title>
      </sec>
      <sec id="sec-8-2">
        <title>Mersenne vortex (default)</title>
      </sec>
      <sec id="sec-8-3">
        <title>Combined recursive</title>
      </sec>
      <sec id="sec-8-4">
        <title>Multiplicative Fibonacci</title>
      </sec>
      <sec id="sec-8-5">
        <title>Halton sequence</title>
      </sec>
      <sec id="sec-8-6">
        <title>Hammersley set</title>
      </sec>
      <sec id="sec-8-7">
        <title>Sobol sequence</title>
        <p>Dimension of space</p>
        <p>25 65
10
136 1702
118 1476
86 1099
111 1375
103 1258
84 1005
10274
11320
9920
9811
8917
8497</p>
        <p>100
64987
103951
68357
61159
51478
66123</p>
        <p>Table 1 contains comparing same pseudo-random and quasi-random generators for Rastrigin
problem solving at 10, 25, 65 and 100 dimensions of continuous space. Instead, Figure 3 shows the
time dependence of the best solution of the sets covering problem for the first 10,000 iterations.</p>
        <p>All quasi-random generators show better result than pseudo-random generators. This result is also
proved by increasing the dimension of the problem when filling the solution space becomes an
important factor. Using of quasi-random generators accelerates on average the search for the global
optimum by 12-20% of the time.
3.2.</p>
      </sec>
    </sec>
    <sec id="sec-9">
      <title>Parental individuals selection</title>
      <p>
        One of the main factors that determine the efficiency of the optimization algorithm, based on the
artificial immune system, is the choice of parental individuals used to obtain clones. In [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] it is
proposed to use the selection operator based on Saaty’s pairwise comparisons. In this case, the matrix
of candidate pairwise comparisons for offspring was filled only for decisions that are included to the
first half of the segment [ 
;
      </p>
      <p>]. The rating of each of candidate j is calculated by</p>
      <p>= 1 + ⌊ (  ) ∙  ⌋,
where k - the dimension of the scale (default k = 9).
 1/  is used, and for the rest -</p>
      <p>=  1 / 1 .</p>
      <p>Construction of a matrix of pairwise comparisons for the first line of which an expression  1 =


of the current generation.</p>
      <p>
        The HINO-SF algorithm analysis with the described selector resulted in some limitations:
with a small population among potential parents there are 1 or 2 applicants, the solution of
which is better than the average value of the segment [ 
;  
];
among the clones, the total majority (or all) will be descendants of one of the representatives
In fact, the selection operator described in [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] deprives the multi-agent search AIS heuristic of its
one advantages - search by several agents in different parts of the search area simultaneously.
      </p>
      <p>
        A modified selector, based on the method of Saaty’s pairwise comparisons, also described in [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ],
was used to evade this problem. The matrix of pairwise comparisons is based on the following
principle:
      </p>
      <p>The calculation of the membership function is as follows:</p>
      <p>|  −   | + 1,   ≥ 


= {

1

 ( (  )) = √∏</p>
      <p>.
,   &lt; 


 =1
(1)
(2)
(3)</p>
      <p>
        The most wide-spread types of selectors, such as rank (ordinal) and tournament (roulette), were
used for comparison of several problems solving, as well. Competitive ones are proved to be highly
efficient among other evolutionary algorithms - GA [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ], in particular.
      </p>
      <p>The comparison of these selection operators is demonstrated in Table 2. In this table, the selection
operators are compared by two main factors: the number of iterations (generations) to the first global
optimum occurrence and the number of calls to the objective function. Both values are calculated on
average over 10 runs with different random initial values. Both combinatorial and continuous
problems, as well as different dimensions are considered.</p>
      <p>According to the results in Table 2, the following conclusions can be outlined:
• The proposed selection operator based on the Saaty’s scale pairwise comparisons method is
equally effective in both binary and continuous space;</p>
      <p>• In some problems (minimum graph tree, Ackley problem) the proposed selector shows a higher
efficiency than the tournament 1.5 - 3 times). The rest of the problems use the same time and access
to the target function.
3.3.</p>
    </sec>
    <sec id="sec-10">
      <title>Crossover applying</title>
      <p>The proposed AIS algorithm mainly differs from other analogues on the application of an adaptive
crossover operator to clones, which involves the exchange of genetic information. It is to diversify the
search for the best solution in the algorithm. Therefore, the more successfully it will perform, the
greater speed is to find the global optimum. The choice of crossover type depends on the space in
which it is applied to. The combinatorial problems were solved with approaches which differ from
other ones applied to the solutions of continuous problems. The most essential factors for evaluating
the solutions efficiency are the time spent searching for a solution and the value of the deviation of the
objective function from the known global optimum.</p>
    </sec>
    <sec id="sec-11">
      <title>3.3.1. Crossover choice for combinatorial problems</title>
      <p>Having used the selection, founded on modified Saaty operator, as well as the standard values of
the mutation operators (their settings are drawn below) the previously described binary crossovers are
compared. One-point, two-point and uniform crossovers were considered.</p>
      <p>A comparison of the solution determining speed (in the number of iterations) and the deviation
from the exact solution (in percent) is illustrated in Figure 4. The data in the figure is averaged over
10 runs with random initial populations.</p>
      <p>According to the results, the two-point crossover happens to be the best for all combinatorial
problems solutions. On the one hand, it collects the most detail about the parental genes (as well as
the single-point crossover), but also provides some diversification due to two breaks in the genome.
At the same time, offspring, produced with a uniform crossover, significantly differs from each of the
parents for considered problems. The solutions are discarded in the new search area and in most cases
are much worse than the parental individuals, so they are no longer used.</p>
      <p>Particular attention was paid to the probability   , which the crossover is applied to the newly
formed clone with. In a wide range from 0 to 0.9 (sometimes up to 0.95), higher   probability
values resulted in a faster search for the global optimum regardless of problem type and its dimension.
At the same time, while   = 1, the search efficiency drops rapidly. Therefore, the crossover
operator is recommended to use with probability differs from 1.</p>
    </sec>
    <sec id="sec-12">
      <title>3.3.2. Crossover choice for the continuous problems</title>
      <p>
        Having applied the selection founded on the modified Saaty operator, the generation value   =
5, the number of offsprings   = 50, the adaptive mutation operator described in [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] with the degree
of mutation  = 0,1, crossovers for continuous space are compared. Such crossovers as
direct, uniform, linear, arithmetic, heuristic, Laplace and SBX were tested.
      </p>
      <p>A comparison of the solution determining speed (in the number of iterations) and the deviation
from the exact solution is illustrated in Figure 5. The data in the figure is averaged over 10 runs with
random initial populations.</p>
      <p>
        Analysis of the results in Figure 5 proves that the Laplace and heuristic crossovers perform 2-2.5
times more iterations, while not providing a sufficient quality result on different problems of various
dimensions. The SBX crossover, which is used in the basic algorithm [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ], and which is most often
used in continuous space, is conventionally effective. It almost always provides a transition to the
global optimum and requires a minimum of iterations for its achieving.
      </p>
      <p>
        Meanwhile, a uniform crossover, which selects the parental genes with equal probability, resulted
in higher values of accuracy on large-scale problems. Similar results were obtained earlier in the study
of optimal settings of genetic algorithms [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]. It is further proposed to use a uniform crossover to solve
problems in multidimensional continuous space, as this operator is an order of magnitude faster than a
complex SBX crossover.
3.4.
      </p>
    </sec>
    <sec id="sec-13">
      <title>Mutation width and type</title>
      <p>The main operator responsible for finding a solution in the proposed algorithm, as in most
implementations of the method of artificial immune systems, is the mutation operator. An adaptive
operator, which randomly applies either a narrow
mutation to search in the current solution
neighborhood, or a broad mutation to change the search area, is proposed. As in the case of the
crossover operator, the mutation operator form significantly depends on the space, where the problem
is solved. The choice of mutation range performance (wide or narrow) is based on the usefulness
assessment of the current cell on the utility assessments scale on the current iteration throughout the
population. The scale is determined for each j-th cell individually with the formula


 = (1 −</p>
      <p>)</p>
      <p>−  
−  
,
 = [1:   ],
(4)
where  − clone number in the population of clones;   – clone population size;  − current iteration
number;  
− provided maximum number of iterations;  
− the worst value of the objective
function in the generation;</p>
      <p>− the best value of the objective function in the generation;   −
the value of the objective function for the current clone;  − the maximum allowable proportion of
broad mutations in the total.</p>
      <p>If a random  (0; 1) is less than 
transition to a new search area.
clone j (search in the solution neighborhood). Otherwise - a broad mutation is performed - the</p>
      <p>
        For continuous space, it is proposed to use a combination of a broad Gaussian mutation and a
narrow polynomial distribution mutation. Both are described in detail in [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ]. It was found that the
value of γ, included in expression (4), should be low values of 0.1… 0.25 for the low dimension of the
problem and increase up to 0.8… 0.9 at high dimension for continuous problems. It is explained with
the fact that in a small dimension space, the main factor is to consider the neighborhood of the best
current solutions. At 60 or more measurements, exactly the broad mutation demonstrates a significant
impact on the speed of global optimum finding.
      </p>
      <p>following (4), a narrow mutation is performed for</p>
      <p>Having analyzed formula (4), it is easy to notice that the best representatives in the generation at
the initial stage of the algorithm are prone to narrow search mutations, at the final - to wide. Instead,
the worst members of the generation almost always perform only broad mutations.
The dependence of the 
 characteristic against the current iteration number, the clone
adaptability degree and the admissible maximum proportion of wide mutations is illustrated by the
graph in Figure 6.</p>
      <p>Analyzing the obtained dependences, it is noticeable that in case of low γ values the worst clones
in the population are almost always make a wide mutation, and successful clones make a narrow
mutation. Increasing the dimension, the mutation nature choice is changing significantly. In
multidimensional continuous space in case γ = 0.9 at the beginning of the calculation, the clones
behave similarly to that described above. As we approach the end of the calculation, the proportion of
narrow mutations decreases even for the best of the clones by 10 times.</p>
      <p>When solving combinatorial problems, a point mutation of a certain number of genes is performed
as a narrow mutation. The number of inverted genes is determined with the formula

( ) = {


(1 +   0



−
,
2 ∙   0


) ,
  &lt;  
⁄2
,</p>
      <p>
        (5)
  ≥  
⁄2
bits);   0
recommended   0
where t - number of current iteration; 
– the minimum level of mutation (usually 1 or 2
– the prior probability of mutation operator using (global parameter of algorithm, in [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ]
= 0,5 … 1) and
      </p>
      <p>– maximum number of generations.</p>
      <p>Then, regardless of the time of algorithm on the initial iterations mutation will be per-formed for
all clones, on final - less than for a half of them.</p>
      <p>Extensive mutation in combinatorial problems can be realized through saltation. It is the process of
the exchange between a certain number of bits, which are located at a certain distance from each other
in the genome. Saltation mutation is effective due to its saving the total number of 0 and 1 in the
genotype, thus not rejecting the current solution too far in the search space.</p>
      <p>The mutation radius adapting issue in continuous space remains to stay opened. For today, the
authors do not manage to substantiate any pattern. The best solution time and accuracy are achieved
with different settings for various problems and multidimensional space.
3.5.</p>
    </sec>
    <sec id="sec-14">
      <title>Whether a local search is required</title>
      <p>
        One-dimensional local search, proposed in [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ], was used in the study of the algorithm to a limited
extent. The matter is that the Fibonacci local search method calls the objective function dozens of
times to find a solution in only one coordinate in the above one-dimensional notation. Therewith, this
search method is not goal-oriented and it can be applied only to problems in continuous space.
Analyzing the results of the previous studies [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ], quasi-Newtonian methods which required to the
using of first- and second-order derivatives are more effective in comparison with zero-order methods
(Fibonacci, Nelder-Mead, Powell, etc.). In this case, local search operator should not be applied
unconditionally to all offsprings, but it must be used with some probability only when traditional
crossover and mutation operators are ineffective for a certain number of generations.
      </p>
      <p>The topic of further research should be such a goal-oriented local search operator determination
that would improve the objective function of a random descendant without spending significant
computation time.</p>
      <p>
        At the same time, the HINO-SF algorithm works effectively without the local search applying for
combinatorial problems solving. The limited local search based on the hill climbing method is
proposed earlier in [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. It reduces the search time for a solution in the number of the algorithm
iterations. However, the number of calls to the objective function increases in comparison with the
option without local search. As a result, the astronomical time to obtain a solution, using local search,
is several times longer than without its applying.
3.6.
      </p>
    </sec>
    <sec id="sec-15">
      <title>Population compression</title>
      <p>
        The population compression operator determines the success of finding solutions in the paradigm
of artificial immune systems. It prevents premature convergence and coming to a halt in local
optimums, which is inherent in all evolutionary algorithms. In most sources [
        <xref ref-type="bibr" rid="ref10 ref7 ref9">7, 9, 10</xref>
        ], the so-called
similarity radius of individuals, which determines the minimum distance in the solution space
between individuals that are passed from generation to generation, is determined empirically. Most
authors define it as a constant, or as a quantity that decreases evenly.
      </p>
      <p>It is proposed to link the similarity radius with the offspring mutation radius in the current
generation. Thus, the mutation operator will operate in the current solution neighborhood, and the
population compression operator - outside it. This ensures diversity in each of the generations. The
main operator, which is responsible for the search intensification, as in most implementations of the
artificial immune systems method, continue to be the mutation operator in any case. At the same time,
the population compression operator assumes the role of diversifier.</p>
      <p>Solving continuous problems, it was found that the optimal value of the compression operator
radius is getting greater meanwhile space dimension is increasing. Optimality should be understood as
the speed of global optimum achieving while maintaining population diversity. The dependence of the
compression radius on the space dimension in solving the Rastrigin problem is shown in Figure 7.</p>
      <p>Also, in Figure 7 it is clearly noticeably the best of several dozen regression models that describe
the dependence of the compression operator radius on the space dimension. The power model with a
power coefficient close to ½ is obtained. Therefore, the desired dependence can be assumed to have
the form
 ( 
) =  0 2√ 
,
(6)
where   –problem dimension in continuous space ;  0 – a constant term, depending on the
problem type and must be obeyed empirically.</p>
      <p>In addition to that, the tendency illustrated in Figure 6 was not refuted, but also received an extra
development. In particular, scaling the knapsack problem and using different numbers of individuals
in the population, it was found that the best results in terms of convergence and repeatability are
provided by the use as a coefficient in (6) the value</p>
      <p>0 = 1⁄  (7)</p>
      <p>Thus, it is further recommended to set the radius in the population compression operator
proportional to the radical of the space dimension and inversely proportional to the population
individuals number.</p>
    </sec>
    <sec id="sec-16">
      <title>4. Discussion</title>
      <p>Based on the results of numerous experiments in solving known global optimization problems, the
developed HINO-SF algorithm was investigated for its individual operators effectiveness. The
following results were obtained:
• Applying of quasi-random generators instead of pseudo-random ones for first and new
individuals generation during the algorithm reduces the global optimum searching time by 12-20%.</p>
      <p>• Adopting of the modified selection operator, based on the Saaty`s paired comparison method,
allows to improve the solution finding dynamics by 1.5-3 times, comparing with the previous
realization and to achieve greater reliability than in case of the rank or tournament selection operators
using.</p>
      <p>• The global optimum finding becomes more reliable applying the crossover operator. The
twopoint crossover is the best for combinatorial problem, uniform crossover – for continuous problems
relatively.</p>
      <p>• The proposed adaptive mutation operator adjusts the narrow and wide mutations ratio, depending
on the clone fitness and the iteration number. As soon as the space dimensionality increases, the
proportion of wide mutations should be increased.</p>
      <p>• The existing one-dimensional local search operator is considered to be ineffective and should be
replaced. Applying one of the quasi-Newtonian methods instead of the zero-order methods might be
promising.</p>
      <p>• The population compression operator should be adaptive, considering the mutation radius at the
current iteration, population size and space dimension.</p>
      <p>
        The work [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] rejects the significance of high-dimensional combinatorial problems solving,
adopting heuristic algorithms.
      </p>
      <p>The search for the global optimum and several local optimal solutions close to global one, which
can be used as an effective alternative, is an essential scientific and technical problem. Its solution,
using a developed and investigated optimization algorithm based on an artificial immune system
modeling might make specific social and economic impact.</p>
    </sec>
    <sec id="sec-17">
      <title>5. Conclusions</title>
      <p>The article includes the investigation of the global optimization algorithm based on artificial
immune systems modeling and the most effective settings of its individual operators. For the first
time, an attempt of finding a universal approach to solving both combinatorial and continuous
problems applying the proposed algorithm was implemented, using a common approach to the
algorithm operators.</p>
      <p>The most efficient settings, based on the minimum solution search time criteria, are recommended
for search agent generation, parental individuals selection, mutation, crossover, local search and
population compression operators.</p>
      <p>A significant reduction in the number of objective function evaluations and the algorithm
performance time was achieved. It allowed the proposed algorithm to become competitive towards
other widely used global optimization methods. The proposed settings of the algorithm operators
allow to reduce the optimization time by 2-4 times without losing accuracy for some solving
problems.</p>
      <p>The set of several test functions of various dimensions in continuous and binary spaces was used
for the algorithm testing, which demonstrated in improved time criteria performance. It is reasonable
to outline that the greatest benefits from algorithm settings usage can be obtained in large dimension
problems.</p>
      <p>The practical value of the work is to make possible the developed algorithm applying for global
optimization problems solving, especially in case of a high dimension mathematical model of a
problem. In particular, the approach can be applied to solve planning and control problems in
metallurgical production to optimize technological processes.</p>
      <p>The direction of further research is the best local search operator selection as well as the adaptation
mechanism formation of the mutation operator as the main search element of the AIS optimization
algorithm. Other essential research direction is the formation of an adaptive population compression
operator, which would help to discover suboptimal solutions.</p>
    </sec>
    <sec id="sec-18">
      <title>6. Acknowledgements</title>
      <p>This publication was prepared as a part of the scientific theme 0121U109788 "Problems of
modeling, optimization, and decision making in complex systems of different nature", which is
implemented on the System Analysis and Control Department at Dnipro University of Technology.</p>
    </sec>
    <sec id="sec-19">
      <title>7. References</title>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>D.</given-names>
            <surname>Dasgupta</surname>
          </string-name>
          , (
          <year>1998</year>
          ).
          <source>Artificial Immune Systems and Their Applications</source>
          .
          <source>DOI:10.1007/978-3- 642-59901-9.</source>
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>R.M.</given-names>
            <surname>Rasli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.A.A.</given-names>
            <surname>Aziz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.M.</given-names>
            <surname>Razali</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.M.</given-names>
            <surname>Norwawi</surname>
          </string-name>
          &amp; N. Basir, “
          <article-title>A Preliminary Survey on Artificial Immune Systems (AIS)”</article-title>
          .
          <source>International Journal of Academic Research in Business and Social Sciences. V. 9 I. 14</source>
          (
          <year>2019</year>
          ):
          <fpage>121</fpage>
          -
          <lpage>144</lpage>
          . DOI:
          <volume>10</volume>
          .6007/IJARBSS/v9-i14/
          <fpage>6835</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <surname>Sh</surname>
          </string-name>
          .
          <string-name>
            <surname>Gao</surname>
          </string-name>
          “
          <article-title>A Multi-Learning Immune Algorithm for Numerical Optimization”</article-title>
          .
          <source>IEICE Transactions on Fundamentals of Electronics Communications and Computer Sciences. E98-A</source>
          (
          <year>2015</year>
          ):
          <fpage>362</fpage>
          -
          <lpage>377</lpage>
          . DOI:
          <volume>10</volume>
          .1587/transfun.E98.A.
          <volume>362</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <surname>T.</surname>
          </string-name>
          <article-title>Li “Local feature selection based on artificial immune system for classification”</article-title>
          . Applied Soft Computing. V.
          <volume>87</volume>
          (
          <year>2020</year>
          ):
          <fpage>105989</fpage>
          , DOI: 10.1016/j.asoc.
          <year>2019</year>
          .
          <volume>105989</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>J.</given-names>
            <surname>Gao</surname>
          </string-name>
          , G. He,
          <string-name>
            <given-names>R.</given-names>
            <surname>Liang</surname>
          </string-name>
          &amp; Zh. Feng, “
          <article-title>A quantum-inspired artificial immune system for the multiobjective 0-1 knapsack problem”</article-title>
          .
          <source>Applied Mathematics and Computation</source>
          . V.
          <volume>230</volume>
          , (
          <year>2014</year>
          ):
          <fpage>120</fpage>
          -
          <lpage>137</lpage>
          , https://doi.org/10.1016/j.amc.
          <year>2013</year>
          .
          <volume>12</volume>
          .088
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Peng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.-L.</given-names>
            <surname>Lu</surname>
          </string-name>
          “
          <article-title>Hybrid learning clonal selection algorithm”</article-title>
          . Information
          <string-name>
            <surname>Sciences</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          <year>296</year>
          (
          <year>2015</year>
          ):
          <fpage>128</fpage>
          -
          <lpage>146</lpage>
          , DOI: 10.1016/j.ins.
          <year>2014</year>
          .
          <volume>10</volume>
          .056.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>W.</given-names>
            <surname>Li</surname>
          </string-name>
          , GG. Wang, &amp;
          <string-name>
            <surname>A.H. Gandomi</surname>
          </string-name>
          “
          <article-title>A Survey of Learning-Based Intelligent Optimization Algorithms”</article-title>
          .
          <source>Arch Computat Methods Eng</source>
          (
          <year>2021</year>
          ).
          <source>DOI: 10.1007/s11831-021-09562-1</source>
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>S.E.</given-names>
            <surname>Avramenko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.A.</given-names>
            <surname>Zheldak</surname>
          </string-name>
          ,
          <string-name>
            <surname>L.S.</surname>
          </string-name>
          <article-title>Koriashkina “Guided hybrid genetic algorithm for solving global optimization problems”</article-title>
          . Radio Electronics, Computer Science, Control.
          <volume>2</volume>
          (
          <year>2021</year>
          ):
          <fpage>174</fpage>
          -
          <lpage>188</lpage>
          . DOI:
          <volume>10</volume>
          .15588/
          <fpage>1607</fpage>
          -3274-2021-2-18.
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>Weiwei</given-names>
            <surname>Zhang</surname>
          </string-name>
          , Weizheng Zhang, Gary G.
          <article-title>Yen &amp; HongLei Jing “A cluster-based clonal selection algorithm for optimization in dynamic environment”</article-title>
          ,
          <source>Swarm and Evolutionary Computation</source>
          ,
          <volume>50</volume>
          (
          <year>2019</year>
          ):
          <fpage>100454</fpage>
          , DOI: 10.1016/j.swevo.
          <year>2018</year>
          .
          <volume>10</volume>
          .005.
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>T.A.</given-names>
            <surname>Zheldak</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.V.</given-names>
            <surname>Slesarev</surname>
          </string-name>
          &amp;
          <string-name>
            <surname>I.G.</surname>
          </string-name>
          <article-title>Gulina “The algorithm of artificial immune system simulation with Saaty selection operator and one­dimensional local search”</article-title>
          .
          <source>Naukovyi Visnyk Natsionalnoho Hirnychoho Universytetu</source>
          .
          <volume>5</volume>
          (
          <year>2016</year>
          ):
          <fpage>149</fpage>
          -
          <lpage>156</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>R.Q.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Jiang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.D.</given-names>
            <surname>Wang</surname>
          </string-name>
          &amp;
          <string-name>
            <given-names>A.P.</given-names>
            <surname>Roskilly</surname>
          </string-name>
          , “
          <article-title>Energy saving technologies and massthermal network optimization for decarbonized iron and steel industry: A review”</article-title>
          ,
          <source>Journal of Cleaner Production</source>
          . V.
          <volume>274</volume>
          . (
          <year>2020</year>
          ):
          <volume>122997</volume>
          , https://doi.org/10.1016/j.jclepro.
          <year>2020</year>
          .122997
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <surname>C-M. Pintea</surname>
          </string-name>
          ,
          <article-title>Advances in Bio-inspired Computing for Combinatorial Optimization Problem</article-title>
          .
          <source>Springer. Intelligent Systems Reference Library</source>
          .
          <volume>57</volume>
          (
          <year>2014</year>
          ).
          <source>DOI:10.1007/978-3-642-40179-4</source>
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>R.</given-names>
            <surname>Rodrigues</surname>
          </string-name>
          “
          <article-title>A Simple Hack to Boost any Global Optimization Algorithm”</article-title>
          ,
          <source>Towards Data Science</source>
          . (
          <year>2021</year>
          ). URL: https://towardsdatascience.com
          <article-title>/a-simple-hack-to-boost-any-globaloptimization-algorithm-bdea461b87</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>H.</given-names>
            <surname>Maaranen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Miettinen &amp; M.M.</surname>
          </string-name>
          <article-title>Makela “Quasi-Random Initial Population for Genetic Algorithms”</article-title>
          ,
          <source>Computers and Mathematics with Applications</source>
          .
          <volume>47</volume>
          . (
          <year>2004</year>
          )
          <fpage>1885</fpage>
          -
          <lpage>1895</lpage>
          . DOI:
          <volume>10</volume>
          .1016/j.camwa.
          <year>2003</year>
          .
          <volume>07</volume>
          .011
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <surname>OR-Library</surname>
          </string-name>
          .
          <article-title>The collection of test data sets for a variety of Operations Research (OR) problems</article-title>
          . URL: http://people.brunel.ac.uk/~mastjjb/jeb/info.html
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>O.V.</given-names>
            <surname>Egorova</surname>
          </string-name>
          &amp;
          <string-name>
            <given-names>V.</given-names>
            <surname>Ye</surname>
          </string-name>
          . Snytuk “
          <article-title>Bagatovymirna tekhnologiya spryamovanoi optymizacii” [Multidimensional technology of directed optimization]</article-title>
          .
          <source>International Conference “Computational intelligence”</source>
          . (
          <year>2015</year>
          ):
          <fpage>88</fpage>
          -
          <lpage>95</lpage>
          . (In Ukrainian).
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>