<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Optimal Planning in Systems Consisting of Rational Agents</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Igor Sinitsyn</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Anatoliy Doroshenko</string-name>
          <email>a-y-doroshenko@ukr.net</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Serhii Pashko</string-name>
          <email>pashko1955@gmail.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Institute of Software Systems of the National Academy of Sciences of Ukraine</institution>
          ,
          <addr-line>Glushkov prosp. 40, build. 5, Kyiv, 03187</addr-line>
          ,
          <country country="UA">Ukraine</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>The paper is devoted to mathematical methods of planning in systems consisting of rational agents. An agent is an autonomous object that has sources of information about the environment and influences this environment. A rational agent is an agent who has a goal and uses optimal behavioral strategies to achieve it. It is assumed that there is a utility function, which is defined on the set of possible sequences of actions of the agent and takes values in the set of real numbers. The goal of a rational agent is to maximize the utility function. If rational agents form a system, then they have a common goal and act in an optimal way to achieve it. Agents use the optimal solution of the optimization problem, which corresponds to the goal of the system. The problem of linear programming is considered, in which the number of product sets produced by the system is maximized. To solve the nonlinear problem of optimizing the production plan, the conditional gradient method is used, which at each iteration uses a posteriori estimation of the error of the solution and can stopping the calculation process after reaching the required accuracy. Since the rational agents that are part of the system can have separate optimality criteria, multi-criteria optimization problems appear. The article discusses methods for solving such problems, among which is a humanmachine procedure assiciated to the conditional gradient method and at each iteration uses information from the decision maker (DM). The difficulties of this approach are that the DM is not able to make decisions many times under the condition of a significant number of iterations of the nonlinear programming method. The article proposes to replace DM with an artificial neural network. Non-linear and stochastic programming methods are used to find optimal parameters of this network.</p>
      </abstract>
      <kwd-group>
        <kwd>1 Method</kwd>
        <kwd>planning</kwd>
        <kwd>rational agent</kwd>
        <kwd>system</kwd>
        <kwd>multi-criteria optimization</kwd>
        <kwd>neural network</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>In recent years, an agent-oriented approach has been successfully developed within the framework
of the theory of artificial intelligence. Agents are considered rational or intelligent and can form
multi-agent systems and act to achieve a common goal. Agent systems are studied from different
points of view: from the standpoint of the theory of conflict processes, information theory, social
psychology, software engineering, in the context of concepts of electronics. The proposed paper
considers the aspect of optimizing the actions of agents as part of a multi-agent system.</p>
      <p>The paper [1] shows that the main types of activities related to the management of systems of
rational agents and individual agents are cooperation (formation of agent systems), planning and
coordination of agent actions, system placement and recognition. Problems of cooperation, planning
and coordination of agents' actions are studied in detail in the monograph [2]. Recognition tasks are
studied in [3 – 6], placement tasks are studied in papers [7, 8]. In [9], the concept of a rational agent
is defined and used.</p>
      <p>To solve optimization planning problems, that is, to find optimal action plans in multi-agent
systems consisting of rational agents, mathematical programming methods are used, in particular
methods of linear, nonlinear, stochastic, discrete programming [10 – 14]. Multi-criteria optimization
problems and corresponding mathematical methods are important for multi-agent systems [15, 16].
In paper [16], human-machine procedures are used to solve multi-criteria problems, in which the
method of nonlinear programming is combined with the work of a DM. The difficulty of this
approach is that the DM is not able to make decisions multiple times under the condition of a
significant number of iterations of the nonlinear programming method.</p>
      <p>This paper considers the problem of optimal planning of actions of a multi-agent system
consisting of rational agents. Planning optimization problems and corresponding mathematical
methods for their solution are described. In multi-criteria optimization problems, it is proposed to
use an artificial neural network instead of DM, and methods of determining the optimal values of
the parameters of such a network are proposed.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Agents and multi-agent systems</title>
      <p>The concept of a rational agent is widely used in the theory of artificial intelligence, economics,
game theory and has a unifying meaning, allowing to define the main tasks facing agents and
multiagent systems, the relationships between these tasks, etc. In [9], an autonomous object that
perceives the environment with the help of sensors and influences this environment with the help
of executive mechanisms is considered an agent. The agent's program works in a computing device
and receives data from sensors, recognizes and analyzes the data, calculates the optimal strategy of
the agent's behavior, and issues commands to executive mechanisms. An agent can be a computer
program, a robot, or a person. The following definition of a rational agent is given in [9]: “For each
possible perceptual sequence, a rational agent must choose the actions that are expected to
maximize its performance, given the facts provided by the perceptual sequence and all the built-in
knowledge the agent has”.</p>
      <p>Usually, there are different permissible sequences of actions of an agent that lead to a goal. In
this case, it is appropriate to consider that there is a utility function defined on a set of action
sequences of an agent (or a system of agents), which takes values from a set of real numbers. A
rational agent is an agent who, for the sake of achieving a goal, uses the optimal behavior strategy,
maximizing the utility function. A multi-agent system is a system consisting of rational agents who
have a common goal and use an optimal strategy to achieve it. It can be assumed that an
optimization problem is formulated for the system and agents form a behavior strategy using the
optimal solution of this problem. Let us give examples of systems consisting of rational agents:
•
•
•
the system of state administration bodies, enterprises, institutions and organizations of the
country's defense-industrial complex;
the system of economies of different countries developing in cooperation;
a group of drones chasing a target.</p>
      <p>Note that the system of agents may include a central agent that performs some of the
management functions.</p>
      <p>In the literature, there are definitions of agents of other types that differ from rational agents. In
[2], the concept of an intelligent agent is defined, which should have the following properties:
•
•
•
reactivity, i.e. the ability to perceive the state of the surrounding environment and act
accordingly;
proactivity, i.e. the ability to identify one's own initiative;
social activity, i.e. the ability to interact with other agents to achieve a goal.</p>
      <p>Agents are considered autonomous. Autonomy means that the agent's behavior is determined
not only by the environment, but also to a large extent by the properties of the agent.</p>
      <p>The definition of an agent is called "weak" if it contains only the described features, that is,
autonomy, reactivity, proactivity, social activity [17]. If, in addition to the above signs, there are
additional ones in the definition of the agent, then the definition is called "strong". Additional
properties may include:
•
•
•
•
•
•
desires, i.e. situations desirable for the agent;
intentions, i.e. what needs to be done to satisfy desires or to fulfill obligations to other
agents;
goals, i.e. a set of intermediate and final goals of the agent;
obligations to other agents;
knowledge, i.e. a part of knowledge that does not change during the agent's existence;
beliefs, i.e. a part of the agent's knowledge that can change.</p>
      <p>We assume that agents have the above and possibly other properties to the extent necessary to
solve problems and use the resulting solutions.</p>
      <p>Planning is the development of a method of action of a multi-agent system and individual agents
in the future depending on the situations that may arise, the choice of an effective method of action,
optimal allocation of resources. Centralized planning, distributed development of a centralized plan,
distributed development of a distributed plan are possible [18]. Centralized planning is performed
by a dedicated agent that has the necessary resources and information.</p>
      <p>Distributed development of a distributed plan means that there is no dedicated agent, agents
build individual plans interacting with each other. In this case, it is considered that the agents have
limited computing capabilities, information about the local environment, and limited
communication capabilities. Examples include groups of autonomous vehicles, mobile sensor
networks, routing in data networks, transportation systems, multiprocessor computing, energy
systems [19].</p>
      <p>Central planning is discussed next.</p>
    </sec>
    <sec id="sec-3">
      <title>3. Optimization planning models in multi-agent systems and corresponding numerical methods</title>
      <p>Consider optimization problems of planning in multi-agent systems. Such problems can be
applied, in particular, in the process of planning the activities of the defense-industrial complex of
Ukraine. The defense-industrial complex is a system consisting of interrelated research centers,
state organizations, and industrial enterprises that produce goods for military purposes. We
consider each element of this system to be a rational agent. Suppose there is a dedicated agent (for
example, a ministry) that has the information needed for planning.</p>
      <p>
        Let the system contain m rational agents that produce several types of products. Some agents
may not produce final products while performing managerial functions. Let n be the number of
types of products, xij denote the value of the j -th product produced by the i -th agent, let α j be
the given specific weight of the j -th product in the set of final products, x = (x11, x12,..., xmn ). It is
necessary to maximize the number of manufactured sets of final products
m
∑ xij
min i=1
j=1,...,n α j
→
x
max
(
        <xref ref-type="bibr" rid="ref1">1</xref>
        )
subject to
m
∑ xij
i=1
α j
y →
x, y
      </p>
      <p>
        max,
≥ y,
j = 1,..., n,
(
        <xref ref-type="bibr" rid="ref2">2</xref>
        )
(
        <xref ref-type="bibr" rid="ref3">3</xref>
        )
(
        <xref ref-type="bibr" rid="ref4">4</xref>
        )
(
        <xref ref-type="bibr" rid="ref5">5</xref>
        )
(
        <xref ref-type="bibr" rid="ref8">8</xref>
        )
(
        <xref ref-type="bibr" rid="ref9">9</xref>
        )
(
        <xref ref-type="bibr" rid="ref10">10</xref>
        )
(
        <xref ref-type="bibr" rid="ref11">11</xref>
        )
x ∈ X . (
        <xref ref-type="bibr" rid="ref6">6</xref>
        )
      </p>
      <p>
        Problem (
        <xref ref-type="bibr" rid="ref4">4</xref>
        ) – (
        <xref ref-type="bibr" rid="ref6">6</xref>
        ) is a linear programming problem and is solved by linear programming
methods. Note that the numbers appearing in problems (
        <xref ref-type="bibr" rid="ref1">1</xref>
        ) - (
        <xref ref-type="bibr" rid="ref3">3</xref>
        ) and (
        <xref ref-type="bibr" rid="ref4">4</xref>
        ) - (
        <xref ref-type="bibr" rid="ref6">6</xref>
        ) represent part of the
agents' knowledge and beliefs, and the objective function expresses the intentions of the system.
      </p>
      <p>
        Suppose, in the problem (
        <xref ref-type="bibr" rid="ref1">1</xref>
        ) – (
        <xref ref-type="bibr" rid="ref3">3</xref>
        ), instead of the objective function (
        <xref ref-type="bibr" rid="ref1">1</xref>
        ), some objective function
f (x) is used. It is necessary to solve the problem
f (x) → max. (
        <xref ref-type="bibr" rid="ref7">7</xref>
        )
      </p>
      <p>x∈X
We consider the function f (x) to be concave (that is, convex upwards) on the set X and smooth
on some neighborhood of the set X . Smoothness means that the function is defined and has
continuous partial derivatives with respect to all variables, and an open set containing the X is
called a neighborhood of X . To solve this problem, it is advisable to use the conditional gradient
method (Frank and Wolf method) [10, 11], which consists of the following. Let's choose x0 ∈ X . At
the s -th step ( s = 0,1,2,... ) we calculate
x s = arg max ∇f (xs ), x ,</p>
      <p>x∈X
ts = arg max f (xs + t(x s − xs )),</p>
      <p>0≤t≤1
xs+1 = xs + ts (x s − xs ).
n
∑ aijk xij ≤ bik , i = 1,..., m, k = 1,...,l,
j=1
xmin ≤ xij ≤ ximjax , i = 1,.., m,
ij
j = 1,..., n.</p>
      <p>Here aijk is the cost of the k -th resource for the production of the j -th product of unit value at the
i -th enterprise, bik is the stock of the k -th resource at the i -th enterprise, l is the number of
types of resources, ximjin , ximjax are the minimum and maximum permissible quantities of the j -th
product produced at the i -th enterprise, 0 ≤ ximjin ≤ ximjax &lt; ∞.</p>
      <p>
        Let X be the set of vectors x that satisfy conditions (
        <xref ref-type="bibr" rid="ref2">2</xref>
        ), (
        <xref ref-type="bibr" rid="ref3">3</xref>
        ). Using an additional variable y, we
write the problem (
        <xref ref-type="bibr" rid="ref1">1</xref>
        ) – (
        <xref ref-type="bibr" rid="ref3">3</xref>
        ) as follows:
Here ∇f
means the gradient of the function f ,
a,b
      </p>
      <p>is the scalar product of vectors
n
a = (a1,..., an ) and b = (b1,...,bn ), i.e. a,b = ∑ a jbj .</p>
      <p>
        j=1
We denote by x∗ the optimal solution of problem (
        <xref ref-type="bibr" rid="ref7">7</xref>
        ). The following inequality holds
f (x∗ ) − f (xs ) ≤ ∇f (xs ), x s − xs .
      </p>
      <p>Indeed, the concavity of the function f (x) implies an inequality</p>
      <sec id="sec-3-1">
        <title>Obviously,</title>
      </sec>
      <sec id="sec-3-2">
        <title>Last two inequalities imply (11).</title>
        <p>f (x∗ ) − f (xs ) ≤ ∇f (xs ), x∗ − xs .</p>
        <p>∇f (xs ), x∗ ≤ ∇f (xs ), x s .</p>
        <p>The partial boundaries of the sequence {xs} are the maximum points of the function f (x) on
the set X . In [10] it is proved that</p>
        <p>
          f (x∗) − f (xs ) = O(1/ s); (
          <xref ref-type="bibr" rid="ref12">12</xref>
          )
this estimate cannot be improved. It follows from (
          <xref ref-type="bibr" rid="ref12">12</xref>
          ) and the results of [12] that the conditional
gradient method is not suboptimal. It is advisable to use this method for solving problem (
          <xref ref-type="bibr" rid="ref7">7</xref>
          ) due to
the possibility of applying estimate (
          <xref ref-type="bibr" rid="ref11">11</xref>
          ), which allows stopping the computational process after
reaching the required accuracy, and due to its simplicity and satisfactory speed of convergence in
practice. The conditional gradient method is used in the next section as a component of
multicriteria optimization.
        </p>
        <p>
          If the function f is not smooth or instead of the gradient (generalized gradient) of the function
f its stochastic analogue is known, non-smooth optimization and stochastic programming
methods [11 – 13] are used to solve problem (
          <xref ref-type="bibr" rid="ref7">7</xref>
          ). If, instead of problem (
          <xref ref-type="bibr" rid="ref7">7</xref>
          ), a discrete programming
problem is considered, branch-and-bound methods, heuristic algorithms, random search methods
with local optimization, and polynomial approximate schemes can be used [14].
        </p>
        <p>The system of economies of different countries interacting with each other through the export
and import of goods, services, and capital can be considered a system of rational agents, provided
that the governing bodies of the countries make rational economic decisions in market conditions.
There are problems of optimizing the interaction of a separate economy with other economies and
with the world market. Corresponding mathematical models and numerical methods are considered
in works [20, 21].</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Multi-criteria optimization and human-machine procedures</title>
      <p>
        In addition to a common goal, each rational agent of a multi-agent system can have its own
optimality criterion. The problem is to find a quality solution that satisfies all agents. Let the
extremal problem for the i -th agent have the form
fi (x) → max, (
        <xref ref-type="bibr" rid="ref13">13</xref>
        )
      </p>
      <p>x∈X
where i ∈ I , I = {1,..., m}, x = (x1,..., xn ), X is a convex closed bounded set of n -dimensional
Euclidean space. We assume that the criteria are non-negative and normalized, that is, each criterion</p>
      <sec id="sec-4-1">
        <title>The minimum of values α i fi (x) is maximized:</title>
        <p>fi (x) is replaced by fi (x) / f *i , where f *i = max fi (x).</p>
        <p>x∈X</p>
        <p>
          To solve a multi-criteria extremal problem, the convolution method can be applied, replacing
the set of criteria (
          <xref ref-type="bibr" rid="ref13">13</xref>
          ) with a single criterion. Such replacement can be performed in several ways
[15]:
•
•
•
minα i fi (x)
i∈I
→
x∈X
        </p>
        <p>max.
m
∑α i fi (x)
i=1
m
∏α i fi (x)
i=1
→
x∈X</p>
      </sec>
      <sec id="sec-4-2">
        <title>The weighted sum of values fi (x) is maximized:</title>
        <p>Product is maximized:
fk (x)
→
x</p>
        <p>max,
fi (x) ≥ fimin , i ∈ I \ {k},</p>
        <p>x ∈ X .</p>
        <p>Here fimin are given numbers, k ∈ I.
solving an extremal problem
⎛ m p ⎞
⎜ ∑α i fi (x) − fi ⎟
⎝ i=1 ⎠
Here p ≥ 1 is given number.</p>
        <p>1/ p
→
x∈X
min.</p>
        <p>The objective programming method consists of setting the desired values of criteria f1,..., fm and
In the concession method, the criteria are numbered in order of decreasing importance. In the first
step, the first criterion is maximized, f1* = max f1(x), where X1 = X. In the second step, the set
x∈X1
X 2 = {x : x ∈ X1, f1(x) ≥ f1* − Δ } is determined, where Δ1 is the amount of concession for the
1
first criterion, and the problem
f2* = max f2 (x) is solved. In the third step, the set</p>
        <p>x∈X 2
X 3 = {x : x ∈ X 2 , f2 (x) ≥ f2* − Δ2} is determined, where Δ2 is the concession amount for the
second criterion, and the problem f3* = max f3(x) is solved. The process continues until the last
x∈X3
m -th step.</p>
        <p>
          Consider the human-machine method of multi-criteria optimization [16]. This method uses a
non-linear programming method, for which the directions of movement and step sizes are
determined using a DM. We will consider the problem of multi-criteria optimization as follows:
(
          <xref ref-type="bibr" rid="ref14">14</xref>
          )
f (x) = u( f1(x),..., fm (x)) → max.
        </p>
        <p>x∈X
Here u is the utility function, X is a convex closed bounded set of n -dimensional Euclidean
space. We assume that the function f (x) is concave on X . The following conditions [16] are
sufficient for the concavity of the function f on the set X :
•
•
the function u is concave, and the functions fi , i = 1,..., m, are linear;
the functions u and fi , i = 1,..., m, are concave, and the function u is nondecreasing in
each variable.</p>
        <p>Let us assume that the conditions are fulfilled that guarantee the existence of continuous partial
derivatives of the function u( f1(x),..., fm (x)) and the validity of the rule for their calculation [22].</p>
        <p>
          To solve problem (
          <xref ref-type="bibr" rid="ref14">14</xref>
          ), we will use the conditional gradient method (
          <xref ref-type="bibr" rid="ref8">8</xref>
          ) – (
          <xref ref-type="bibr" rid="ref10">10</xref>
          ). According to the
rule for calculating partial derivatives, we have
m ⎛ ∂u ⎞
∇f (xs ) = ∇u( f1(xs ),..., fm (xs )) = ∑⎜⎜
i=1 ⎝ ∂fi ⎠⎟⎟ ∇fi (xs ).
        </p>
        <p>
          s
Here, (∂u ∂fi )s − the i -th partial derivative of u − is calculated at the point ( f1(xs ),..., fm (xs )),
and the gradient ∇fi (x s ) is calculated at the point xs. The linear function ∇f (xs ), x , included in
(
          <xref ref-type="bibr" rid="ref8">8</xref>
          ), is not fully known, since the values of (∂u ∂fi )s are not known. Note that in (
          <xref ref-type="bibr" rid="ref8">8</xref>
          ) the value of
∇f (xs ), x can be multiplied by a positive number. In particular, it can be divided by any positive
coefficient (∂u ∂fi )s. (You can also choose a negative coefficient, such a case is treated similarly).
Without loss of generality, we consider that the first coefficient (∂u ∂f1 )s is chosen as the divisor.
Let's write (
          <xref ref-type="bibr" rid="ref8">8</xref>
          ) as follows:
where wis =
(∂u ∂fi )s .
(∂u ∂f1 )s
        </p>
        <p>m
x s = arg max ∑ wis∇fi (xs ), x ,</p>
        <p>x∈X i=1</p>
        <p>
          In (
          <xref ref-type="bibr" rid="ref15">15</xref>
          ), all quantities, except for wis , are considered known. The values of wis , showing the
relative importance of the first and i -th criteria, are determined using DM, after which optimization
(
          <xref ref-type="bibr" rid="ref15">15</xref>
          ) is performed. In [16], two ways of determining wis are proposed.
        </p>
        <p>Suppose the function u satisfies the conditions of the theorem on the total increment of a
function of several variables [22]. Let us assume that at point ( f1(xs ),..., fm (xs )) criterion f1
received an increment Δ1, criterion fi − an increment Δi , and the values of the remaining criteria
did not change. We have</p>
        <p>Δu = u( f1 (x s ) + Δ1 , f 2 (x s ),..., fi−1 (x s ),
fi (xs ) + Δi , fi+1(xs ),..., fm (xs )) − u( f1(xs ),..., fm (xs )) = (∂u ∂f1 )s Δ1 +</p>
        <p>+ (∂u ∂fi )s Δi + o( (Δ1)2 + (Δi )2 ).</p>
        <p>Let DM choose the values Δ1, Δi so that Δu = 0. In this case, we have approximately
(∂u ∂f1 )s Δ1 + (∂u ∂fi )s Δi = 0, that is</p>
        <p>Δ
wis = − 1 .</p>
        <p>
          Δi
(
          <xref ref-type="bibr" rid="ref15">15</xref>
          )
(
          <xref ref-type="bibr" rid="ref16">16</xref>
          )
(
          <xref ref-type="bibr" rid="ref17">17</xref>
          )
        </p>
        <p>The second way of calculating wis is based on the fact that the gradient of a function indicates
the direction of its fastest growth. Let all criteria except the first and i -th remain unchanged, the
first and i -th grow by amounts δ 1,δ i respectively, and the vector (δ 1,0,...,0,δ i ,0,...,0) indicates
the direction of the fastest growth. In this case, approximately</p>
        <p>δ
wis = i .</p>
        <p>
          δ 1
At step (
          <xref ref-type="bibr" rid="ref9">9</xref>
          ) of the conditional gradient method, the problem is solved
u( f1(xs + t(x s − xs )),..., fm (xs + t(x s − xs )) ) → max .
0≤t≤1
Here, the objective function is a function of one variable t . One graph shows the curves
(18)
f j (xs + t(x s − xs )), j = 1,2,...,m, on the segment [0,1]. Analyzing the constructed graphs, the
DM selects the best value of t from the segment [0,1].
        </p>
        <p>So, the DM calculates the values of Δi or δ i , the optimal value of t in problem (18) and,
possibly, selects the starting point x 0 . All other calculations are performed by a computer program.
In [16], an example of the application of the described human-machine method of multi-criteria
optimization to the organization of the educational process at one of the university faculties is
given. Human-machine procedures can be combined not only with the conditional gradient method,
but also with other methods of nonlinear programming.</p>
        <p>
          Obviously, DM cannot always qualitatively perform the tasks assigned to it in the process of
multi-criteria optimization. In addition, the nonlinear programming method can perform a large
number of iterations, which is unacceptable for DM. Therefore, it is advisable to use an artificial
intelligence system instead of DM. In particular, it is possible to use an artificial neural network that
calculates the values of wis , using formulas (
          <xref ref-type="bibr" rid="ref16">16</xref>
          ) and (
          <xref ref-type="bibr" rid="ref17">17</xref>
          ), or calculates the gradient of the function
f in another way, and the optimal value of t in problem (18). Such networks are considered in the
next section
        </p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Use of artificial neural networks in optimization methods</title>
      <p>Artificial neural networks (hereinafter referred to as neural networks) have become widespread
⎛ r ⎞
as part of artificial intelligence systems. We consider a function y = f ⎜⎜ ∑ wj x j ⎟⎟ to be a
⎝ j=1 ⎠
mathematical model of an artificial neuron, where the real numbers x1,..., xr are the inputs of the
neuron, a real number y is the output, real numbers of w1,..., wr − weighting coefficients. The
scheme of an artificial neuron is shown in Figure 1.</p>
      <p>A neural network combining a large number of artificial neurons is capable of solving complex
problems. We assume that the network has n inputs and one output and calculates some function
f (x1,..., xn ). Let I n = [0;1]n be the n -dimensional unit cube. The following theorem was proved in
[23].</p>
      <p>Theorem 1. For any integer n ≥ 2 there exist such continuous real functions ψ ij ( x), defined on
the unit segment I 1 = [0;1], that each defined on the n -dimensional unit cube I n continuous real
function f (x1,..., xn ) is represented in the form</p>
      <p>2n+1 ⎛ n ⎞
f (x1,..., xn ) = ∑ϕ i ⎜⎜ ∑ψ ij (x j ) ⎟⎟,</p>
      <p>i=1 ⎝ j=1 ⎠
where the functions ϕ i ( y) are real and continuous.
(19)</p>
      <p>Note that the functions ψ ij (x j ) do not depend on the function f . Formula (19) can be
presented in the form of a neural network (Figure 2), where artificial neurons are represented by
rectangles; in this case, the weighting coefficients equal to one. So, the values of any continuous
real function f (x1,..., xn ), defined on the n -dimensional unit cube I n , can be calculated using a
neural network containing 2n2 + 3n + 2 artificial neurons in which the values of the functions
ψ ij ,ϕ i , f are calculated.</p>
      <p>function f (x1,..., xn ) .</p>
      <p>We call a function σ (t) sigmoid if the condition</p>
      <p>⎧1, t → +∞,
σ (t) → ⎨</p>
      <p>⎩0, t → −∞
is satisfied. For example, σ (t) = 1/(1 + e−t ) is a sigmoid function. Let С(I n ) denote the space of
continuous functions on I n . The following theorem was proved in [24].</p>
      <p>Theorem 2. Let σ be any continuous sigmoid function. Finite sums of the form</p>
      <p>N
G(x) = ∑α jσ ( y j , x +θ j ) (20)</p>
      <p>j=1
(where α j ,θ j are real numbers, y j are vectors with real components, y j = ( y j1,..., y jn )) are
dense in С(I n ). In other words, given any f ∈ С(I n ) and ε &gt; 0 there is a sum G(x) of the form
(20) for which</p>
      <p>G(x) − f (x) &lt; ε , ∀x ∈ I n .</p>
      <p>Formula (20) can be represented in the form of a neural network (Figure 3), containing N +1
artificial neurons, which are used to calculate the values of the function G. Theorems 1, 2 justify
the suitability of neural networks for calculating any continuous functions. The neural network,
which is built on the basis of Theorem 1 and accurately calculates the value of the function f ,
contains a limited number of neurons. Functions ϕ i ,ψ ij , the values of which are calculated in
neurons, can be different and not have continuous derivatives. The neural network, which is built
on the basis of Theorem 2 and approximately calculates the value of the function f , contains
artificial neurons that use a single sigmoid function σ . The number of artificial neurons in such a
neural network is not limited. If you choose a sigmoid function σ (t), that has a continuous
derivative, then the quantity G will have continuous partial derivatives for all quantities α j ,θ j
and for all components of vectors y j . This circumstance can play an important role in the process of
finding optimal values of these parameters.</p>
      <p>F (α , y,θ ) = 2E{qα j xkσ ʹ( y j , x +θ j )},</p>
      <p>F (α , y,θ ) = 2E{qα jσ ʹ( y j , x +θ j )},</p>
      <p>Suppose the probability distribution is given on the set X and x is a random element [25]; here
x ∈ X , X is a closed bounded set of n -dimensional Euclidean space. Let α = (α 1,...,α N ),
N
y = ( y11,..., yN1, y12,..., yNn ), θ = (θ1,...,θ N ), q(α , y,θ , x) = ∑α jσ ( y j , x +θ j )− f (x). Let's
j=1
choose the parameters α , y,θ as the optimal solution of the problem</p>
      <p>F (α , y,θ ) = Eq2 (α , y,θ , x)
→
α , y,θ
min .</p>
      <p>(21)
Here E is a sign of mathematical expectation.</p>
      <p>Suppose that the sigmoid function σ (t) has a continuous derivative and the conditions
sufficient for differentiation under the sign of mathematical expectation are met. We have
∂</p>
      <p>F (α , y,θ ) = 2E{qσ ( y j , x +θ j )},</p>
      <p>j = 1,..., N , k = 1,..., n.</p>
      <p>Here σ ʹ(t ) is the derivative of function σ (t ). So, it is easy to calculate the vector g(α , y,θ , x), the
mathematical expectation of which is equal to the gradient ∇F (α , y,θ ) of the function F (α , y,θ ),
therefore, stochastic programming methods [11 – 13] can be used to solve problem (21). In the case
when the objective function of problem (21) is not convex, with the help of these methods we will
obtain a local minimum. In this case, you should use the method of random search with local
optimization [14], in which local optima are calculated using the stochastic programming method.</p>
      <p>Another approach to determining parameters α , y,θ consists in solving the minimax problem
[26]</p>
      <p>F1 (α , y,θ ) = max q(α , y,θ , x)
x∈X
→
α ,y,θ
min .</p>
      <p>
        Suppose that the objective function u( f1(x),..., fm (x)) of problem (
        <xref ref-type="bibr" rid="ref14">14</xref>
        ) can be calculated by a
computer program. Let us choose f (x) = u( f1(x),..., fm (x)) and calculate the parameters α , y,θ
of function (20) using one of the described methods. We can use the appropriate neural network
instead of DM to solve the multi-criteria optimization problem (
        <xref ref-type="bibr" rid="ref14">14</xref>
        ).
      </p>
    </sec>
    <sec id="sec-6">
      <title>6. Conclusions</title>
      <p>The article deals with the problem of optimal planning of actions in multi-agent systems
consisting of rational agents. Definitions of the concepts of agent, rational agent, system consisting
of rational agents are given. Properties of agents and ways of developing action plans of a
multiagent system are considered. The focus is on centralized planning. Optimization problems of
planning and the conditional gradient method are described. Methods for solving multicriteria
optimization problems are considered, including a human-machine procedure in which the DM
takes part. It is proposed to use an artificial neural network instead of DM. Methods are proposed
for determining the optimal values of the parameters of such a network.
[18] E.H. Durfee, Distributed problem solving and planning, in: ECCAI Advanced Course on</p>
      <p>Artificial Intelligence, Springer, Berlin, Heidelberg, 2001, pp. 118–149.
[19] J. Shamma, Cooperative control of distributed multi-agent systems, John Wiley &amp; Sons, 2008.
[20] S.V. Pashko, Optimization of Capital Investment Distribution in an Open Economy on the Basis
of the “Input–Output” Model, Cybernetics and Systems Analysis 58 (2022) 225–232.
doi:10.1007/s10559-022-00454-1.
[21] A.M. Gupal, S.V. Pashko, Optimizing the Capital Investment Distribution Based on a Dynamic
Mathematical Model, Cybernetics and Systems Analysis 60 (2024) 375–382.
doi:10.1007/s10559024-00678-3.
[22] G.M. Fikhtengolts, Course of Differential and Integral Calculus, Vol. 1, Nauka, Moscow, 1969.</p>
      <p>(in Russian).
[23] A.N. Kolmogorov, On the representation of continuous functions of several variables in the
form of superpositions of continuous functions of one variable and addition, DAN USSR 114
(1957) 953–956. (in Russian).
[24] G. Cybenko, Approximation by Superpositions of a Sigmoidal function, Mathematics of</p>
      <p>Control, Signals and Systems 2 (1989) 303–314.
[25] A.N. Shiryaev, Probability, Nauka, Moscow, 1980. (in Russian).
[26] V.F. Demyanov, V.N. Malozemov, Introduction to minimax, Nauka, Moscow, 1972. (in Russian).</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>S.V.</given-names>
            <surname>Pashko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>I.P.</given-names>
            <surname>Sinitsyn</surname>
          </string-name>
          ,
          <article-title>Optimal solutions in systems consisted of rational agents</article-title>
          ,
          <source>Artificial Intelligence</source>
          <volume>2</volume>
          (
          <year>2023</year>
          )
          <fpage>16</fpage>
          -
          <lpage>26</lpage>
          . doi:
          <volume>10</volume>
          .15407/jai2023.
          <fpage>02</fpage>
          .
          <fpage>016</fpage>
          . (in Ukrainian).
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>M.</given-names>
            <surname>Wooldridge</surname>
          </string-name>
          ,
          <article-title>An introduction to multiagent systems</article-title>
          , John Wiley &amp; Sons, Ltd, United Kingdom,
          <year>2009</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>M.I.</given-names>
            <surname>Shlesinger</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.A.</given-names>
            <surname>Glavach</surname>
          </string-name>
          ,
          <article-title>Ten lectures on statistical and structural recognition, Nauk</article-title>
          . Dumka, Kyiv,
          <year>2004</year>
          . (in Russian).
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>A.M.</given-names>
            <surname>Gupal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>I.V.</given-names>
            <surname>Sergienko</surname>
          </string-name>
          ,
          <article-title>Optimal recognition procedures, Nauk</article-title>
          . Dumka, Kyiv,
          <year>2008</year>
          . (in Russian).
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>A.M.</given-names>
            <surname>Gupal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.V.</given-names>
            <surname>Pashko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>I.V.</given-names>
            <surname>Sergienko</surname>
          </string-name>
          ,
          <article-title>Efficiency of Bayesian classification procedure</article-title>
          ,
          <source>Cybernetics and Systems Analysis</source>
          <volume>31</volume>
          (
          <year>1995</year>
          )
          <fpage>543</fpage>
          -
          <lpage>554</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>I.V.</given-names>
            <surname>Sergienko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.M.</given-names>
            <surname>Gupal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.V.</given-names>
            <surname>Pashko</surname>
          </string-name>
          ,
          <article-title>Complexity of classification problems</article-title>
          ,
          <source>Cybernetics and Systems Analysis</source>
          <volume>32</volume>
          (
          <year>1996</year>
          )
          <fpage>519</fpage>
          -
          <lpage>533</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>S.V.</given-names>
            <surname>Pashko</surname>
          </string-name>
          ,
          <article-title>Optimal placement of multy-sensor system for threat detection</article-title>
          ,
          <source>Cybernetics and Systems Analysis</source>
          <volume>54</volume>
          (
          <year>2018</year>
          )
          <fpage>249</fpage>
          -
          <lpage>257</lpage>
          . doi:
          <volume>10</volume>
          .1007/s10559-018-0026-z.
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>S.</given-names>
            <surname>Pashko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Molyboha</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Zabarankin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Gorovyy</surname>
          </string-name>
          ,
          <article-title>Optimal sensor placement for underwater threat detection</article-title>
          ,
          <source>Naval Research Logistics</source>
          <volume>55</volume>
          (
          <year>2008</year>
          )
          <fpage>684</fpage>
          -
          <lpage>699</lpage>
          . doi:
          <volume>10</volume>
          .1002/nav.20311.
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>S.</given-names>
            <surname>Russell</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Norvig</surname>
          </string-name>
          ,
          <article-title>Artificial intelligence: a modern approach, 4th Edn</article-title>
          .,
          <string-name>
            <surname>Pearson</surname>
          </string-name>
          , Hoboken, NJ,
          <year>2021</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>B.T.</given-names>
            <surname>Polyak</surname>
          </string-name>
          <article-title>Introduction to optimization</article-title>
          , Nauka, Moscow,
          <year>1979</year>
          . (in Russian).
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>V.S.</given-names>
            <surname>Mikhalevich</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.M.</given-names>
            <surname>Gupal</surname>
          </string-name>
          ,
          <string-name>
            <surname>V.I. Norkin</surname>
          </string-name>
          ,
          <article-title>Methods of non-convex optimization</article-title>
          , Nauka, Moscow,
          <year>1987</year>
          . (in Russian).
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>A.S.</given-names>
            <surname>Nemirovskii</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.B.</given-names>
            <surname>Yudin</surname>
          </string-name>
          ,
          <article-title>Complexity of problems and effectiveness of optimization methods</article-title>
          , Nauka, Moscow,
          <year>1979</year>
          . (in Russian).
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <surname>Yu.M. Ermoliev</surname>
          </string-name>
          ,
          <article-title>Methods of stochastic programming</article-title>
          ,
          <source>Nauka</source>
          , Moscow,
          <year>1976</year>
          . (in Russian).
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>C.</given-names>
            <surname>Papadimitriou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Steiglitz</surname>
          </string-name>
          , Combinatorial optimization: Algorithms and Complexity, Prentice-Hall, Inc.,
          <string-name>
            <surname>Englewood</surname>
            <given-names>Cliffs</given-names>
          </string-name>
          , New Jersey,
          <year>1982</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>M.E.</given-names>
            <surname>Kondruk</surname>
          </string-name>
          ,
          <string-name>
            <surname>M.</surname>
          </string-name>
          <article-title>M. Malyar, Multi-criteria optimization of linear systems</article-title>
          , DVNZ «Uzhgorod National University», Uzhhorod,
          <year>2019</year>
          . (in Ukrainian).
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <surname>A.M. Geoffrion</surname>
            ,
            <given-names>J.S.</given-names>
          </string-name>
          <string-name>
            <surname>Dyer</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Fienberg</surname>
          </string-name>
          ,
          <article-title>An Interactive Approach for Multi-Criterion Optimization, with an Application to the Operation of an Academic Department</article-title>
          , Management Science
          <volume>19</volume>
          (
          <year>1972</year>
          )
          <fpage>357</fpage>
          -
          <lpage>368</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>V.I.</given-names>
            <surname>Gorodetsky</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.S.</given-names>
            <surname>Grushinsky</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.V.</given-names>
            <surname>Khabalov</surname>
          </string-name>
          ,
          <article-title>Multiagent systems: review of the current state of theory and practice, 1998</article-title>
          . URL: https://www.slideshare.net/ rudnichenko/mas10320580.
          <year>2015</year>
          .
          <article-title>(in Russian)</article-title>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>