<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>M. Zgurovsky);</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>A hybrid approach based on swarm intelligence and behavior trees for coordinating autonomous agents</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Michael Zgurovsky</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Yuriy Zaychenko</string-name>
          <email>zaychenkoyuri@ukr.net</email>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Helen Zaichenko</string-name>
          <email>zaichenko.helen@lll.kpi.ua</email>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Oleksii Kuzmenko</string-name>
          <email>oleksii.kuzmenko@ukr.net</email>
        </contrib>
      </contrib-group>
      <pub-date>
        <year>2026</year>
      </pub-date>
      <volume>000</volume>
      <fpage>0</fpage>
      <lpage>0001</lpage>
      <abstract>
        <p>In this paper, a hybrid approach for coordinating autonomous agents is proposed, combining swarm intelligence based on the GBestPSO (Global Best Particle Swarm Optimization) algorithm and behavior trees (BTs). This approach aims to solve the problem of balancing global coordination of actions and local autonomy of each agent in dynamic and uncertain environments. The proposed two-level control system architecture consists of a planning level (GBestPSO-Level) for global optimization and a behavior level (BTLevel) for tactical behavior. The results were partially supported by the National Research Foundation of Ukraine, grant No. 2025.06/0022 "AI platform with cognitive services for coordinated autonomous navigation of distributed systems consisting of a large number of objects".</p>
      </abstract>
      <kwd-group>
        <kwd>multi-agent systems</kwd>
        <kwd>drone swarm</kwd>
        <kwd>autonomous systems</kwd>
        <kwd>hybrid algorithms</kwd>
        <kwd>self-organization 1</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Modern systems with collective interaction of autonomous agents are attracting increasing attention
from researchers due to their ability to solve complex tasks in environments with a high level of
uncertainty [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. In such systems, agents function in a decentralized manner, exchange information,
and jointly achieve goals, which makes them suitable for use in reconnaissance, monitoring, search
and rescue, logistics, facility security, and other fields [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. The key challenge in this context is to
develop coordination methods that can ensure a balance between global coordination of actions and
local autonomy of each agent.
      </p>
      <p>
        Traditional centralized approaches, which assume the existence of a single controller, often prove
ineffective in dynamic environments where there is a possibility of communication loss, system
failure, or the emergence of new obstacles. In contrast, decentralized methods inspired by natural
models of collective behavior, such as swarms of insects or flocks of birds, demonstrate significantly
greater resilience and flexibility. One of the most common tools in this class is the Particle Swarm
Optimization (PSO) algorithm, which allows agents to collectively find effective solutions by
exchanging local and global information [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ].
      </p>
      <p>
        However, PSO alone does not provide a high level of cognition for an individual agent. It allows
determining the direction and trajectory of movement, but does not provide a flexible mechanism
for decision-making in situations where environmental conditions, unforeseen events, or complex
interaction scenarios must be considered. In this context, the use of behavior trees (BT) [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] is of
considerable interest. This approach, widely used in robotics and the gaming industry, provides
modularity, hierarchy, and simplicity in describing agent behavior [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. Using the BT structure, it is
easy to integrate conditions, sequences of actions, and parallel processes, which allows for the
creation of flexible and adaptive control algorithms.
      </p>
      <p>
        The combination of PSO and BT forms the basis of a hybrid approach in which swarm
optimization is responsible for global coordination and trajectory optimization, while behavior trees
are responsible for local decision-making and tactical responses of agents [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. This approach provides
a double level of stability: on the one hand, the system's ability to achieve common goals is preserved
even with the loss of individual agents or communication channels; on the other hand, each agent
has a sufficient level of autonomy to act in complex and dynamic conditions.
      </p>
      <p>
        This work pays particular attention to the application of a modified GBestPSO algorithm, which
is supplemented by mechanisms of self-organization and adaptation to the influence of external
factors that disrupt coordination [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]. In combination with behavior trees, this approach allows the
creation of an architecture capable of solving both strategic tasks at the group level and tactical tasks
at the individual agent level.
      </p>
      <p>To verify the effectiveness of the proposed approach, two application scenarios are considered.
ability to coordinate a collective attack, ensuring synchronization of actions and achievement of the
goal. The second is reconnaissance of enemy territory, illustrating the effectiveness of the method in
space allocation tasks and avoiding duplication of actions while maintaining global coordination.
Both scenarios confirm that the integration of PSO with behavior trees creates the basis for a more
robust and flexible control architecture in multi-agent systems.</p>
      <p>Thus, the research aims to substantiate and demonstrate the capabilities of a hybrid approach
combining swarm intelligence and behavior trees as a promising direction for the development of
technologies for coordinating autonomous agents in complex and dynamic environments.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related works</title>
      <p>
        Research into the collective behavior of autonomous agents combines biological inspiration,
optimization methods, and engineering approaches. Many models are based on the idea of
selforganization, borrowed from observations of animal-flocks of birds, colonies of ants, or swarms of
insects [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]. These principles were subsequently formalized in the form of optimization algorithms
and coordination models, which are now actively used in robotics, multi-agent systems, and
distributed control systems.
      </p>
      <p>
        One of the most popular paradigms is Particle Swarm Optimization (PSO), proposed by Kennedy
and Eberhart in 1995 [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. It models the process of finding the optimal solution through the dynamics
of a swarm of particles moving in the search space, guided by their own experience and the successes
of their neighbors. The GBestPSO modification focuses on the global leader the most successful
particle that determines the direction of development of the entire group [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]. This increases the
convergence speed but makes the system more vulnerable to local minimum and dependent on a
single center of influence.
      </p>
      <p>
        Other swarm algorithms, such as Ant Colony Optimization (ACO) [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] and Bee Colony
Optimization (BCO) [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ], offer different mechanisms for collective decision-making. ACO is widely
used in routing and logistics problems due to its efficiency in discrete spaces. Its key advantage is
the use of the phenomenon of stigmergy indirect interaction through changes in the environment
(e.g., pheromone trails) [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ]. However, the algorithm requires numerous iterations and significant
resources to maintain global consistency, which reduces its suitability in real-time scenarios.
      </p>
      <p>
        Further modifications of swarm algorithms, such as Firefly Algorithm [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ], Cuckoo Search [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ],
or Glowworm Swarm Optimization [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ], have been proposed to better balance global and local search
capabilities. They demonstrate high efficiency in theoretical tests but are rarely used in practical
robotic systems due to the complexity of parameterization and the lack of proven on-board
implementations.
      </p>
      <p>
        In the early stages of multi-agent system development, simpler approaches were actively used.
The Boids model, proposed by Reynolds in 1987, was the first simulation of bird flock behavior based
on three simple rules: alignment, collision avoidance, and attraction to the center of the group [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ].
Despite its simplicity, this approach is still used in simulations and computer graphics, but its
limitation lies in the absence of a cognitive level agents cannot make complex decisions or adapt
to mission changes.
      </p>
      <p>Another direction of development is consensus algorithms, which focus on achieving a
coordinated state among all agents through iterative information exchange. They scale well and have
proven effective in synchronization tasks but have limited functionality in dynamic environments
where not only coordination, but also adaptive strategy change is required.</p>
      <p>
        The leader-follower model, which places responsibility on a key agent, has similar problems [
        <xref ref-type="bibr" rid="ref16 ref17">16,
17</xref>
        ]. The loss of a leader or their temporary unavailability leads to a breakdown in coordination,
making the architecture vulnerable.
      </p>
      <p>
        Recent years have been marked by the active introduction of deep reinforcement learning (DRL)
methods into multi-agent systems [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ]. DRL allows agents to learn optimal strategies through trial
and error, forming policies capable of generalizing new scenarios. Examples include Deep
QNetwork (DQN) [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ], Proximal Policy Optimization (PPO) [
        <xref ref-type="bibr" rid="ref20">20</xref>
        ], and Multi-Agent Deep Deterministic
Policy Gradient (MADDPG) [
        <xref ref-type="bibr" rid="ref21">21</xref>
        ] algorithms.
      </p>
      <p>
        Despite its high potential effectiveness, DRL faces a few limitations. First, training requires
millions of episodes, making it unsuitable for rapid deployment in the field. Second, the models are
fication of autonomous systems. Third,
high computational costs make them unsuitable for implementation on resource-constrained robots
without powerful GPUs or TPUs [
        <xref ref-type="bibr" rid="ref22">22</xref>
        ].
      </p>
      <p>
        Behavior trees (BTs) have become one of the most promising methods for building control
architectures in robotics [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. They successfully combine modularity, hierarchy, and flexibility,
making them a natural development of finite state machines [
        <xref ref-type="bibr" rid="ref23">23</xref>
        ] and HTN (Hierarchical Task
Networks) planners [
        <xref ref-type="bibr" rid="ref24">24</xref>
        ].
      </p>
      <p>BTs allow you to describe behavior through a combination of control nodes and actions that form
a tree with a clearly defined execution logic. Their key advantages are:
•
•
•
•
ease of reusing individual subtrees in different tasks
ability to easily extend the structure with new nodes without complete redesign
absence of a single critical element, even with the loss of individual agents
ability to rebuild behavior logic during mission execution.</p>
      <p>In decentralized architectures, BTs allow each agent to make local decisions, coordinating with
neighbors only at the level of exchanging minimal information. This makes them suitable for
largescale systems where centralized control is impossible.</p>
      <p>A review of the literature shows that none of the approaches is universal. Swarm optimization
algorithms are good at global planning and search, but do not provide flexible reactive behavior.
Conversely, BTs make it easy to model local adaptability and cognitive autonomy but lack
mechanisms for global optimization.</p>
      <p>
        That is why modern research shows a tendency toward integrating different methods. For
example, PSO or GBestPSO can determine the optimal location or distribution of tasks among agents,
after which local execution is coordinated through BTs [
        <xref ref-type="bibr" rid="ref25 ref6">6, 25</xref>
        ]. This combination provides both global
coordination and local autonomy, which is especially important in scenarios with a high level of
uncertainty or when working in dynamic environments.
      </p>
      <p>Thus, the current state of research in the field of collective behavior of autonomous agents can
be characterized as a gradual transition from monolithic approaches to composite architectures that
129
combine the strengths of different methods. Among them, the integration of GBestPSO for global
search and BTs for local autonomy attracts particular attention. This approach not only solves the
problems inherent in each of the methods separately but also forms a new level of stability and
adaptability necessary for building multi-agent systems capable of functioning in real, unpredictable
conditions.</p>
    </sec>
    <sec id="sec-3">
      <title>3. Hybrid system architecture</title>
      <p>Effective coordination of autonomous agents in a dynamic and resource-constrained environment
requires a multi-level control system capable of combining global planning with local reactive
behavior. The proposed architecture is based on the integration of two approaches: Behavior Trees
(BTs), which provide modularity and flexibility at the tactical level, and the GBest Particle Swarm
Optimization (GBestPSO) algorithm, which is responsible for the strategic coordination of the entire
system.</p>
      <p>The architecture has a two-level structure (Fig. 1): the planning level (GBestPSO-Level) and the
behavior level (BT-Level).</p>
      <p>The planning level (GBestPSO-Level) acts as a global optimizer that forms the strategic basis of
the mission. Its main functions are:</p>
      <p>Optimization of the global configuration of the swarm. The GBestPSO algorithm periodically
calculates the best location of agents in space, considering goals and constraints.
Distribution of tasks and roles. Each agent is assigned a subtask that corresponds to its
capabilities and the strategic goals of the system.</p>
      <p>Scenario planning. In complex missions, it is possible to form several scenarios of events with
the subsequent selection of the optimal one.</p>
      <p>Replanning. In the event of significant changes in the environment (for example, the
emergence of new threats or obstacles), GBestPSO is restarted to correct the strategic plan.
the group of agents and coordinates their actions. Unlike centralized approaches, the architecture
ly sent to agents, rather than as a real-time
directive. This allows the system to remain stable even during temporary communication failures.</p>
      <p>The behavior level (BT-Level) determines the individual tactical behavior of each agent, which is
equipped with a behavior tree. This level includes the following modules:
•
•
•
•
•</p>
      <p>Behavior Manager. Performs real-time interpretation of the behavior tree, activates local
actions, and rebuilds logic when conditions change.</p>
      <p>Perception Module. Provides sensor data processing, object recognition, and local
environment mapping.</p>
      <p>Local Planner. Responsible for low-level navigation obstacle avoidance, local SLAM,
trajectory correction.</p>
      <p>Adaptation Module. Allows dynamic restructuring of the behavior tree in case of agent loss,
role change, or communication interruption.</p>
      <p>Communication Module. Implements information exchange with other agents, including the
transmission of states, positions, and parts of the behavior tree.</p>
      <p>BT-level provides instant responsiveness to environmental events and allows agents to remain
operational even when communication with the planning level is lost.</p>
      <p>Thus, GBestPSO provides global coordination, while BTs ensure local autonomy. Their
interaction is organized through standardized messaging protocols, allowing the system to remain
functional even in the event of partial communication loss.</p>
      <p>The connection between the planning and behavior levels is organized in such a way as to avoid
agents' dependence on a single source of information. The algorithm can be described as follows:
Swarm Optimizer (at the GBestPSO level) calculates the globally best strategy (gBest) and transmits
it as a message, and BT-level perceives this information as external mission conditions, which are
reflected in the tree branches.</p>
      <p>In the event of a change in global strategy, the local behavior of agents is automatically adjusted.
If communication with the planning level is lost, agents act based on the last received reference point,
maintaining the ability to complete the mission. To distinguish between critical and auxiliary data,
channels with different priority levels are used: messages about strategy changes have a higher
: agents remain autonomous but can quickly integrate new information when it becomes
available.</p>
      <p>The proposed two-level architecture has some key advantages:
•
•
•
•
•</p>
      <p>Each agent can work without being constantly connected to the global planner.</p>
      <p>Behavior trees are easy to modify and scale for different missions.</p>
      <p>The system remains operational even if individual agents are lost or communications are
temporarily disrupted.</p>
      <p>The architecture allows new algorithms to be integrated without the need for a complete
system overhaul.</p>
      <p>It is suitable for a wide range of applications, from reconnaissance and escort to attack
scenarios or search and rescue operations.</p>
      <p>Thus, the architecture based on the combination of GBestPSO and BTs provides a balance
between global optimization and local adaptability. It demonstrates the ability to work effectively in
conditions of uncertainty, high environmental dynamics, and limited resources. The use of a
hierarchical approach makes the system fault-tolerant, modular, and suitable for a variety of
collective interaction scenarios.</p>
    </sec>
    <sec id="sec-4">
      <title>4. Description of simulation scenarios and settings</title>
      <p>Unmanned Aerial Vehicles (UAVs) were selected as autonomous agents for the study. The simulation
model is based on a hybrid architecture that combines swarm optimization methods (GBestPSO) for
dynamic modelling of drone trajectories and the Behavior Trees (BT) framework for implementing
adaptive and fault-tolerant control logic. Each UAV functions as an autonomous agent controlled by
its own behavior tree, which cyclically processes the current mission status and external influences.
4.1.</p>
      <p>This scenario simulates a coordinated attack by a swarm of drones on a stationary target, consisting
of two consecutive phases: approach and formation of a ring of fire.</p>
      <sec id="sec-4-1">
        <title>4.1.1. Phase 1: approach to the target</title>
        <p>In the initial stage, the drones move toward the target using a modified swarm optimization
algorithm (GBestPSO). The main difference between this algorithm and others is that the
acceleration coefficients that regulate the influence of the leader ( 1) and the target ( 2) are not
constants but are linearly dependent on the current distance of the drone from the corresponding
object.</p>
        <p>The equation for updating the speed of drone  in direction  is as follows:
  ( + 1) =   ( ) +  1 ( (  ( ),   ( ))) ∙  1( ) ∙ [   ( ) −   ( )] +</p>
        <p>+ 2 ( ( ∗( ),   ( ))) ∙  2( ) ∙ [  ∗( ) −   ( )],
where coefficients  1 and  2 are calculated using the following formulas:</p>
        <p>−  1
 1 ( (  ( ),   ( ))) =
 2 ( ( ∗( ),   ( ))) =
 1
‖  (0) −   (0)‖
 2 −  2
‖ ∗(0) −   (0)‖
∙ (  ( ) −   ( )) +  1
∙ (  ∗( ) −   ( )) +  2</p>
        <p>The UAVs follow these routes until the leader of the swarm (the UAV closest to the target) reaches
the specified attack radius  0.</p>
      </sec>
      <sec id="sec-4-2">
        <title>4.1.2. Phase 2: firing by the ring</title>
        <p>Once the leader reaches the target's vicinity with a radius of  0 (100 200 m), the swarm enters the
phase of rotation around it. Movement in this phase is described in a polar coordinate system relative
to the target, which is considered the center.</p>
        <p>The coordinates of the leader   in the Cartesian system can be calculated based on its angular
position   ( ) and radius  0:
 1 ( ) =  0 ∙</p>
        <p>(  ( ))
{ 2 ( ) =  0 ∙  (  ( ))
The angular position is updated at each step of the simulation using the following formula:
  ( + 1) =   ( ) +   ( )∆
where   ( ) is the angular velocity of rotation.
(1)
(2)
(3)
(4)
(5)</p>
        <p>The synchronization stage of the attack is critically important and is provided through the
Decorator~ node in the Behavior Tree. This node allows the final attack to begin only after a certain
minimum number of UAVs (  ∗) reach orbit, ensuring a simultaneous and effective strike.</p>
        <p>The simulation settings for this scenario are shown in Table 1.</p>
        <sec id="sec-4-2-1">
          <title>Communication</title>
        </sec>
      </sec>
      <sec id="sec-4-3">
        <title>4.2. Scenario 2. Reconnaissance of enemy territory</title>
        <p>This scenario demonstrates the capabilities of a swarm for autonomous reconnaissance, where UAVs
effectively divide the search area among themselves. The simulation scenario involves coordinating
a swarm of autonomous agents to conduct reconnaissance of enemy territory. Each UAV is assigned
a specific area for reconnaissance. The results from all areas are combined into a common map of
the terrain, which significantly speeds up the process. The main goal is to ensure complete coverage
of the designated area and localization of reconnaissance targets, considering possible UAV losses
and route replanning.</p>
      </sec>
      <sec id="sec-4-4">
        <title>4.2.1. Problem statement</title>
        <p>Let the surveyed area be defined as a rectangular region with coordinates:</p>
        <p>The swarm consists of  agents, each of which receives an individual area for exploration. Each
area is defined by a strip along the  -axis:</p>
        <sec id="sec-4-4-1">
          <title>Coordinates of the  -th zone:</title>
          <p>This data is transmitted to agents via a communication system.</p>
        </sec>
      </sec>
      <sec id="sec-4-5">
        <title>4.2.2. Routes to reconnaissance areas</title>
        <p>A hybrid GBestPSO+BTs algorithm is used to plan trajectories, where:
{
  ( + 1) =   ( ) +   1( )∆ 
  ( + 1) =   ( ) +   2( )∆ 
where ⃗⃗⃗⃗(⃗⃗⃗⃗⃗) = (  1( ),   2( )) is the velocity vector of UAV  at iteration  , and ∆  is the time

step.
, =  
+ ( − 1)∆</p>
        <p>+  ∆
, =</p>
        <p>The swarm supervisor evaluates the coordinates of the target areas and optimizes the routes to
avoid collisions and ensure minimum arrival time.</p>
      </sec>
      <sec id="sec-4-6">
        <title>4.2.3. Movement in the reconnaissance zone</title>
        <p>UAVs move along strips with a width of  = 2 0, where  0 is the observation radius of the UAV.
Movement along the  axis is defined by the equation:</p>
        <p>After reaching the upper limit of the survey area, the UAV shifts along  by  and returns in the
2
opposite direction to survey the adjacent strip.</p>
        <p>Conditions for completing the survey of the area:
  ( + 1) =   ( ) +   ( )∆
  ( + 1) +</p>
        <p>≥   ,

2
coefficient ( 1( ))</p>
        <p>Communication
)</p>
        <p>)
( )
,  
,</p>
        <sec id="sec-4-6-1">
          <title>Scale</title>
        </sec>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Experimental investigations</title>
      <p>were varied. The main performance metrics were:</p>
      <p>If some UAVs are lost or an area remains unexplored, the hybrid algorithm determines which
UAVs should perform a re-survey. To do this, the position update rule is used:
  ( + 1) =   ( ) +  1( ) 1( 
, −   ( ))
where  1( ) is the cognitive influence coefficient, and  1 ~ (0,1) is a random variable.
After completing the survey, the UAVs return to their initial coordinates (  0,   0).
The simulation is configured using the parameters described in Table 2.
•
•
•
time to attack initiation (number of iterations until the minimum number of UAVs entered
orbit)
total attack completion time (number of iterations until all UAVs targeted the object)
UAV loss during approach and attack</p>
      <p>Value
6-12
2 0
(12)</p>
      <p>uniformity of UAV placement in orbit (average distance between UAVs in orbit).</p>
      <p>During each experiment, the launch of UAV from a certain area of the arena was simulated (Fig.</p>
      <p>After launch, the UAVs were controlled by the internal BTs mechanism and moved to the target
coordinates along the shortest trajectory (Fig. 3).
Behavior Trees (BT)</p>
      <p>The data shows that increasing the number of UAVs speeds up the start of the attack and reduces
the time it takes to complete it due to faster formation of the orbit around the target. UAV losses
remain low, indicating the stability of the behavior tree algorithm.
The graph shows three stages of UAV behavior:
•
•
•</p>
      <p>Approaching UAVs move toward orbit
On Orbit UAVs form a circle around the target</p>
      <p>Attacking UAVs attack the target.</p>
      <p>The next series of experiments was conducted to evaluate the effectiveness of the proposed hybrid
consisted of 15 UAVs that took off from an area located on the right side of the arena and had to
coordinate and gather around the target at a specified point.</p>
      <p>The purpose of the experiment was to measure the swarm's ability to:
•
form a stable orbit around the target
achieve a uniform angular distribution of UAVs in orbit
initiate a coordinated attack when the appropriate conditions are met
minimize UAV losses.</p>
      <p>After the launch of the UAVs, the swarm leader was immediately determined (Fig. 5), which was
at the shortest distance to the target's orbit.</p>
      <p>During the movement to the target orbit, the UAVs approached the leader, forming a swarm (Fig.
6).</p>
      <p>Upon reaching the target orbit, the swarm distributed itself along the orbit (Fig. 7).</p>
      <p>The swarm dispersed in orbit until the minimum number of UAVs required for the attack had
gathered in orbit. After that, a simultaneous attack from different directions began (Fig. 8).</p>
      <p>The key metrics obtained during the experiments are shown in Table 5.</p>
      <p>Figure 9 shows the evolution of the number of UAVs in different states (Approaching, Orbiting,
Attacking) depending on iterations.</p>
      <p>Three distinct phases were observed:
•
•
•</p>
      <p>Phase I (0 50 iterations), during which most UAVs approach the target, with a few devices
beginning to enter orbit.</p>
      <p>Phase II (50 120 iterations), when the number of UAVs in orbit steadily increases, reaching
a uniform angular distribution.</p>
      <p>Phase III (after ~120 iterations), when a coordinated attack begins, leading to a sharp decrease
in the number of UAVs in orbit and an increase in attackers.</p>
      <p>The final series of experiments was conducted to simulate a swarm of UAVs for the
and was tasked with completely surveying a rectangular reconnaissance area located on the left side
of the arena.</p>
      <p>The UAVs were assigned individual reconnaissance zones and coordinated their actions using
behavior trees (BTs), while their movement and collision avoidance were controlled by the GBestPSO
mechanism (Fig. 11).</p>
      <p>Upon reaching the reconnaissance zone, each UAV began surveying a separate area (Fig. 12).</p>
      <p>After successfully surveying the entire reconnaissance area, the UAVs returned to their
permanent deployment location (Fig. 13).</p>
      <p>For each simulation run, metrics were recorded that characterized the speed and quality of task
execution, as well as the system's resistance to UAV losses.</p>
      <p>Based on the research conducted, the following key metric values were obtained (Table 6).</p>
      <p>The dynamics of changes in the number of UAVs depended on the corresponding phase of the
mission:
•
•
•</p>
      <p>Phase I (0 50 iterations) most UAVs move to the areas; the first reconnaissance routes are
gradually activated.</p>
      <p>Phase II (50 180 iterations) the number of UAVs in Reconnaissance mode increases,
gradually covering the entire area.</p>
      <p>Phase III (after ~180 iterations) most UAVs complete their survey of the areas and switch
to Returning or Awaiting mode.</p>
    </sec>
    <sec id="sec-6">
      <title>6. Analysis and discussion</title>
      <p>Experimental investigations confirm the effectiveness of the proposed hybrid approach, which
combines swarm optimization (GBestPSO) and behavior trees (BTs). Analysis of the results for two
key scenarios</p>
      <p>allows us to draw reasonable conclusions about the advantages of this architecture.</p>
      <p>A comparison of the results for the attack scenario demonstrates the clear advantage of the hybrid
method over a system that uses only behavior trees. Although the BT-based approach provides a
141
faster start to the attack (28 45 iterations versus 121 in the hybrid), it is inferior in terms of
coordination quality. Agents controlled only by BTs move toward the target along the shortest
trajectory, resulting in a less organized formation in orbit. In contrast, the hybrid system uses
GBestPSO to pre-cluster the swarm around the leader, which, although it takes more time in the
initial stage, provides significantly better synchronization and uniform angular distribution in orbit.
This, in turn, leads to a more effective and simultaneous attack from different directions, minimizing
the target's chances of countering.</p>
      <p>In the reconnaissance scenario, the hybrid algorithm demonstrated high efficiency in solving
space allocation and fault tolerance problems. The system successfully distributed the
reconnaissance area between UAVs and ensured complete coverage of the territory, even under
conditions of simulated agent losses. The two-level architecture plays a key role here: GBestPSO is
responsible for strategic zone allocation and replanning in case of losses, while BTs control the
tactical behavior of each UAV within its zone. This confirms that the proposed approach not only
solves the problem of global optimization but also provides the local autonomy necessary to adapt
to unpredictable circumstances.</p>
      <p>Thus, the experiments prove the central thesis of the work: the integration of swarm intelligence
for global planning and behavior trees for local execution creates a synergistic effect. GBestPSO
provides the system with strategic coordination, while BTs provide tactical flexibility and
responsiveness. This allows for a balance between global coordination and individual autonomy of
agents, making hybrid architecture a promising solution for complex missions in dynamic
environments.</p>
    </sec>
    <sec id="sec-7">
      <title>7. Practical aspects and implementation</title>
      <p>The proposed hybrid architecture is not only theoretically sound, but also practically implementable
thanks to the use of modern technologies and modularity principles. The implementation is based
on a two-level structure that includes planning (GBestPSO) and behavior (BT) levels, which interact
through standardized messaging protocols.</p>
      <p>The key advantages of the architecture in terms of implementation are modularity and scalability.
Behavior trees are modular by nature, which makes it easy to reuse, extend, and modify the behavior
logic of agents without having to redesign the entire system. Adding new UAVs to the swarm does
not require changing the architecture, as GBestPSO works effectively with different numbers of
agents, and each new agent functions as an independent unit with its own BT.</p>
      <p>An important practical aspect is fault tolerance. The system is designed according to the principle
agent can continue to perform tasks based on the last instructions received. The loss of individual
UAVs also does not lead to mission failure, as the system is capable of redistributing tasks among
active agents.</p>
      <p>Finally, unlike resource-intensive methods such as deep reinforcement learning (DRL), the
proposed approach is computationally efficient. This makes it suitable for implementation on UAV
onboard computers with limited hardware resources, which is critical for practical application in the
field.</p>
    </sec>
    <sec id="sec-8">
      <title>8. Conclusions</title>
      <p>The paper investigated and substantiated the effectiveness of a hybrid approach to coordinating
autonomous agents, integrating swarm intelligence based on the GBestPSO algorithm and a control
architecture based on behavior trees (BTs). The main problem addressed by the study is the need to
ensure a balance between global coordination of group actions and local autonomy of each agent in
dynamic and unpredictable environments.
The hybrid approach was confirmed to be superior. In the attack scenario, the hybrid system,
although requiring more time to prepare, provided a significantly higher level of coordination
and synchronization of the strike compared to the approach based solely on BTs.
Fault tolerance and adaptability were demonstrated. In the reconnaissance scenario, the
system effectively distributed tasks among agents, adapted to UAV losses, and successfully
completed the mission, confirming its reliability.</p>
      <p>The synergy of the two methods was substantiated. It was proven that the combination of
global optimization using GBestPSO and local flexibility of BTs allows creating a system that
is both purposeful and adaptive.</p>
      <p>Thus, the main contribution of this work is to demonstrate that the integration of swarm
intelligence and behavior trees is a promising direction for creating a new generation of multi-agent
systems. Such systems are capable of functioning effectively in complex real-world conditions,
making them suitable for a wide range of applications, from military operations to search and rescue
missions.</p>
    </sec>
    <sec id="sec-9">
      <title>9. Future research and applications</title>
      <p>Despite the successful results, there are several promising areas for further development and
improvement of the proposed hybrid approach. Future research may focus on the following aspects:</p>
      <p>The proposed two-level architecture, where GBestPSO is responsible for strategic planning and
BTs for tactical reactive behavior, demonstrated high efficiency in simulation experiments. Two
application scenarios were analyzed: coordinated swarm attack and territory reconnaissance.</p>
      <p>Key results of the work:
•
•
•</p>
      <p>Expanding the behavior tree library. Creating more complex and versatile sub-trees to
implement a wider range of tactical actions, such as evading electronic warfare systems,
dynamically changing roles (e.g., from reconnaissance to strike UAV), or cooperative
interaction to perform complex tasks.</p>
      <p>Intellectualization of GBestPSO. Improving the swarm optimization algorithm by integrating
mechanisms for adapting to dynamic changes in the environment in real time. For example,
the algorithm parameters could be automatically adjusted depending on the threat level,
obstacle density, or communication channel availability.</p>
      <p>Heterogeneous swarms. Adapting the architecture to manage heterogeneous swarms
consisting of agents with different capabilities (e.g., UAVs for reconnaissance, electronic
warfare, and strikes). This will require the development of more complex mechanisms for
distributing tasks and roles at the GBestPSO level.</p>
      <p>Combination with learning methods. Investigation of the possibilities for synergy between
the proposed approach and elements of machine learning. For example, reinforcement
learning methods can be used for offline optimization of behavior tree parameters or
GBestPSO, which will combine the advantages of transparency and verifiability of classical
methods with the high efficiency of data-based models.
-in-themodel, followed by a transition to full-scale field tests on real UAVs. This will allow verifying
the effectiveness of algorithms in conditions of real communication delays, sensor noise, and
other physical limitations.</p>
    </sec>
    <sec id="sec-10">
      <title>Declaration on Generative AI</title>
      <p>The authors have not employed any Generative AI tools.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>M.</given-names>
            <surname>Falco</surname>
          </string-name>
          , G. Robiolo,
          <article-title>Tendencies in multi-agent systems: A systematic literature review</article-title>
          ,
          <source>CLEI Electronic Journal</source>
          <volume>23</volume>
          (
          <year>2020</year>
          ).
          <source>doi:10.19153/cleiej.23.1</source>
          .1.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>A.</given-names>
            <surname>Farinelli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Rogers</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N. R.</given-names>
            <surname>Jennings</surname>
          </string-name>
          ,
          <article-title>Agent-based decentralised coordination for sensor networks using the max-sum algorithm</article-title>
          ,
          <source>Autonomous Agents and Multi-Agent Systems</source>
          <volume>28</volume>
          (
          <year>2014</year>
          ).
          <source>doi:10.1007/s10458-013-9225-1.</source>
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3] - International
          <source>Conference on Neural Networks</source>
          , vol.
          <volume>4</volume>
          ,
          <string-name>
            <surname>Perth</surname>
          </string-name>
          , WA, Australia,
          <year>1995</year>
          , pp.
          <fpage>1942</fpage>
          <lpage>1948</lpage>
          . doi:
          <volume>10</volume>
          .1109/ICNN.
          <year>1995</year>
          .
          <volume>488968</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>M.</given-names>
            <surname>Colledanchise</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Ogren</surname>
          </string-name>
          ,
          <source>Behavior Trees in Robotics and AI: An Introduction</source>
          ,
          <year>2018</year>
          . doi:
          <volume>10</volume>
          .1201/9780429489105.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>M.</given-names>
            <surname>Iovino</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Scukins</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Styrud</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Ogren</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Smith</surname>
          </string-name>
          ,
          <article-title>A survey of behavior trees in robotics</article-title>
          and
          <source>AI</source>
          ,
          <source>Robotics and Autonomous Systems</source>
          <volume>154</volume>
          (
          <year>2022</year>
          ). doi:
          <volume>10</volume>
          .1016/j.robot.
          <year>2022</year>
          .
          <volume>104096</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>H.</given-names>
            <surname>Karishma</surname>
          </string-name>
          ,
          <article-title>A new hybrid particle swarm optimization algorithm for optimal tasks scheduling in distributed computing system</article-title>
          ,
          <source>Intelligent Systems with Applications</source>
          <volume>18</volume>
          (
          <year>2023</year>
          ). doi:
          <volume>10</volume>
          .1016/j.iswa.
          <year>2023</year>
          .
          <volume>200219</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Shi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. C.</given-names>
            <surname>Eberhart</surname>
          </string-name>
          ,
          <article-title>A modified particle swarm optimizer</article-title>
          ,
          <source>in: Proceedings of the IEEE Conference on Evolutionary Computation (ICEC)</source>
          ,
          <year>1998</year>
          , pp.
          <fpage>69</fpage>
          -
          <lpage>73</lpage>
          . doi:
          <volume>10</volume>
          .1109/ICEC.
          <year>1998</year>
          .
          <volume>699146</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>E.</given-names>
            <surname>Bonabeau</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Dorigo</surname>
          </string-name>
          , G. Theraulaz,
          <source>Swarm Intelligence: From Natural to Artificial Systems</source>
          ,
          <year>1999</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>M.</given-names>
            <surname>Dorigo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Birattari</surname>
          </string-name>
          , T. Stützle,
          <article-title>Ant colony optimization</article-title>
          ,
          <source>IEEE Computational Intelligence Magazine</source>
          <volume>1</volume>
          (
          <year>2006</year>
          )
          <fpage>28</fpage>
          -
          <lpage>39</lpage>
          . doi:
          <volume>10</volume>
          .1109/
          <string-name>
            <surname>MCI</surname>
          </string-name>
          .
          <year>2006</year>
          .
          <volume>329691</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>D.</given-names>
            <surname>Karaboga</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Basturk</surname>
          </string-name>
          ,
          <article-title>A powerful and efficient algorithm for numerical function optimization: Artificial bee colony (ABC) algorithm</article-title>
          ,
          <source>Journal of Global Optimization</source>
          <volume>39</volume>
          (
          <year>2007</year>
          )
          <fpage>459</fpage>
          -
          <lpage>471</lpage>
          . doi:
          <volume>10</volume>
          .1007/s10898-007-9149-x.
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>F.</given-names>
            <surname>Heylighen</surname>
          </string-name>
          ,
          <article-title>Stigmergy as a universal coordination mechanism II: Varieties and evolution</article-title>
          ,
          <source>Cognitive Systems Research</source>
          <volume>38</volume>
          (
          <year>2015</year>
          ). doi:
          <volume>10</volume>
          .1016/j.cogsys.
          <year>2015</year>
          .
          <volume>12</volume>
          .007.
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>X.-S.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <article-title>Firefly algorithms for multimodal optimization</article-title>
          ,
          <source>in: Lecture Notes in Computer Science</source>
          , vol.
          <volume>5792</volume>
          ,
          <year>2010</year>
          , pp.
          <fpage>169</fpage>
          -
          <lpage>178</lpage>
          . doi:
          <volume>10</volume>
          .1007/978-3-
          <fpage>642</fpage>
          -04944-6_
          <fpage>14</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>X.-S.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Deb</surname>
          </string-name>
          ,
          <article-title>Cuckoo search via Lévy flights</article-title>
          ,
          <source>in: Proceedings of the World Congress on Nature and Biologically Inspired Computing (NABIC)</source>
          ,
          <year>2009</year>
          , pp.
          <fpage>210</fpage>
          -
          <lpage>214</lpage>
          . doi:
          <volume>10</volume>
          .1109/NABIC.
          <year>2009</year>
          .
          <volume>5393690</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>N.</given-names>
            <surname>Zainal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Zain</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Radzi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Udin</surname>
          </string-name>
          ,
          <article-title>Glowworm swarm optimization (GSO) algorithm for optimization problems: A state-of-the-art review</article-title>
          ,
          <source>Applied Mechanics and Materials</source>
          <volume>421</volume>
          (
          <year>2013</year>
          )
          <fpage>507</fpage>
          -
          <lpage>511</lpage>
          . doi:
          <volume>10</volume>
          .4028/www.scientific.net/AMM.421.507.
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>C. W.</given-names>
            <surname>Reynolds</surname>
          </string-name>
          ,
          <article-title>Flocks, herds and schools: A distributed behavioral model</article-title>
          ,
          <source>in: Proceedings of -34. doi:10.1145/37401</source>
          .37406.
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>G.</given-names>
            <surname>Xie</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Hu</surname>
          </string-name>
          ,
          <string-name>
            <surname>C.-D. Wang</surname>
          </string-name>
          ,
          <article-title>Consensus enhancement for multi-agent systems with rotating-segmentation perception</article-title>
          ,
          <source>Applied Intelligence</source>
          <volume>53</volume>
          (
          <year>2022</year>
          ). doi:
          <volume>10</volume>
          .1007/s10489-022-03687- x.
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>M.</given-names>
            <surname>Naserian</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Ramazani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Khaki</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Moarefianpour</surname>
          </string-name>
          ,
          <article-title>Leader-follower consensus control for a nonlinear multi-agent robot system with input saturation and external disturbance</article-title>
          ,
          <source>Systems Science &amp; Control Engineering</source>
          <volume>9</volume>
          (
          <year>2021</year>
          )
          <fpage>260</fpage>
          -
          <lpage>271</lpage>
          . doi:
          <volume>10</volume>
          .1080/21642583.
          <year>2021</year>
          .
          <volume>1897959</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>C.</given-names>
            <surname>Zhu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Dastani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <article-title>A survey of multi-agent deep reinforcement learning with communication</article-title>
          ,
          <source>Autonomous Agents and Multi-Agent Systems</source>
          <volume>38</volume>
          (
          <year>2024</year>
          ).
          <source>doi:10.1007/s10458- 023-09633-6.</source>
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>V.</given-names>
            <surname>Mnih</surname>
          </string-name>
          , et al.,
          <article-title>Human-level control through deep reinforcement learning</article-title>
          ,
          <source>Nature</source>
          <volume>518</volume>
          (
          <year>2015</year>
          )
          <fpage>529</fpage>
          -
          <lpage>533</lpage>
          . doi:
          <volume>10</volume>
          .1038/nature14236.
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>X.</given-names>
            <surname>Jin</surname>
          </string-name>
          , W. Zhengxiao,
          <article-title>Proximal policy optimization based dynamic path planning algorithm for mobile robots</article-title>
          ,
          <source>Electronics Letters</source>
          <volume>58</volume>
          (
          <year>2021</year>
          ). doi:
          <volume>10</volume>
          .1049/ell2.
          <fpage>12342</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>R.</given-names>
            <surname>Lowe</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Tamar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Harb</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Abbeel</surname>
          </string-name>
          ,
          <string-name>
            <surname>I.</surname>
          </string-name>
          <article-title>Mordatch, Multi-agent actor-critic for mixed cooperative-competitive environments</article-title>
          ,
          <source>in: Proceedings of the 31st International Conference on Inc.</source>
          ,
          <string-name>
            <surname>Red</surname>
            <given-names>Hook</given-names>
          </string-name>
          , NY,
          <year>2017</year>
          , pp.
          <fpage>6382</fpage>
          -
          <lpage>6393</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>J.</given-names>
            <surname>Kober</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Bagnell</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Peters</surname>
          </string-name>
          ,
          <article-title>Reinforcement learning in robotics: A survey</article-title>
          ,
          <source>The International Journal of Robotics Research</source>
          <volume>32</volume>
          (
          <year>2013</year>
          )
          <fpage>1238</fpage>
          -
          <lpage>1274</lpage>
          . doi:
          <volume>10</volume>
          .1177/0278364913495721.
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <given-names>A.</given-names>
            <surname>Ulusoy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Smith</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Belta</surname>
          </string-name>
          ,
          <article-title>Optimal multi-robot path planning with LTL constraints: Guaranteeing correctness through synchronization</article-title>
          , in: Springer Tracts in Advanced Robotics, vol.
          <volume>104</volume>
          ,
          <year>2012</year>
          , pp.
          <fpage>447</fpage>
          -
          <lpage>462</lpage>
          . doi:
          <volume>10</volume>
          .1007/978-3-
          <fpage>642</fpage>
          -55146-8_
          <fpage>24</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [24]
          <string-name>
            <given-names>D.</given-names>
            <surname>Nau</surname>
          </string-name>
          , T.-
          <string-name>
            <surname>C. Au</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          <string-name>
            <surname>Ilghami</surname>
            ,
            <given-names>U.</given-names>
          </string-name>
          <string-name>
            <surname>Kuter</surname>
            ,
            <given-names>J. W.</given-names>
          </string-name>
          <string-name>
            <surname>Murdock</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Wu</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Yaman</surname>
          </string-name>
          ,
          <article-title>SHOP2: An HTN planning system</article-title>
          ,
          <source>Journal of Artificial Intelligence Research</source>
          <volume>20</volume>
          (
          <year>2003</year>
          )
          <fpage>379</fpage>
          -
          <lpage>404</lpage>
          . doi:
          <volume>10</volume>
          .1613/jair.1141.
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          [25]
          <string-name>
            <given-names>F.</given-names>
            <surname>Medeiros</surname>
          </string-name>
          , pyAutonomousAgent:
          <article-title>An academic tool for modeling autonomous agent behaviors using behavior trees</article-title>
          ,
          <source>Journal of Aerospace Technology and Management</source>
          <volume>16</volume>
          (
          <year>2024</year>
          ). doi:
          <volume>10</volume>
          .1590/jatm.v16.
          <fpage>1352</fpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>