<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>A Formal Analysis of Hierarchical Planning with Multi-Level Abstraction</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Michael Staud</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>StaudSoft UG</institution>
          ,
          <addr-line>D-88213 Ravensburg</addr-line>
          ,
          <country country="DE">Germany</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Ulm University, Institute of Artificial Intelligence</institution>
          ,
          <addr-line>James-Franck-Ring, 89081 Ulm</addr-line>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2025</year>
      </pub-date>
      <abstract>
        <p>We analyze Hierarchical World State Planning (HWSP), a novel algorithm that tackles the scalability limitations of Hierarchical Task Network (HTN) planning. By combining multi-level state abstraction with predictive task decomposition, HWSP reduces exponential search space growth. We formalize predictive separable tasks, classify planning domains, and derive tight bounds on search complexity. Our results show that HWSP transforms exponential interactions into linear coordination, enabling near-linear top-level planning efort in favorable domains. This enables eficient planning in complex domains that were previously intractable, opening up new possibilities for real-world applications.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Hierarchical World State Planning</kwd>
        <kwd>HTN Planning</kwd>
        <kwd>Multi-Level Abstraction</kwd>
        <kwd>Predictive Task Decomposition</kwd>
        <kwd>Domain Classification</kwd>
        <kwd>Search Complexity Analysis</kwd>
        <kwd>Scalability in AI Planning</kwd>
        <kwd>Backtracking Behavior</kwd>
        <kwd>Separable Abstract Tasks</kwd>
        <kwd>Probabilistic Prediction Models</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Complex real-world domains, such as urban planning or logistics problems, are challenging due to their
scale and intricacy. One solution is to use a Hierarchical Task Network (HTN) planning approach to
reduce the size of the search space. However, these problems are often still too large—both in memory
and computational demands—to be handled by today’s planners. One new approach is the Hierarchical
World State Planning (HWSP) algorithm [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], which reduces the search space by abstracting the world
state. The problem is then divided into smaller, manageable subproblems.
      </p>
      <p>In this paper, we provide a theoretical analysis of HWSP’s planning eficiency across varying domain
structures. We introduce a classification of domains – EX1, EX2, and EX3 – that captures how separable
tasks interact and when backtracking is necessary. This classification yields new insights into the
structural properties that afect planning complexity.</p>
      <p>Our results establish tight upper bounds on the number of states visited during search under each
domain class. We contrast these results with classical HTN planning, showing that under certain
assumptions (e.g., sibling task independence), HWSP transforms the exponential interaction between
sibling tasks into linear coordination, while preserving the necessary exponential search within each
separable task’s local scope. Furthermore, we incorporate a probabilistic analysis of binary decomposition
predictors, showing how predictor quality impacts search efort in EX2 domains.</p>
      <p>By grounding HWSP in a rigorous formal framework and connecting it to classical HTN planning
theory, we provide a new foundation for understanding the complexity of hierarchical planning. Our
ifndings demonstrate that HWSP can ofer substantial improvements in planning eficiency by exploiting
structural domain properties and leveraging prediction mechanisms.</p>
      <p>We organize the paper as follows: After reviewing related work in Section 2, Section 3 introduces
the formal framework of Hierarchical World State Planning (HWSP), including its planning model,
semantics, and interface definitions. We also present a city planning domain that is used to illustrate the
key concepts. In Section 4, we define three domain classes (EX1 – EX3) that constrain how separable
abstract tasks behave, particularly with respect to decomposition and backtracking. Section 5 analyzes
the backtracking behavior of HWSP under these domain classes. Section 6 presents the theoretical
results comparing HWSP and classical HTN planning.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related Work</title>
      <p>
        Early work by Korf [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] demonstrated that decomposing a planning problem into subproblems can
yield significant improvements in search eficiency. This idea was further developed by Knoblock [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ],
who introduced algorithms for automatically creating abstraction hierarchies and proved that solving
problems on multiple levels of abstraction can be exponentially more eficient than flat planning. The
concept of factored planning, introduced by Brafman and Domshlak [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ], also shows that problem
factorization can lead to exponential reductions in efort.
      </p>
      <p>
        Macro-operators have also been recognized as a powerful tool for improving HTN planning eficiency.
Korf’s seminal paper [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] on macro-operators showed that learning efective macro-operators can
reduce the efective solution depth and transform planning processes with exponential complexity into
polynomial ones. More recent work by Botea et al. [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] demonstrated the efectiveness of automatically
learned macro-operators in improving planning performance.
      </p>
      <p>
        Hierarchical reinforcement learning (HRL) has also been explored as a means of improving planning
performance. Dietterich’s foundational paper from 2000 introduced the MAXQ framework [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ], which
formalizes how a Markov Decision Process (MDP) can be decomposed into a temporal hierarchy of
subtasks. More recent work by Kulkarni et al. [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] established that HRL can achieve state-of-the-art
performance on complex game environments.
      </p>
      <p>
        Backtracking and planning dynamics in HTNs have also been extensively studied. Nau et al.’s
work [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] examined the impact of task ordering on HTN search, while Olz and Bercher [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] introduced
a look-ahead technique for reducing backtracking in HTN plan search. Commitment strategies in
HTN planning have also been explored by Tsuneto et al. [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ], who compared diferent strategies for
variable and method commitment. Theoretical analyses by Alford et al. [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ] and Geier and Bercher [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ]
have provided valuable insights into the complexity of HTN planning and the efects of task ordering
on search eficiency. McCluskey [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ] introduced EMS, which uses object transition sequences and
sort-based independence to enable local reasoning about how dynamic objects change state during
hierarchical planning.
      </p>
      <p>
        Beyond classical HTN planning, several foundational and conceptually related approaches have
addressed hierarchical abstraction. Sacerdoti’s ABSTRIPS system [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ] and Knoblock’s work on automatic
abstraction hierarchy construction [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ] pioneered world-state abstraction in planning, demonstrating
that solving abstract problems first can drastically reduce search. Angelic hierarchical planning [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ]
further introduced optimistic high-level actions with abstract efect summaries, similar to separable
abstract tasks but emphasizing cost bounds for optimization rather than multi-level abstraction for
scalability. Recent work has also explored compiling HTNs to SAT for optimal search [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ], defining
standard hierarchical languages like HDDL [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ], and integrating HTN planning into competitive
evaluations via the IPC track.
      </p>
    </sec>
    <sec id="sec-3">
      <title>3. Framework Definitions</title>
      <p>
        This Section describes the Hierarchical World State Planning (HWSP) algorithm, which extends the
traditional planning paradigm by introducing a multi-level world state [
        <xref ref-type="bibr" rid="ref1 ref20">20, 1</xref>
        ]. The HWSP algorithm is
based on the concept of separable abstract tasks, which enable more eficient planning by partitioning
the planning process into smaller sub-processes across multiple levels of abstraction. The following
version of the HWSP algorithm is a more formal one compared to the original paper [
        <xref ref-type="bibr" rid="ref21 ref22">21, 22</xref>
        ]. In this
framework, we assume a progressive planning strategy in which decomposition proceeds front-to-back:
tasks are handled in their plan order, and the world state is updated strictly forward from the initial
state.
      </p>
      <p>Let C, V , P and A denote the set of all constants, variables, predicates and atoms, respectively. A
literal is an atom or its negation, while an atom is a predicate applied to a tuple of terms. Here, a term
may be a constant or a variable. Every predicate p belongs to the set P .</p>
      <sec id="sec-3-1">
        <title>3.1. Example: City Planning</title>
        <p>We introduce a running example to help illustrate the concepts of this paper. This domain models a
hierarchical city planning problem with three levels of detail (abstraction): City, Quarter, and Building.
Streets are not explicitly modeled. Each level consists of a 2D grid of square cells. A quarter consists of a
4× 4 block of building cells, and a city consists of a 4× 4 block of quarters – forming a non-overlapping,
nested hierarchy (see Figure 1).</p>
        <p>Some cells may be blocked, preventing construction; this is represented by the predicate blocked(x),
which holds in the initial state for blocked cells. Each level contains objects that can be constructed via
tasks:
• City Level: small-city, big-city
• Quarter Level: industrial-quarter, residential-quarter, park
• Building Level: living-house, playground, football-field, powerplant, factory</p>
        <p>To demonstrate the algorithm’s ability to handle diferent domain classes, we can modify this base
domain in three key ways. First, we can introduce blocked cells, which may cause tasks to fail and
lead to diferent planning outcomes. Second, we can enable electricity constraints, where a single
powerplant – placed in any quarter – can power the entire city. Third, power propagation is expressed
explicitly through the world state: a quarter is considered powered if its city contains a powerplant,
captured by a specific atom in the world state.</p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Planning Domain</title>
        <p>A planning domain is denoted as D = (Ta, Tp, M, L, Ld, Dder, Φ) where Ta is the set of all abstract
tasks (including both standard and separable tasks), TS ⊂ T a denotes the set of separable abstract tasks,
Tp is the set of primitive tasks, M is the set of methods, and L ∈ N the number of detail levels (i.e.,
abstraction levels) present in the domain. The detail level function Ld : P ∪ Ta ∪ Tp → N assigns each
predicate or task a detail level ℓ. The derived predicates are defined in Dder, with their corresponding
logical formulas specified in Φ.</p>
        <p>Both abstract and primitive tasks are tuples in the form t(τ ) = ⟨prec t(τ ), eff t(τ )⟩ , where each task
t(τ ) ∈ T a ∪ Tp has a precondition prect(τ ) and an efect eff t(τ ) . For separable abstract tasks t ∈ TS ,
we write eff t(τ ) = meff t(τ ) ∪ peff t(τ ) where meff t are mandatory efects and peff t are predicted
(non-guaranteed) efects. In the planning process, these are predicted by the prediction function. A
plan step is a uniquely labeled task l : t(τ ).</p>
        <p>Example The task residential-quarter has mandatory efects meff t such as “provides housing capacity”
(which must be achieved for the quarter to be valid). It could have predicted efects peff t such as
“increases local property values”.</p>
        <p>A method is a tuple m = ⟨ta(τ m), Pm⟩, where ta(τ m) is the abstract task it can decompose, and Pm
is a series of plan steps (total-ordered), with τ m as the parameters of the method.</p>
        <p>A state s is an element of P(A), where P(A) denotes the power set of A. We require that a
separable abstract task tS at detail level ℓ decomposes only into tasks at detail level ℓ + 1, while
other tasks decompose at the same level ℓ. Hence, separable tasks must occur at levels ℓ &lt; L. Let
s ↓ ℓ = {a ∈ s|Ld(a) = ℓ} denote the projection of the state s to the detail level ℓ.
Example In our domain, we have L = 3 detail levels which includes city (level 1), quarter (level 2),
and building (level 3). Separable abstract tasks TS include small-city, big-city at level 1, and
residentialquarter, industrial-quarter, park at level 2. The restrictions on the methods regarding levels means that
a city task can only create quarters not buildings directly.</p>
        <sec id="sec-3-2-1">
          <title>City</title>
          <p>(Level 1)</p>
        </sec>
        <sec id="sec-3-2-2">
          <title>Quarter (Level 2)</title>
          <p>Available Tasks:
- Create Big City
- Create Small City
Available Tasks:
- Create Industrial</p>
        </sec>
        <sec id="sec-3-2-3">
          <title>Quarter</title>
          <p>- Create Residential</p>
        </sec>
        <sec id="sec-3-2-4">
          <title>Quarter</title>
        </sec>
        <sec id="sec-3-2-5">
          <title>Block of</title>
        </sec>
        <sec id="sec-3-2-6">
          <title>Buildings (Level 3)</title>
          <p>Available Tasks:
- Create Living</p>
        </sec>
        <sec id="sec-3-2-7">
          <title>House</title>
          <p>- Create</p>
        </sec>
        <sec id="sec-3-2-8">
          <title>Powerplant - Create Factory</title>
          <p>
            Let Dder ⊂ P be the set of derived predicates [
            <xref ref-type="bibr" rid="ref23">23</xref>
            ], which includes all predicates that are not on the
most concrete (non-abstract) detail level. For each derived predicate Pder ∈ Dder at detail level ℓ, there
exists a logical formula ϕ Pder ∈ Φ involving only predicates from the next detail level (ℓ + 1). Formally:
Pder(x1, ..., xn) ≡ ϕ
          </p>
          <p>Pder (pℓ+1(x1), ..., pℓ+1(xn))
where:
• pℓ+1 represents a predicate at detail level ℓ + 1,
• x1, ..., xn are variables representing objects or entities in the domain,
• ϕ Pder is a first-order logic formula that combines the predicates from detail level ℓ + 1 using logical
operators (e.g., ∧, ∨, ¬).</p>
          <p>Each derived predicate Pder at detail level ℓ can only depend on predicates that are strictly at level ℓ + 1
(this is more restricted than in PDDL 2.2). Furthermore, no cycles are allowed, even across multiple
predicates.</p>
          <p>Example Derived predicates in our examples are the predicates on the city and quarter level. On the
quarter level has_power(quarter) is in a world state if the corresponding sub-rectangle at the next detail
level (building level) contains a building classified as a power plant.</p>
          <p>We restrict our attention to domains with a finite decomposition depth. The decomposition depth
represents the longest chain of tasks achievable through method application. The maximum
decomposition depth in any domain D, denoted Dmax(D), is assumed to be less than or equal to ∆ . Every
domain considered in this paper satisfies Dmax(D) ≤ ∆ for a fixed but arbitrary natural number ∆ . A
planning problem is defined as a pair Π = (s 0, t0), where s0 ∈ P (A) is the initial world state, and
t0 ∈ Ta is the initial abstract task to be decomposed.</p>
          <p>
            To enable eficient decomposition, HWSP incorporates predictor functions [
            <xref ref-type="bibr" rid="ref1">1</xref>
            ] that estimate separable
abstract task outcomes.
          </p>
          <p>Definition 1 (Predictor functions). HWSP can use learned estimators that operate on separable abstract
tasks. Given the current detail level ℓ state s ↓ ℓ and a separable task tS ∈ TS the predictor function returns
π(s, t S ) =</p>
          <p>ε(s, tS ), σ(s, t S ) ∈ P (A) × {true, false},
where
where
• ε : P(A) × T S → P(A) is an efect predictor that proposes a set of non-mandatory efects (an
estimate for peff tS ), and
• σ : P(A) × T S → {true, false} is the decomposition predictor whose output is to decide in a
planning process if a separable abstract task can be decomposed or not.</p>
          <p>
            The predictor can be fixed-efect rules, Conv2D, ResNet or any other approximation function. Predictor
functions in HWSP were first introduced and formally described by Staud [
            <xref ref-type="bibr" rid="ref1">1</xref>
            ].
3.2.1. HTN plans with labelled occurrences
Definition 2 (HTN plan with decomposition tree). A plan (or partial plan) is a tuple
          </p>
          <p>Q = V, E, r, λ T , λ M , ≺
• (V, E, r) is a finite rooted tree whose root is r ∈ V ,
• every task node v ∈ VT ⊆ V is labeled by λ T (v) ∈ Ta ∪ Tp
• every method node v ∈ VM ⊆ V (the remaining vertices) is labeled by a method λ M (v) ∈ M , and
• a strict total order ≺ over the task nodes, which is consistent with the child order of each method
node.</p>
          <p>Edges alternate strictly between task and method nodes: if (u, v) ∈ E and u is a task node, then v is a
method node and λ M (v) must be a method for λ T (u); conversely, all children of a method node are task
nodes that constitute the method plan steps.</p>
          <p>Definition 3 (Separable task predicate). is_sep(Q, v) :=
λ T (v) ∈ TS .</p>
          <p>Let state_at(Q, v) be the world state obtained by applying all efects of task nodes w ≺ v that are
fully decomposed to the initial state s0. By Definition 7, all task nodes w such that w ≺ v must be fully
decomposed before v is considered for decomposition. Hence, the cumulative world state state_at(Q, v)
is well-defined and includes only efects of primitive tasks. We define:
Definition 4 (Decomposability Test). Let Q be a plan and v ∈ VT a task node with label λ T (v) = tτ ()¯ .
Let s := state_at(Q, v). Then:</p>
          <p>can_decompose(Q, v) := true
if there exists a sequence of method applications starting from tτ ()¯ such that: the precondition of every
task in the resulting decomposition tree holds in the state at which it is applied, the resulting decomposition
tree contains only primitive tasks at the leaves, and all relevant depth and abstraction-level constraints are
satisfied.</p>
          <p>Sibling-rewriting relation For two plans Q, Q′ and a separable abstract task occurrence s ∈ VT
of Q we write Q ⇝ s Q′ if Q′ is obtained from Q by only replacing the subtree rooted at s; all other
subtrees (including the one currently under investigation) are left unchanged.</p>
          <p>Definition 5 (Prior-task predicate). For any two task occurrences v, w ∈ V T in a plan Q, define
prior(Q, v, w) := v ≺ · · · ≺ w
where ≺ is the total order over task nodes that are part of the plan and ≺ · · · ≺ means that there exists a
chain of order relations. In other words, prior(Q, v, w) is true if the execution of Q executes v (or some
descendant of v) before w.</p>
          <p>We now define method decomposition under the constraint that planning is progressive. So each plan
is guaranteed to have been decomposed in such a way that the current world state can be computed
from the initial world state:
Definition 6 (Fully Decomposed Task). Let Q = ⟨V, E, r, λ T , λ M , ≺⟩ be a plan, and let v ∈ VT be a
task node. We say that v is fully decomposed in Q if all leaf task nodes in its descendant subtree rooted at
v are primitive tasks, i.e.,</p>
          <p>is_decomposed(Q, v) := ∀u ∈ descendantsQ(v) ∩ VT , λ T (u) ∈ Tp.</p>
          <p>Definition 7 (Progressive Method Decomposition). Let Q = ⟨V, E, r, λ T , λ M , ≺⟩ be a plan and u ∈ VM
a method node with parent task tτ ()¯ = λ T (parent(u)). The function decompose(Q, u) returns a new
plan Q′ only if:
1. All task nodes w ∈ VT such that w ≺ tτ ()¯
i.e.:</p>
          <p>(i.e., tasks prior to the parent of u) are fully decomposed,
∀w ∈ VT , (prior(Q, w, tτ ())¯ → is_decomposed(Q, w)).</p>
          <p>2. The siblings of parent(u) that appear before parent(u) in ≺ are also fully decomposed.
This condition ensures that the planner progresses only when all previous tasks are fully resolved into
executable steps. The decomposition proceeds as follows:
1. Add fresh task nodes s1, . . . , sm as children of u (ordered left-to-right).
2. Update ≺ to replace the parent task’s position with s1 ≺ · · · ≺ s m in a way that preserves the
ordering relative to parent(u)’s siblings.</p>
          <p>3. If u already has children (idempotence), the plan is unchanged.</p>
          <p>Otherwise, the decomposition is not allowed.</p>
          <p>The above definitions will be used in subsequent proofs.</p>
        </sec>
      </sec>
      <sec id="sec-3-3">
        <title>3.3. Planning Process</title>
        <p>
          The most important diference between HTN planning and HWSP is the planning process [
          <xref ref-type="bibr" rid="ref1">1</xref>
          ]. In HTN
planning, we decompose abstract tasks until only primitive tasks remain. In HWSP, decomposition
stops at separable abstract tasks, unless we are on the most detailed level. Instead of decomposing these
further, the planner uses a predictor function to estimate their efects and proceeds as if they were
primitive tasks at that level. This allows the planner to generate a top-level plan composed of separable
abstract tasks, which are expected to work under typical conditions.
        </p>
        <p>For each separable abstract task, a new planning process is launched to generate a subplan that
achieves its mandatory efects meff t ⊆ D der. As the efects are derived predicates, the actual goal is the
union of their logical formulas ϕ meff t . Backtracking is only necessary if this subplan fails. To maintain
a consistent world state, planning processes emit one separable abstract task at a time. Informally, each
task emitted is treated as a final decision and is only generated after confirming that a (local) plan
exists to achieve the required efects. While separable abstract tasks are semantically equivalent to
regular abstract tasks, they are treated diferently by the planner. Each planning process receives the
local world state, s ↓ (ℓ + 1) (projected to the next detail level) as input, and its goal is to achieve the
mandatory efects meff t of the separable abstract task that spawned it.</p>
        <p>Definition 8 (Planning Process Solution). A solution for planning process at level ℓ is a plan Q where
every leaf node is either: (1) a primitive task (λ T (v) ∈ Tp) if ℓ + 1 = L, or (2) a separable abstract task
(λ T (v) ∈ TS ) if ℓ + 1 &lt; L. All preconditions of separable abstract tasks, treated as primitive at this level,
hold during sequential execution from the initial state.</p>
        <p>Definition 9 (Planning Process Interface). A planning process (PP) is a 5-tuple (D, I, ℓ, I, F ) where:
• D = (Ta, Tp, M, L) is the planning domain
• I ∈ IS is the internal state of the planning process. IS is the set of all internal states and this depends
on the concrete implementation.
• ℓ ∈ {1, ..., L} is the current detail level
• I : IS × P(A) → I S × (T a ∪ TS ∪ Tp ∪ {null})</p>
        <p>Given the current internal state and the world state, this function returns the updated internal state
and either the next task to be added to the plan, or null to indicate failure.
• F : IS × P(A) × T S → IS × N</p>
        <p>Handles the failure of a previously generated separable task t ∈ TS under the given world state.
Returns the updated internal state and the number of previously generated tasks to retract (i.e., undo).</p>
        <p>The next call to I will generate a new alternative task.</p>
        <p>Internally, depending on the implementation, F will trigger backtracking or replanning. After F is
called, the next call to I will return either a new task to replace the failed one or null if replanning
fails entirely.</p>
        <p>Implementation Notes.</p>
        <p>
          This interface can be realized through diferent search strategies:
• MCTS-based: Uses Monte Carlo tree search where backtracking simply means that we go back
and select a task that has a smaller visit count than the discarded one.
• HTN with repair: Rolls back the currently failed plan, discards the failed task, and then attempts
to complete a new plan by reusing parts of the previously failed one. [
          <xref ref-type="bibr" rid="ref24">24</xref>
          ].
        </p>
        <p>Example In the city example, a planning process generates a plan to construct the buildings at the
highest detail level 3 when given a quarter task as input.</p>
        <p>Relationship to classical HTN planning. If the domain provides only one level (L = 1) and contains
no separable abstract tasks, HWSP degenerates into standard HTN planning: the single PP is invoked
exactly once, no spawning occurs, and the algorithm behaves identically to a flat HTN search.</p>
      </sec>
      <sec id="sec-3-4">
        <title>3.4. Planning Process Management</title>
        <p>The HWSP algorithm manages planning processes through a stack-based approach. The operational
lfow proceeds as follows:
1. Creation: When encountering a separable abstract task, a new planning process is created at the
next detail level, with the efects of the separable task as its goals.
2. Execution: Only the topmost process on the stack is active, producing tasks that are either added
to the main plan (if primitive) or used to spawn new processes (if separable abstract task).
3. Termination: A process terminates upon achieving its goals or failing. On success, its parent
process resumes; on failure, backtracking occurs.</p>
        <p>The main planning algorithm initializes with the first planning process at detail level one, an empty
plan, and the initial world state. While the stack contains processes, the algorithm retrieves the next
task from the topmost process. If the process is finished, it is removed from the stack. If a process
fails to produce a task, the algorithm backtracks according to the domain class rules. For primitive
tasks, the algorithm adds them to the plan and updates the world state with their efects. For separable
abstract tasks, it creates a new planning process and pushes it onto the stack. The algorithm terminates
successfully when the goal is achieved, or with failure when the stack becomes empty without achieving
the goal.</p>
        <p>Definition 10 (Overall Solution). A solution to the overall planning problem Π = (s 0, t0) is a plan Q
where: (1) every leaf node is a primitive task (λ T (v) ∈ Tp), and (2) all primitive task preconditions hold
during sequential execution from s0.</p>
        <p>Example In the city domain, the main planning algorithm would be coordinating the planning process
so that a complete city is built.</p>
        <p>As demonstrated in Figure 2, when a planning process encounters a failure, the backtracking behavior
is determined by the domain class.</p>
        <p>State projection restricts each planning process to atoms at the appropriate detail level, reducing
memory usage. The behavior of HWSP is further influenced by domain-specific properties, which we
classify into EX1–EX3 in the Section 4.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Domain Classes</title>
      <p>
        The performance of planning algorithms, such as Hierarchical Task Network (HTN) planning and
its extensions like Hierarchical World State Planning (HWSP), can be significantly influenced by the
structure of the planning domain. This observation aligns with the downward refinement property
introduced by Bacchus and Yang [
        <xref ref-type="bibr" rid="ref25">25</xref>
        ], which emphasizes how domain structure impacts the feasibility
and eficiency of hierarchical refinement. To formalize these influences in the context of separable
abstract tasks, we introduce a classification of domain classes based on task interaction and backtracking
behavior. In particular, we analyze how sibling tasks, parallel subtasks at the same abstraction level,
may or may not interfere with one another.
      </p>
      <p>Let’s consider our urban planning example again, where a city manager needs to plan the construction
of new buildings in diferent quarters of the city. To demonstrate the algorithm’s ability to handle
diverse domain scenarios, we modify this base domain in three key ways:
• First, we introduce blocked cells, which may cause tasks to fail and lead to diferent planning
outcomes.
• Second, we enable electricity constraints, where a single powerplant—placed in any quarter—can
power the entire city.
• Third, power propagation can be expressed using derived predicates Dder: a quarter is considered
powered if its city contains a powerplant, captured by has_power(quarter).</p>
      <p>These modifications highlight distinct domain classes, such as spatial constraints, resource allocation,
and hierarchical dependencies. We propose several domain classes, which we will describe using
illustrative examples related to placing buildings in a quarter:
1. EX1: Execution Guaranteed</p>
      <p>Intuitive Definition: Execution (decomposition of the separable abstract task) is guaranteed
once preconditions are fulfilled.</p>
      <p>Example: In our urban planning scenario, if a quarter plan fits (i.e., the preconditions of the
separable abstract task are fulfilled), its construction is guaranteed without any unforeseen
interference from other quarters or tasks.</p>
      <p>Definition 11 (Domain class EX1). A domain D is in EX1 if for every plan Q and every separable
task occurrence v ∈ VT of Q, if is_sep(Q, v) ∧ preconditions of λ T (v) are satisfied in state_at(Q, v),
then can_decompose(Q, v) = true.
2. EX2: Execution Can Fail but Won’t Benefit from Backtracking into Siblings
Intuitive Definition: Execution can fail, but a diferent solution from another planning process
on the same detail level won’t help.</p>
      <p>Example: In a more realistic urban planning scenario, parts of the land within a quarter might
not be buildable due to unforeseen reasons (e.g., hidden environmental hazards). As a result,
the construction of a specific building – and possibly the entire quarter – may fail, even if
all preconditions appeared to be satisfied. However, changing the plan for a diferent quarter
(i.e., a sibling task) will not influence the outcome. If such a failure occurs, the planner must
backtrack to the parent task (e.g., the city-level plan) and select a new decomposition that avoids
the problematic quarter altogether. The key EX2 property is that the decomposability of one
separable abstract task (e.g., a quarter) is independent of the concrete plans chosen for earlier or
parallel sibling tasks.</p>
      <p>Definition 12 (Domain class EX2). D is in EX2 if for all plans Q, separable occurrences v ∈ VT and
for all sibling abstract task occurrences s with the same parent as v in the plan’s strict total order ≺ over
task nodes with Ld(s) = Ld(v), is_sep(Q, s)∧is_sep(Q, v)∧Q ⇝ s Q′ =⇒ can_decompose(Q, v)
= can_decompose(Q′, v).</p>
      <p>In words, replacing the internal subtree of an earlier separable sibling s does not afect the
decomposability of a later separable task at the same detail level.
3. EX3: Execution Can Fail and May Benefit from Backtracking</p>
      <p>Intuitive Definition: Execution can fail, and another solution from a prior planning process
might help.</p>
      <p>Example: In a complex urban planning scenario, the placement of a building in one quarter could
influence the feasibility of constructing buildings in adjacent quarters due to shared resources or
spatial constraints. If the construction of a building fails because there is no power, creating an
industrial quarter with a power plant prior will help.</p>
      <p>Definition 13 (Domain class EX3). The class EX3 includes all planning domains.</p>
      <p>It represents the general case where no specific assumptions about separable task independence
are made. Note that introducing derived predicates such as has_power(quarter) can explicitly
model task dependencies, allowing EX3 domains to be reclassified as EX2 if sibling tasks become
independent under these conditions.</p>
      <p>Separable abstract tasks isolate planning to local islands of the world state. How strongly such islands
interact determines how much global backtracking remains necessary. We capture this idea by three
domain classes, EX1 – EX3, that impose increasingly weak independence assumptions on separable
tasks.</p>
      <p>Proposition 1 (Independence chain). EX1 ⊂ EX2 ⊂ EX3.</p>
      <p>Proof. Immediate from the definitions: EX1 implies the EX2 invariance because the antecedent of
Def. 12 is stronger; conversely any counter-example to EX2 is by definition a witness for EX3.</p>
    </sec>
    <sec id="sec-5">
      <title>5. Backtracking</title>
      <p>In Hierarchical World State Planning (HWSP), backtracking plays a pivotal role due to its unique
approach to task decomposition and execution. While HWSP shares similarities with traditional
Hierarchical Task Network (HTN) planning, its backtracking mechanism in the main planning system
is distinct.</p>
      <p>In HWSP, backtracking occurs when a planning process fails to find a viable solution. At this point,
the algorithm must reconsider its previous decisions and explore alternative paths. This is where the
domain classes (EX1, EX2, and EX3) come into play, as they significantly influence the backtracking
behavior.</p>
      <p>• EX1: In domains where execution is guaranteed once preconditions of a separable abstract task
are met, backtracking is relatively straightforward. Since each task’s success is assured, the main
planner never needs to backtrack.
• EX2: In EX2 domains, the failure of a separable abstract task requires specific backtracking
behavior. Let us analyze why EX2 leads to this behavior through a logical proof.</p>
      <p>Proposition 2 (EX2 Backtracking Behavior). In an EX2 domain, if a separable abstract task v fails
during decomposition, the planner must backtrack to the parent planning process without considering
siblings of v.</p>
      <p>Layer 1
Layer 2</p>
      <p>Planning Processes
small-city</p>
      <p>big-city
industrial. residential.</p>
      <p>park
residential
Layer 3
power. factory
living living
playg. football</p>
      <p>playg. living
Backtrack to big-city without
replanning park (EX2 property)
Main Plan
power. factory living living playg. football playg. living
Progressive World State
&lt;Set of atoms&gt;
! failure because
preconditions
for living-house
are not fullfilled</p>
      <p>Proof. By Definition 12, for any plan Q and separable task v, modifying any sibling s
(with s ≺ v ) does not afect the decomposability of v. Formally: can_decompose(Q, v) =
can_decompose(Q′, v) where Q ⇝ s Q′.</p>
      <p>
        If v fails (can_decompose(Q, v) = false), altering any prior sibling s through backtracking
cannot make can_decompose(Q′, v) = true. Thus, the only recourse is to inform the parent
process of v’s failure. The parent must then generate a new separable abstract task to replace v or
one of its siblings. If the parent exhausts all alternatives and fails, the failure propagates upward
recursively. This eliminates the need to explore siblings of v, as their modifications cannot resolve
v’s failure under EX2’s independence. The algorithm remains sound and complete [
        <xref ref-type="bibr" rid="ref20">20</xref>
        ].
• EX3: In these domains, where execution can fail and backtracking may help, we must perform
backtracking as in the original HWSP algorithm [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ].
      </p>
    </sec>
    <sec id="sec-6">
      <title>6. Results</title>
      <p>
        This section establishes formal upper bounds on the search efort of Hierarchical World-State Planning
(HWSP) when executed on the three domain classes EX1 ⊂ EX2 ⊂ EX3 and contrasts those bounds with
classical HTN planning. All logarithms are base 2; plan-existence complexities follow the taxonomy of
Erol et al. [
        <xref ref-type="bibr" rid="ref22">22</xref>
        ].
      </p>
      <sec id="sec-6-1">
        <title>6.1. Complexity model</title>
        <p>Consider a fixed detail level ℓ. Let nℓ denote the number of sibling separable task occurrences emitted
at that level, and let each sibling have at most kℓ decomposable methods. The refinement of one sibling
induces a local search tree of branching factor bℓ+1 and depth hℓ+1; we abbreviate Bℓ+1 = bℓh+ℓ+11 .</p>
        <p>Let N denote the total input size (i.e., the size of the domain and the initial task network). We assume
that nℓ = O(N ); that is, the number of sibling tasks at each level grows at most linearly with the input.
The analysis that follows is worst-case and tight: for every bound, we can construct an infinite family
of instances whose planners visit exactly the stated number of states.</p>
        <p>We analyze the number of states visited during search, denoted as:
• SHTN: states visited by classical HTN planning
• SEX1, SEX2, SEX3: states visited by HWSP in the respective domain classes</p>
      </sec>
      <sec id="sec-6-2">
        <title>6.2. State-space bounds</title>
        <p>Theorem 3 (EX3 versus HTN). For any level ℓ</p>
        <p>SHTN = Θ
SHTN = Bℓn+ℓ1, SEX3 = nℓ kℓnℓ Bℓ+1, SEX3
Bℓn+ℓ1−1 !
nℓ kℓnℓ
Proof. Classical HTN planning inlines the local search of each sibling.1 Consequently the global search
tree contains Bℓ+1 possibilities for the first sibling, Bℓ+1 for the second, and so on, for a total of Bℓn+ℓ1
states.</p>
        <p>In EX3, the parent process still explores every combination of high-level alternatives, hence kℓnℓ
parent choices. However, once such a combination has been chosen, each sibling is solved independently
by its own local search; the parent never needs to interleave those searches. The resulting state count is
therefore nℓ kℓnℓ Bℓ+1. Taking the quotient of the two expressions yields the claimed factor.
Theorem 4 (EX2 versus EX3). For any level ℓ</p>
        <p>SEX2 = nℓ kℓ Bℓ+1,</p>
        <p>SEX3 = Θ
SEX2
knℓ−1 !
ℓ
nℓ
Proof. Definition 12 states that substituting an earlier sibling by an alternative decomposition never
changes the decomposability of a later sibling. Hence, when a sibling fails, the planner backtracks
exactly one level up, replaces only that sibling, and leaves all others untouched. Worst case it will test
every one of the kℓ methods before a success (or final failure) is found. Since there are nℓ siblings, at
most nℓ kℓ parent alternatives are generated, each of which spawns one local search of size Bℓ+1. The
product yields SEX2. Dividing by SEX3 establishes the gap.</p>
        <p>The bounds given in Theorems 3 and 4 are tight in the following sense: for each setting of nℓ, kℓ, and
Bℓ+1, there exists a family of domains and initial task networks for which the number of visited states
matches the bound asymptotically, i.e., S = Θ(·).Θ(S) states.</p>
        <p>HTN Tightness Example. Let the initial task be a sequence of nℓ abstract tasks, each with kℓ methods,
where each method expands to a local plan space of size Bℓ+1 (e.g., a binary tree of depth log2 Bℓ+1). If
no heuristic guidance is used, the planner must exhaustively search all Bℓn+ℓ 1 combinations.
EX3 Tightness Example. Consider a domain with nℓ sibling tasks (e.g., city quarters). Each has
kℓ decomposition methods, exactly one of which builds a power plant. Suppose that every quarter
requires electricity to decompose, and power flows only from left to right: quarter qj can be decomposed
only if some earlier sibling qi with i &lt; j contains a power plant. This induces a causal dependency
chain across siblings: choosing non-power plant methods early may prevent later decompositions,
forcing backtracking and revision of earlier choices. In the worst case, it explores all kℓnℓ global method
assignments; for each one, it performs nℓ independent subsearches of size Bℓ+1. The total number
of visited states is therefore nℓ kℓnℓ Bℓ+1, which matches the upper bound for EX3. The cross-sibling
dependency violates the EX2 invariance condition (Definition 12), so this domain lies strictly in EX3.</p>
        <p>Thus, the upper bound for SEX3 is tight, and the factor of nℓ is necessary in all general EX3 domains
where sibling tasks do not share causal links.
1We assume classical HTN planning operates under the same restrictions as HWSP: acyclic networks, progressive
decomposition, and bounded depth, ensuring a fair comparison.</p>
        <p>EX2 Tightness Example. Construct a domain with nℓ separable sibling tasks (e.g., one per city
quarter), each having kℓ distinct task alternatives. For each sibling, only one of its kℓ alternatives is
decomposable, while the others are designed to fail due to local, hidden constraints (e.g., unobservable
predicates or infeasible layouts). These constraints are specific to each sibling and cannot be inferred
or afected by the other siblings’ choices. When a sibling task fails, the parent planning process must
replace it with another alternative, up to kℓ times in the worst case. Each attempted task triggers a local
planning process that explores a subsearch space of size Bℓ+1. Since all sibling tasks operate over disjoint
parts of the world state and have no shared predicates, changing one sibling’s decomposition does not
influence the decomposability of another—satisfying the EX2 invariance condition (Definition 12). The
total number of visited states in this construction is nℓ · k ℓ · B ℓ+1, matching the upper bound for EX2
domains and proving its tightness.</p>
        <p>On EX1. EX1 guarantees downward refinement for the root plan: once the preconditions of a separable
task are satisfied, its decomposition cannot fail. Inside the local planning process spawned for that task,
however, non-monotonic search (e.g. local backtracking) is still permitted. Therefore EX1 improves the
constant factors of Theorem 4 but does not change its asymptotic form. More precisely, the number of
search states in EX1 is SEX1 = nℓ Bℓ+1, as each of the nℓ separable tasks is successfully decomposed
on the first attempt, requiring a single local search of size B ℓ+1 per task.</p>
        <p>Corollary 5 (Strict inclusion chain). It holds SEX1 &lt; SEX2 &lt; SEX3 &lt; SHTN as each “&lt;” removes one
exponential factor in the input size N , assuming nℓ = O(N ).</p>
      </sec>
      <sec id="sec-6-3">
        <title>6.3. Connection to classical complexity theory</title>
        <p>
          Bacchus and Yang’s Downward Refinement Property (DRP) [
          <xref ref-type="bibr" rid="ref25">25</xref>
          ] formalizes a key structural criterion
for eficient hierarchical planning: once a high-level plan is found, it can be refined monotonically
without modifying earlier decisions. Our EX1 domain class satisfies DRP by construction, as successful
decomposition is guaranteed once preconditions are met.
        </p>
        <p>EX2 domains do not satisfy DRP in the strict sense, since the planner may backtrack and replace
earlier siblings. However, they enforce a weaker property we term internal decomposition invariance:
the feasibility of decomposing a task is invariant under changes to the internal decomposition of prior
siblings at the same abstraction level. This allows EX2 domains to retain many of the practical benefits
of DRP, such as reduced backtracking and localized refinement, but without full top-down monotonicity.</p>
        <p>
          Korf’s macro-operator theorem [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ] predicts an exponential-to-linear collapse in ideal DRP-compatible
settings, which applies fully to EX1 and partially to EX2. Erol et al. [
          <xref ref-type="bibr" rid="ref22">22</xref>
          ] showed that plan existence is
semi-decidable for unrestricted HTNs but becomes PSPACE-complete under acyclic, non-interleaved
constraints—the same assumptions that characterize our EX3 class. Structural restrictions like those in
EX2 further reduce the search space, pushing the complexity down to NP-complete or even as low as P
in well-structured domains.
        </p>
        <p>Here, N is the combined size of the initial task network and domain description; p and c are fixed
polynomials.</p>
      </sec>
      <sec id="sec-6-4">
        <title>6.4. Efect of a (fallible) binary predictor</title>
        <p>In HWSP every separable task tS comes with a binary decomposition predictor σ : P(A) × T S →
{true, false}. Intuitively, σ(s, t S) = true says that local planning ought to succeed.2 Mis-predictions
are the only reason why an EX2 planner ever needs to backtrack.</p>
        <p>Error model. For the sake of analysis we assume that, conditioned on the current state, the predictor
returns the correct answer with probability p (accuracy) and the wrong answer with probabilityp¯ = 1−p .
2We treat σ as a black box: it may be a learned classifier, a logical test or a mix thereof.</p>
        <p>Theorem 6 (Expected search efort with a fallible predictor). Let SE⋆X2 = nℓBℓ+1 be the number of
search states when the predictor is perfect. Assume that, each time a separable task is encountered, the
predictor is correct with probability p and wrong with probabilityp¯ = 1 − p (independently of previous
calls). If the predictor errs, the parent process tries a diferent alternative; at most kℓ − 1 further alternatives
exist.</p>
        <p>Then the expected number of visited states is bounded by E[SEX2(p)] ≤ S E⋆X2 1 +p¯(k ℓ − 1) .
Proof. Let nℓ be the number of separable task occurrences at level ℓ, and let each task require a local
search of size Bℓ+1 once a valid decomposition is selected. Under a perfect predictor, each task is
decomposed successfully on the first try, so the total number of visited states is: SE⋆X2 = nℓ · B ℓ+1.. Now
consider a fallible predictor with accuracy p, and letp¯ = 1 − p denote the probability of an incorrect
prediction. If the predictor fails (e.g., produces a false positive), the planner must backtrack and try a
diferent decomposition. Since each task has at most kℓ possible alternatives, the number of additional
alternatives to try after a misprediction is at most kℓ − 1.</p>
        <p>Thus, the expected number of task attempts for each separable task is at most: 1 · p + (1 + β) ·
p,¯ with β ≤ k ℓ − 1 . Substituting, we obtain the upper bound: Expected attempts per task ≤ 1 +
p¯ · β ≤ 1 +p¯ · (k ℓ − 1) . Multiplying this overhead by the base cost SE⋆X2 yields: E [SEX2(p)] ≤
SE⋆X2 · (1 +p¯ · (k ℓ − 1)) , which proves the claim.</p>
        <p>A perfect predictor collapses EX2 to EX1. Setting p = 1 makes the overhead term (1 +p¯β) equal
to 1, so every separable task is decomposed successfully on the very first try. This eliminates all global
backtracking and re-establishes the Downward-Refinement Property for the top-level plan – precisely
the hallmark of the EX1 class (Def. 11). Formally, the planner’s behavior becomes observationally
equivalent to running HWSP on an EX1 domain. □
Interpretation. Theorem 6 shows that the predictor’s accuracy enters the search complexity as a
linear scalar factor. Even modest accuracies (say p = 0.9) leave the Θ kℓnℓ−1 gap of Theorem 4 largely
intact, becausep¯ ≤ 0.1 . Only adversarially bad predictors (p → 0) would degrade EX2 back to EX3
behaviour.</p>
      </sec>
    </sec>
    <sec id="sec-7">
      <title>7. Conclusion</title>
      <p>We analyzed Hierarchical World State Planning (HWSP), a planning framework that extends classical
HTN planning through multi-level state abstraction and separable abstract tasks. While HWSP was
introduced in prior work, its theoretical properties and complexity characteristics had not been formally
studied.</p>
      <p>Our main contribution is a comprehensive theoretical analysis of HWSP’s planning dynamics. We
introduced a novel classification of planning domains, EX1, EX2, and EX3, based on structural constraints
on separable abstract tasks and their backtracking behavior. This classification enables a deeper
understanding of when HWSP achieves substantial reductions in search complexity compared to
classical HTN planning.</p>
    </sec>
    <sec id="sec-8">
      <title>Declaration on Generative AI</title>
      <p>During the preparation of this work, the author(s) used ChatGPT and LLaMA 3.3 70B in order to:
grammar and spelling check. After using these tools, the author(s) reviewed and edited the content as
needed and take full responsibility for the publication’s content.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>M.</given-names>
            <surname>Staud</surname>
          </string-name>
          ,
          <article-title>Integrating Deep Learning Techniques into Hierarchical Task Planning for Efect and Heuristic Predictions in 2D Domains</article-title>
          , in: HPLAN Workshop,
          <year>2023</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>R. E.</given-names>
            <surname>Korf</surname>
          </string-name>
          ,
          <article-title>Planning as Search: A Quantitative Approach</article-title>
          ,
          <source>AI</source>
          87
          <volume>33</volume>
          (
          <year>1987</year>
          )
          <fpage>65</fpage>
          -
          <lpage>88</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>C. A.</given-names>
            <surname>Knoblock</surname>
          </string-name>
          ,
          <article-title>Learning Abstraction Hierarchies for Problem Solving</article-title>
          , in: AAAI,
          <year>1990</year>
          , pp.
          <fpage>923</fpage>
          -
          <lpage>928</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>R. I.</given-names>
            <surname>Brafman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Domshlak</surname>
          </string-name>
          , Factored Planning: How, when, and When Not, in: AAAI, volume
          <volume>6</volume>
          ,
          <year>2006</year>
          , pp.
          <fpage>809</fpage>
          -
          <lpage>814</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>R. E.</given-names>
            <surname>Korf</surname>
          </string-name>
          ,
          <string-name>
            <surname>Macro-Operators</surname>
          </string-name>
          :
          <article-title>A Weak Method for Learning</article-title>
          ,
          <source>Artificial Intelligence</source>
          <volume>26</volume>
          (
          <year>1985</year>
          )
          <fpage>35</fpage>
          -
          <lpage>77</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>A.</given-names>
            <surname>Botea</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Enzenberger</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Müller</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Schaefer</surname>
          </string-name>
          , Macro-FF:
          <article-title>Improving AI Planning with Automatically Learned Macro-Operators</article-title>
          ,
          <source>Journal of Artificial Intelligence Research</source>
          <volume>24</volume>
          (
          <year>2005</year>
          )
          <fpage>581</fpage>
          -
          <lpage>621</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>T. G.</given-names>
            <surname>Dietterich</surname>
          </string-name>
          ,
          <article-title>Hierarchical Reinforcement Learning with the MAXQ Value Function Decomposition</article-title>
          ,
          <source>Journal of Artificial Intelligence Research</source>
          <volume>13</volume>
          (
          <year>2000</year>
          )
          <fpage>227</fpage>
          -
          <lpage>303</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>T. D.</given-names>
            <surname>Kulkarni</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Narasimhan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Saeedi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Tenenbaum</surname>
          </string-name>
          ,
          <article-title>Hierarchical deep reinforcement learning: Integrating temporal abstraction and intrinsic motivation</article-title>
          ,
          <source>Advances in Neural Information Processing Systems</source>
          <volume>29</volume>
          (
          <year>2016</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>D.</given-names>
            <surname>Nau</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Munoz-Avila</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Cao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Lotem</surname>
          </string-name>
          , S. Mitchell,
          <article-title>Total-Order Planning with Partially Ordered Subtasks</article-title>
          , in: IJCAI, volume
          <volume>1</volume>
          ,
          <year>2001</year>
          , pp.
          <fpage>425</fpage>
          -
          <lpage>430</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>C.</given-names>
            <surname>Olz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Bercher</surname>
          </string-name>
          ,
          <article-title>A Look-Ahead Technique for Search-Based HTN Planning: Reducing the Branching Factor by Identifying Inevitable Task Refinements</article-title>
          ,
          <source>in: Proceedings of the International Symposium on Combinatorial Search</source>
          , volume
          <volume>16</volume>
          ,
          <year>2023</year>
          , pp.
          <fpage>65</fpage>
          -
          <lpage>73</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>R.</given-names>
            <surname>Tsuneto</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Erol</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Hendler</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Nau</surname>
          </string-name>
          ,
          <article-title>Commitment strategies in hierarchical task network planning</article-title>
          ,
          <source>in: Proceedings of the National Conference on Artificial Intelligence (AAAI)</source>
          ,
          <year>1996</year>
          , pp.
          <fpage>536</fpage>
          -
          <lpage>542</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>R.</given-names>
            <surname>Alford</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Bercher</surname>
          </string-name>
          , D. Aha,
          <article-title>Tight Bounds for HTN Planning</article-title>
          ,
          <source>in: Proceedings of the International Conference on Automated Planning and Scheduling</source>
          , volume
          <volume>25</volume>
          ,
          <year>2015</year>
          , pp.
          <fpage>7</fpage>
          -
          <lpage>15</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>T.</given-names>
            <surname>Geier</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Bercher</surname>
          </string-name>
          ,
          <article-title>On the Decidability of HTN Planning with Task Insertion</article-title>
          , in: IJCAI,
          <year>2011</year>
          , pp.
          <fpage>1955</fpage>
          -
          <lpage>1961</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <surname>T. L. McCluskey</surname>
          </string-name>
          ,
          <article-title>Object Transition Sequences: A New Form of Abstraction for HTN Planners</article-title>
          , in: AIPS,
          <year>2000</year>
          , pp.
          <fpage>216</fpage>
          -
          <lpage>225</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>E. D.</given-names>
            <surname>Sacerdoti</surname>
          </string-name>
          ,
          <article-title>Planning in a Hierarchy of Abstraction Spaces</article-title>
          ,
          <source>Artificial intelligence 5</source>
          (
          <year>1974</year>
          )
          <fpage>115</fpage>
          -
          <lpage>135</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>C. A.</given-names>
            <surname>Knoblock</surname>
          </string-name>
          , Automatically Generating Abstractions for Planning,
          <source>Artificial Intelligence</source>
          <volume>68</volume>
          (
          <year>1994</year>
          )
          <fpage>243</fpage>
          -
          <lpage>302</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>B.</given-names>
            <surname>Marthi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Russell</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. A.</given-names>
            <surname>Wolfe</surname>
          </string-name>
          , Angelic Hierarchical Planning:
          <article-title>Optimal and Online Algorithms</article-title>
          , in: ICAPS,
          <year>2008</year>
          , pp.
          <fpage>222</fpage>
          -
          <lpage>231</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>G.</given-names>
            <surname>Behnke</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Höller</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Biundo</surname>
          </string-name>
          ,
          <article-title>Finding optimal solutions in htn planning-a sat-based approach</article-title>
          ., in: IJCAI,
          <year>2019</year>
          , pp.
          <fpage>5500</fpage>
          -
          <lpage>5508</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>D.</given-names>
            <surname>Höller</surname>
          </string-name>
          , G. Behnke,
          <string-name>
            <given-names>P.</given-names>
            <surname>Bercher</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Biundo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Fiorino</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Pellier</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Alford</surname>
          </string-name>
          ,
          <string-name>
            <surname>HDDL:</surname>
          </string-name>
          <article-title>An Extension to PDDL for Expressing Hierarchical Planning Problems</article-title>
          , in
          <source>: Proc. of the AAAI Conference on AI</source>
          , volume
          <volume>34</volume>
          ,
          <year>2020</year>
          , pp.
          <fpage>9883</fpage>
          -
          <lpage>9891</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>M.</given-names>
            <surname>Staud</surname>
          </string-name>
          , Urban Modeling via Hierarchical Task Network Planning,
          <source>in: HPLAN Workshop</source>
          ,
          <year>2022</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>M.</given-names>
            <surname>Ghallab</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Nau</surname>
          </string-name>
          , P. Traverso,
          <source>Automated Planning: Theory and Practice</source>
          , Elsevier,
          <year>2004</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>K.</given-names>
            <surname>Erol</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Hendler</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. S.</given-names>
            <surname>Nau</surname>
          </string-name>
          ,
          <article-title>Complexity Results for HTN Planning</article-title>
          ,
          <source>Annals of Mathematics and Artificial Intelligence</source>
          <volume>18</volume>
          (
          <year>1996</year>
          )
          <fpage>69</fpage>
          -
          <lpage>93</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <given-names>S.</given-names>
            <surname>Edelkamp</surname>
          </string-name>
          ,
          <source>J. Hofmann, PDDL 2</source>
          .
          <article-title>2: The Language for the Classical Part of IPC-4</article-title>
          , in: International Planning Competition,
          <year>2004</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [24]
          <string-name>
            <given-names>D.</given-names>
            <surname>Höller</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Bercher</surname>
          </string-name>
          , G. Behnke,
          <string-name>
            <surname>S.</surname>
          </string-name>
          <article-title>Biundo, HTN Plan Repair Using Unmodified Planning Systems</article-title>
          ,
          <source>in: HPLAN (ICAPS)</source>
          ,
          <year>2018</year>
          , pp.
          <fpage>26</fpage>
          -
          <lpage>30</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          [25]
          <string-name>
            <given-names>F.</given-names>
            <surname>Bacchus</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <article-title>The Downward Refinement Property</article-title>
          , in: IJCAI,
          <year>1991</year>
          , pp.
          <fpage>286</fpage>
          -
          <lpage>293</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>