<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Thinking in front of the box: Towards intelligent robotic action selection for navigation in complex environments using image-schematic reasoning</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Mihai Pomarlan</string-name>
          <email>pomarlan@uni-bremen.de</email>
          <xref ref-type="aff" rid="aff3">3</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Stefano De Giorgis</string-name>
          <email>stefano.degiorgis2@unibo.it</email>
          <xref ref-type="aff" rid="aff5">5</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Maria M. Hedblom</string-name>
          <email>maria.hedblom@ju.se</email>
          <xref ref-type="aff" rid="aff4">4</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Mohammed Diab</string-name>
          <email>m.diab@imperial.ac.uk</email>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Nikolaos Tsiogkas</string-name>
          <email>nikolaos.tsiogkas@kuleuven.be</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Core Lab ROB</institution>
          ,
          <addr-line>Flanders Make, Gaston Geenslaan 8, 3001 Heverlee</addr-line>
          ,
          <country country="BE">Belgium</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Department of Mechanical Engineering, KU Leuven</institution>
          ,
          <addr-line>Celestijnenlaan 300, B-3001 Heverlee (Leuven)</addr-line>
          ,
          <country country="BE">Belgium</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Imperial College London</institution>
          ,
          <addr-line>London</addr-line>
          ,
          <country country="UK">UK</country>
        </aff>
        <aff id="aff3">
          <label>3</label>
          <institution>Institute of Artificial Intelligence, University of Bremen</institution>
          ,
          <country country="DE">Germany</country>
        </aff>
        <aff id="aff4">
          <label>4</label>
          <institution>Jönköping Artificial Intelligence Laboratory, Jönköping University</institution>
          ,
          <country country="SE">Sweden</country>
        </aff>
        <aff id="aff5">
          <label>5</label>
          <institution>University of Bologna</institution>
          ,
          <addr-line>Via Zamboni 32, Bologna (BO)</addr-line>
          ,
          <country country="IT">Italy</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>One of the problems an agent faces when operating in a partially known, dynamic, sometimes unpredictable environment is to keep track of aspects of the world relevant to its task, and, if possible, restrict its attention to only these aspects. We present our first steps towards constructing a system that combines image schematic knowledge and reasoning with reactive robotics, and which enables perception that focuses on, and keeps track of, relevant entities and relationships. While our approach is more reasoning intensive than is usual in reactive robotics, the formalism we use for inference is fast and allows an agent to adjust, in real time, the complexity of its action selection procedures according to the complexity of the relevant part of the environment. We illustrate our approach with a few simulated examples of robots performing navigation tasks. In some examples, interaction with obstacles is necessary to complete the navigation tasks, adding complexity to the scenario.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>Among the problems that a robotic agent has to deal with, navigation seems to be one of the
most robustly solved. As an example, it was only in 2019 that, after 15 years of operation, the
Mars Rover finally was declared “dead” because of dust preventing its solar cells from operating.
Through all that time, it had to move autonomously on the martian surface; the communication
lag with Earth was too long to make teleoperation feasible.</p>
      <p>
        Clearly, algorithmic support exists to generate paths in environments with many obstacles.
Algorithms such as A*[
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] or its version adapted to changing environments D*[
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] can eficiently
search for shortest paths in (discretized) two-dimensional spaces. Randomized planning
algorithms such as RRT[
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] can handle spaces with more degrees of freedom in a “probabilistically
complete” fashion, eventually finding a path, if one exists, and in practice they find one quickly.
      </p>
      <p>
        However, what such algorithms address is the problem of finding a path in an environment
that is conceptualized as beyond the robot’s ability to change. Also, an understanding of
the environment in terms of its possibilities for action and change are beyond the scope of a
navigation module. This becomes important when path and trajectory planning and control
methods, such as RRT[
        <xref ref-type="bibr" rid="ref3">3</xref>
        ], A*[
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], TEB[
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] etc., fail to provide a solution. If there is no deeper
understanding of the reasons of the failure, there is little indication of how to overcome it.
      </p>
      <p>
        For instance, boxes could be pushed out of the way; other agents (humans or autonomous
agents) standing in the way might be asked to leave; and a box that was not a problem before
should be pushed away to make room for another and so on. Such actions may also have
consequences for other tasks the robot may be pursuing. To handle such situations, a robot
would rely on higher level planning systems, such as state transition planning systems in which
actions are treated as atomic components [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] described symbolically in something such as
Planning Domain Definition Language (PDDL) [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ].
      </p>
      <p>However, planning operations are time-ineficient with their complexity growing
exponentially in relation to the situation as both perception systems and action execution systems
require constant updating to changes in the environment and the agent’s own states. It is
important in such cases to keep track of the relevant data about the environment, and, if possible,
only the relevant data – as anything else is a potentially costly distraction. Further, in order
for a planning method to work it will need knowledge of both the environment as it is, and
of possibilities of action in general. In consequence, methods based on planning alone are
seldom enough to deal with the complex and dynamic real-world scenarios. For instance, in
a situation in which a box is blocking the planned trajectory, planning alone will not be able
to adapt to this change, unless the system has a deeper understanding of what this blockage
means, how it efectively can bypass it or move it, and, likely, it will also need to focus on a
comparatively small part of the environment for the planning search to be feasible. Thus, for
dificult situations extending over longer time periods or that require unexpected manipulations
of the environment, conventional planning methods will not be enough. At the same time, for
many routine operations where repeated perceptual search and drawing complex inferences
are unnecessary, the solution can instead be sought on the behavior level.</p>
      <p>
        In this paper, we explore how much decision-making power can be created exclusively from
the perception inputs and behavioral outputs of the robot, on which very simple inference
mechanisms operate. We do not claim a slower, deliberative layer that performs planning is
unnecessary. We claim that such a layer should be supported by a behavioral layer that keeps its
focus on relevant parts of an environment, where this relevance is decided by the task a robot
performs, and the situations it finds itself in. Our work is inspired by the reactive, behavior-based
robotics approach pioneered by Rodney Brooks [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] with two-fold novel contributions.
      </p>
      <p>
        Firstly, we employ inference techniques that, while tractable and much simpler than planning,
allow for more dynamic rewiring of behavior than what is traditionally seen in reactive robots.
In particular, our methods allow the robot to adjust the set of entities and relationships it
considers relevant in a given situation. Secondly, we build on recent work on the formalization
of image schemas [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] (sensori-motor patterns of embodied experience [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]) and encode knowledge
for robotic perception/attention reconfiguration. The hypothesis we pursue in the long term
is that the information present in the image-schematic knowledge will allow our system to
reconfigure “the rules” of environmental changes, as opposed to an approach that would
hardcode information about objects and their interactions.
      </p>
      <p>In this paper, we take a first step in this direction. We are here interested in how to
automatically detect potential disturbances to the nominal task execution from a robot. This requires
the robot to decide which elements of the environment are interesting based on the executed
task and accordingly monitor them. In short, what we are tackling in this paper is:
1. a robot navigation problem, namely identify elements of a certain “saliency” in the
navigation environment (paths, blockage, etc.) and ask questions about their status;
2. managing the action-selection loop, in this case restricted to selecting what perception
queries to perform;
3. cost-eficiency problem, where the method of inference is close to the behavior-level, to
enable the robot to respond quickly to changes in the environment;
4. commonsense reasoning problem, since the knowledge for inference is expressed in
image-schematic terms to assist in its reuse and adaptability of the robot.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Theoretical Foundation and Related Work</title>
      <sec id="sec-2-1">
        <title>2.1. Embodied Cognition and Image Schemas</title>
        <p>
          Embodied cognition is a commonly applied research paradigm for cognitive robotics (e.g. see
[
          <xref ref-type="bibr" rid="ref10">10</xref>
          ]). It proposes that all forms of cognition, and by formal extension, robotic behavior, can
be traced back to abstracted information in the embodied experiences [
          <xref ref-type="bibr" rid="ref11">11</xref>
          ]. Through repeated
exposure to particular spatiotemporal relationships between objects, agents, and environments,
salient features are extracted and formed into conceptual schemas that can be repeatedly reused
for analyzing similar scenarios. These abstracted sensori-motor patterns are called image
schemas and have been defined by Mandler as “...dynamic analog representations of spatial
relations and movements in space” [12, p. 591]. To this end, they have been applied to describe
information transfer in analogical reasoning [
          <xref ref-type="bibr" rid="ref13">13</xref>
          ] and to describe the cognition structure in
event conceptualization [
          <xref ref-type="bibr" rid="ref14">14</xref>
          ].
        </p>
        <p>
          While stemming from cognitive linguistics, image schemas have seen increased popularity to
solve issues surrounding symbol grounding and commonsense reasoning problems in artificial
intelligence [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ] and more interestingly, due to their relationship to afordances and functional
relationships [
          <xref ref-type="bibr" rid="ref15">15</xref>
          ], research to integrate them into cognitively-inspired robotics systems to
simulate more intelligently behaving robots (e.g. [
          <xref ref-type="bibr" rid="ref16 ref17">16, 17</xref>
          ]).
        </p>
        <p>
          Due to their interdisciplinary foundation, there exists no agreed-upon list of image schemas.
Likewise, the most optimal method to formally represent these abstract concepts in terms of their
internal taxonomy and relational meronymy remains uncertain (for examples see [
          <xref ref-type="bibr" rid="ref18 ref19 ref8">8, 18, 19</xref>
          ]).
        </p>
        <p>
          In this paper, we consider a minimal scenario, as exposed in detail in Sec. 4.3, namely a robot
moving on wheels from a starting point to some destination, which should be able to move
unless it encounters some blockage in the form of an obstacle (object) in front of it, entering
in a relation of contact with it. To represent this apparently simple scenario, from a cognitive
robotics perspective, it requires the introduction of some of the most relevant image schemas,
here presented together with their conceptual components:
Source_Path_Goal: movement from a source or starting point, following a series of
contiguous locations leading to a goal or endpoint [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ]. In its dynamic form, it is present in all
forms of movement of an entity [
          <xref ref-type="bibr" rid="ref20">20</xref>
          ].
        </p>
        <p>
          Contact: the physical relationship in which the surfaces of two entities are touching.
Support: Contact between two objects in the vertical dimension [
          <xref ref-type="bibr" rid="ref21">21</xref>
          ], assuming some
consequences related to the forces exerted by the above and below entities being such that the
supported entity does not fall.
        </p>
        <p>
          Blockage: the complex construct in which planned movement is prevented by an obstacle. It
can be described as a force vector encountering a barrier and then taking any number of
directions [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ].
        </p>
      </sec>
      <sec id="sec-2-2">
        <title>2.2. Task and Motion Planning</title>
        <p>Task and Motion Planning (TAMP) plans the tasks to be done by the robot, both at an abstract
level, considering only atomic actions, and at a geometric and/or control level that considers
how these actions are to be carried out in detail. TAMP techniques are used to adapt to the
actual state of the environment, and comply with constraints both at task and geometric levels.</p>
        <p>
          There are two dominant approaches in the task planning domain, one based on search
algorithms and the other based on knowledge and reasoning. The former mainly uses PDDL [
          <xref ref-type="bibr" rid="ref22">22</xref>
          ] to
describe the world. Although this way of description can easily handle tasks with many actions,
and integrate the geometric (motion) constraints [
          <xref ref-type="bibr" rid="ref23">23</xref>
          ], it makes a closed world assumption,
i.e., if some facts about the world are not known or change, a planner may not be able to find
a solution. This limitation means that robots are not able to begin a task until all objects in
the environment are known and the actions the robot can do on them are completely defined.
The latter, knowledge enabled approach, has emerged as a new domain of planning, aiming at
making the robot able to flexibly perform manipulation tasks like [
          <xref ref-type="bibr" rid="ref24">24</xref>
          ]. This approach can easily
integrate the knowledge from the environment and adapt the action to be done accordingly.
        </p>
        <p>The increasing emphasis on real-world applications has led researchers to develop algorithms
and systems that more closely match realistic planning problems in which manipulation skills
play a significant role. Manipulation problems are referred to as problems in which robots
handle objects using a set of primitives, e.g., pushing, picking, or placing. Due to task constraints,
the limitation of generic motion planning emerges, and the robot is required to displace objects
when there are no feasible solutions between two robot configurations. This requires TAMP
to deal with diferent types of robotic manipulation problems, ranging from single or multiple
collaborative mobile robots navigating among movable obstacles to complex higher-dimensional
table-top manipulation problems carried out by dual-arm robots or mobile manipulators.</p>
        <p>
          The key challenge in real scenarios that include manipulation skills is to make the robot
adaptive to the changes that happen in the world. This requires a sophisticated system that
updates the state of the entities in the environment and feed the planning system with the
changes that occur. In this line, researchers have investigated how to keep updating the state
of the environment. A manipulation planning framework with perception capability has been
proposed [
          <xref ref-type="bibr" rid="ref25">25</xref>
          ] and [26]. The former integrates a multi-modal sensory system to infer the
line-of-sight, and non-line-of-sight objects and store the records in the database as experiential
knowledge to adapt the robot behaviour in similar situations based on the new states of the
world. The latter optimizes over Cartesian frames defined relative to target objects. The resulting
plan remains valid even if the objects are moving and can be executed by reactive controllers
that adapt to these changes in real-time. In [27] learning manipulation skills from a single
demonstration are proposed where a robot is shown a manipulation skill once and should then
use only a few trials on its own to learn to reproduce, optimize, and generalize that same skill.
        </p>
      </sec>
      <sec id="sec-2-3">
        <title>2.3. Situational awareness and assessment</title>
        <p>For a robot to successfully operate in the world and achieve its goals, a general understanding of
the surrounding environment and the dynamics of the processes taking place there, are required.
In a similar fashion, humans that execute tasks in the real world, are using cognitive processes
to form an understanding of their surroundings that allows them to take appropriate actions
towards their goals. This task-specific knowledge and understanding is formally referred to
as situation awareness (SAW) [28], and is studied in the scope of understanding how humans
perform tasks so that accidents can be minimized. It is divided in three levels:
Level 1: Refers to the perception of elements from the environment that are relevant to the
executed task.</p>
        <p>Level 2: Refers to the comprehension of the current situation based on the perceived elements,
and any relations that are formed among them.</p>
        <p>Level 3: Refers to the projection of the current situation to the future, allowing the individual
to make predictions regarding the possible evolution of it.</p>
        <p>Given the situation awareness, an individual can plan its actions accordingly, so that they
achieve their goals. It must be noted that the three levels of situation awareness are not forming
a linear relationship, i.e. a complete knowledge of one is not required to form the other. On
the contrary, they all work in parallel, continuously updated, as the situation evolves. Once
the individual needs to make a decision, they can use the current best estimate of each level to
guide their decision process.</p>
        <p>To achieve situation awareness, an active process of gathering knowledge is required. Such
an active process is referred as situation assessment (SAS). During situation assessment, an
individual focuses their attention towards the elements of the environment that are important for
the task at hand. In addition, they make the required connections between perceived elements
given background knowledge. In an analogous way, an artificial agent, such as a robot, can
assess the situation given its task, by using sensors to measure the environment, and algorithms
to process the sensor data.</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Problem setting</title>
      <p>In this paper we focus only on reconfiguring an agent’s perception, as opposed to aslo its actions.
Our setup then is that a “monitoring” agent, implementing our image-schematic approach,
watches over a collection of robots each navigating to its own goal. The robots do not interact
with the monitoring agent. Even so, the agent will keep track of what is relevant for navigation,
and seek to form an understanding of the relevant relationships between objects such that it
can foresee and suggest solutions for problems in navigation.</p>
      <sec id="sec-3-1">
        <title>3.1. Competency Questions</title>
        <p>The operation of our monitoring agent involves reasoning to direct its perceptive attention
and update its understanding of the world. To develop this reasoning procedure, we have first
defined a set of competency questions it must address. As part of its behavior, a robot must
focus its perception on aspects of the environment that are plausibly important (or, “relevant”)
for the achievement of its goals. Or, in the case of our monitoring agent, it should be able to take
the perspective of the robots it monitors. Following are the competency questions we defined:</p>
        <sec id="sec-3-1-1">
          <title>1. which entities are directly involved in the robot’s goals? 2. what relevant relations exist between entities directly involved in the robot’s goals? 3. what are other relevant entities and relationships in the environment? 4. are there entities/relationships that cease to be relevant?</title>
          <p>Questions 1 and 2 are there to provide a basic understanding of a situation. When navigating,
there is a trajector and a goal for example, and relevant relations pertain to their spatial
arrangement. However navigation situations can become more complex, hence the need for
question 3, which allows entities to enter into the agent’s attention. Thus, an obstacle becomes
relevant, and so may be an other entity that prevents the obstacle from being moved out of the
way. Question 4 is important because it provides a cleanup mechanism for the agent’s attention.</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Proposed approach</title>
      <p>We implement our monitoring agent as a combination of perception procedures that are called
depending on the results of a reasoning process. The reasoning process takes into account
information about the current situation of the environment, the robot’s goals, and previously
available perception data about objects. It is fast, and integrated into the perception-action loop
of the agent. The knowledge going into the reasoning process also comes from definitions of
image schemas and their relationships. In the next subsections we will provide detail about the
image schematic knowledge and the organization of the image schematic reasoning process.</p>
      <sec id="sec-4-1">
        <title>4.1. Frame Semantics and Description &amp; Situation Pattern</title>
        <p>
          To model image schemas as framal structures we adopted Fillmore’s frame semantics [29]. Frame
semantics has been most influential as a combination of linguistic descriptions and contextual
knowledge to describe cognitive representation of phenomena occurring in the world. Frames
are schematic structures representing prototypical and recurrent elements in order for some
entity, event or situation to realize. Fillmore describes frame-based structures to notions such
as experiential gestalt [
          <xref ref-type="bibr" rid="ref13">13</xref>
          ], stating that frames can refer to a unified framework of knowledge or
a coherent schematization of recurrent experience. For example a simple action like walking,
to be represented in its framal structure, would require some necessary roles such as: a subject
of the action (agent) and some spatial extension covered by the walk (path), but also some
external elements such as the duration of the walk (time), and optional elements could be
expressed, e.g. the weather conditions during the action, the curvilinear shape of the walk, the
unexpected stumbling stone encountered, nesting more complex knowledge in more specific
frames and scenarios. Lexical units (words and phrases) are associated with frames in order to
fully understand their semantics, based on mental schemes representing these evoked scenes. In
FrameNet [30], frames are also explained as situation types. Therefore, in our reasoning module,
image schemas are represented as frames, while their spatial primitives are represented as roles,
and the detection of one of its roles, allows the inference, and gestaltic activation, of the whole
image schema, as described in Section 4.2. We furthermore reuse the Description and Situation
ontology design pattern [31], which allows the introduction of a constructivist perspective in
the ontological module. In our work, as described in section 4.3, the occurrence of e.g. some
robot moving towards its Goal is treated as the occurrence of an image-schematic situation,
namely, in the above mentioned example, as a Source_Path_Goal situation.
4.2. ISL2OWL
Any formal research aiming to utilize the semantic richness of image schemas has to deal with
the complexity of formally representing the full range of abstract conceptualisations associated
with these spatio-temporal relationships. One formalization method developed for this purpose
is the Image Schema Logic, ISL . Greatly simplified, it combines spatial mereology in the form
of Region Connection Calculus (RCC) [32] in Euclidean 3D space, with Qualitative Trajectory
Calculus (QTC) to describe relative movement between objects and regions [33] framed with
Linear Temporal Logic (LTL) over the Reals to allow for sequential changes (For a complete
account of the logic’s syntax and semantics see [
          <xref ref-type="bibr" rid="ref8">8, 34</xref>
          ]).
        </p>
        <p>In our work, we abstract away from the richer representation in ISL  by transposing them
into the Web Ontology Language (OWL2) following a frame semantics approach [29], called
ISL2OWL1. While this reduces the semantic richness of the image-schematic concepts to its
minimal structural elements, it also allows us to seamlessly integrate them into our cognitive
robotics framework. Thus, in ISL2OWL ontological module each Image Schema is modeled
as a frame, taking as necessary roles its spatial primitives. E.g. Source_Path_Goal takes as
necessary roles three elements: Source, Path, and Goal. Due to the gestalt, frame-based nature
of image-schemas, the activation of one of spatial primitive triggers the activation of the whole
image schema; e.g. knowing there is a goal also means knowing there is a path and a source.</p>
        <sec id="sec-4-1-1">
          <title>1The current ISL2OWL version used in this work is available at:</title>
          <p>https://raw.githubusercontent.com/mpomarlan/robontics2022/main/src/ISL2OWL_4_Robontics.ttl ;
while ISL2OWL full graphs are available at:
https://github.com/StenDoipanni/ISAAC/tree/main/ISL2OWL</p>
        </sec>
      </sec>
      <sec id="sec-4-2">
        <title>4.3. Image Schematic Reasoning for Action Selection</title>
        <p>In order to answer the competency questions mentioned in Section 3.1, and to do so fast enough
to be useful for a robot during its activity, we defined an image schematic reasoning layer (ISRL).
It has its own ontological module, is written in a simpler formalism than OWL-DL and in which
inference is guaranteed tractable, and contains some simple heuristics about image schematic
situations and how these connect to each other. The tractability comes from the low complexity
of the inference problem for this formal system – linear time in the size of the knowledge base
fragment used to answer a query– as opposed to planning which in the worst case requires
time exponential in the size of a planning problem.</p>
        <p>The ISRL is queried once every perception-action loop of a robot or agent. It takes as input
prior knowledge obtained via a process of inference and perception during the previous iteration,
and background knowledge, i.e. knowledge that is believed true from other sources.</p>
        <p>In this paper, we focus on the perception side, so the output of ISRL is posterior knowledge
about how the situations around an agent are developing and a perception tasklist. The perception
tasklist provides information about how a robot’s perception system should reconfigure itself to
track environmental entities and relationships that are inferred to be relevant. The perception
module would then act on this tasklist and produce new knowledge about the environment.
The new perception knowledge, together with the inferred posterior knowledge will become
the prior knowledge for the next iteration of reasoning with ISRL. In a more general case, the
ISRL would also decide, for each of a robot’s actuators (or logical groupings of actuators, such
as base, arms, head etc.) a set of actions to perform. Figure 1 shows an overview of how the
ISRL is integrated into a robot’s/agent’s perception-action loop.</p>
        <p>The main benefits of this approach is that it is reactive, adaptable, and situation-aware.
Reactivity here means simply that it can operate fast enough to be useful for quick cognitive
processes that have to deal with a changing and sometimes surprising environment. It is
adaptable in that the complexity of the perception queries scale, in a controlled way, depending
on what is deemed relevant at each particular time. Finally, situation-awareness here means
that the approach has and makes use of current knowledge of the situation, in other words a
top-down understanding of the environment, to filter and make sense of a bottom-up stream of
facts about that environment so that it can update its higher level understanding of the situation.</p>
        <p>In more detail, the formalism in which ISRL is written is that of defeasible rules, with
some limitations. A rule is formed of an antecedent, which is a conjunction of terms, and
an antecedent, which is one term. A term asserts either that some predicate is true on some
collection of arguments (e.g.,   (, )), or that the predicate is false on a collection
of arguments (e.g. −   (, )). The meaning of a rule is that, if its antecedent is
established, then the consequent is defeasibly provable. We refer the reader to [35] for details
about the proof theory of defeasible logic and how conflicts between rules are resolved. We
have chosen the combination “ambiguity propagation”, “team defeat”, and “loop detection”.</p>
        <p>One limit we place on our formal system implementation, is that predicates can have at most
two arguments, which makes them formally similar to the triples commonly used in knowledge
representation. Another limitation concerns the use of variables in the expression of a rule: all
variables in the consequent must appear in the antecedent as well, i.e. no “existential rules”.</p>
        <p>The rules in our system include, partially, information coming from ISL2OWL and domain
knowledge about our running example for this paper, which is navigation by robots on a planar
surface where obstacles may be present. The rules have been written by hand, however see
section 6 for a perspective on how they could be at least in part generated by combining several
knowledge sources. The knowledge we took from ISL2OWL is as follows:</p>
        <p>The state of the environment is considered formally as a dul:Situation, which may
have other dul:Situation participants – in particular, to account for the image schematic
relations between ots such as an object being supported by another, or blocked by another. Each
modification of relationships among entities is an dul:Event. This structure allows inferences
about participants in image schematic situations.</p>
        <p>Image schematic situations have particular kinds of participants. E.g., Source_Path_Goal
has a isrl:goal (among others), a isrl:BLOCKAGE has a isrl:blocked etc. A situation
can have multiple image schematic meanings, and the same object may play several roles in it
(e.g., both isrl:blocked and isrl:goal).</p>
        <p>If two objects are not falling and are in Contact such that one is isc:above the other
allows inferring a isrl:SUPPORT Situation with the two objects as participants.</p>
        <p>The domain knowledge for our running example is that robots that move their wheels should
move unless they are blocked, and that to block a robot – or other object more generally – then
some obstacle should exist in front of the robot/object and be in contact with it. This helps in
creating rules with which to infer what questions to ask of perception.</p>
        <p>In order to decide what questions to ask, each object currently in the attention of the
robot/agent is associated to a corresponding question individual, e.g. isrl:about(q_box,box).
Through inference, various information becomes attached to the question individual, such as
what kind of question it is – and it can be several kinds of question, because several
questions may be relevant about an object. For example, if we know an object is the trajector of a
isrl:SOURCE_PATH_GOAL, it is relevant to ask whether it is moving, and what is in front of it.
isA(?S,SOURCE_PATH_GOAL), trajector(?S,?R), about(?Q,?R) =&gt; isA(?Q,AskIsMoving)
isA(?S,SOURCE_PATH_GOAL), trajector(?S,?R), about(?Q,?R) =&gt; isA(?Q,AskIsInSpatialRelation)
isA(?S,SOURCE_PATH_GOAL), trajector(?S,?R), about(?Q,?R) =&gt; hasMode(?Q,frontOf)
isA(?S,SOURCE_PATH_GOAL), trajector(?S,?R), about(?Q,?R) =&gt; hasRelatum(?Q,?R)</p>
        <p>Once ISRL completes inference, all the inferred symbolic facts are passed on to a perception
module. Primarily, this module is interested in triples asserting more detailed types for question
individuals, because these assertions decide what functions will be called to answer the queries.
However, the symbolic context provided by the other facts is also important to decide how will
these functions answer their queries. For example, to return a list of objects that are in a given
spatial relationship with another, the semantics of this spatial relation needs to be described.
That is, it is not just interesting for perception to know that an individual q_turtle is of type
AskIsInSpatialRelation, but also what kind of spatial modality[36] is being queried.</p>
        <p>Even in our fairly simple scenarios discussed in section 5, this polysemous nature of spatial
relations is important. It is relevant to query what is in front of the robot when it moves; it is
also important to query what is in front of an obstacle in the robot’s path, because whatever
is in front of the obstacle might prevent the robot from pushing the obstacle out of its way.
However, these two meanings of the “in front of” relation are not the same. In the first, we have
a reference object that provides also the reference axis for what in front means – the robot, and
its forward direction. In the second case, the reference object is the obstacle, but the reference
axis is the robot’s (the blocked object’s) forward direction.</p>
        <p>Finally, the perception module returns a list of new facts, such as states of (relative) motion of
objects, movement of actuators, relative spatial relations. These facts will get propagated to the
next iteration of the perception-action loop, where they will be fed into ISRL’s inference, and the
process repeats. Note that this allows a robot’s scope of attention to change as appropriate to
the needs of a situation. That is, objects may be added to its attention scope by being mentioned
in the facts produced by perception. Objects can also drop from the attention scope when there
are no relevant questions to ask about them and they do not get mentioned in the answers to
the relevant questions.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Evaluation</title>
      <p>To evaluate the proposed method for reasoning for situation assessment a series of experiments
are performed using mobile robots in a simulated world performing a navigation task. Each
robot is supposed to go to a particular goal, but the structure of the environment and the robots’
actual actions will difer in the various scenarios.</p>
      <p>Our scenarios run in PyBullet [37], a physics based simulation environment widely used in
robotics research. The simulated robots are Turtlebot 2 robots, equipped with a depth camera
sensor. An example setting of an empty environment with two robots trying to reach two goals
can be seen in figure 2a. The simulated environments used in the scope of this work, represent
situations where robots are required to physically interact with their environment in order to
achieve their navigation goals. For example in figure 2b the purple robot needs to understand
that its path is blocked by an obstacle and evade it. In the final scenario, seen in figure 2c the
path of the purple robot to its goal is blocked by a box trapped between walls such that the
robot cannot push it out of the way. In the current setting any classical path planning approach
would fail. Instead, the orange robot can understand that and can help by pushing the obstacle
out of the way.</p>
      <p>Our use case is of a “monitor” agent that, being aware that the two robots have goals they
should navigate towards, keeps track of relevant entities and relationships from the environment.
(a)
(b)
(c)</p>
      <p>It does so by performing inferences based on spatial relations as reported by perception, the
current knowledge of image schematic situations, the action-selection ontological query system
and image schematic reasoning. That is, at every iteration, perception is instructed to monitor
whether the robots are trying to move, and actually moving towards their goal. However, more
complex inferences and queries to the perception system become available when needed.</p>
      <p>A query that the monitor agent will make of perception is whether there are objects in front
of a robot. If an object is in front of the robot, a further query is asked – is this object in contact?
A situation of a robot that attempts to move, but does not, with a contacting object in front
of it is one of blockage, and in such a situation the neighborhood of the blocking object also
becomes relevant – in particular, what’s “in front” of the blocking object and might prevent it
from being pushed aside. In short, the monitoring agent expands its attention as the situation
becomes more complex by adding more objects to the list of objects it tracks, in a controlled
and motivated manner.</p>
      <p>The opposite – removing objects from attention – is also done once the reasons for which an
object was deemed relevant cease to hold. For example, once the box in figure 2b is no longer in
front of the robot, it is no longer an object about which perception queries are asked.</p>
    </sec>
    <sec id="sec-6">
      <title>6. Conclusion</title>
      <p>This work presented a method for robots to understand and reason about the current situation
evolving around them. Such an understanding is important, as it allows focusing attention and
reacting to situations correctly, as well as coping with cases in the environment that would cause
classical task, path, and trajectory planning approaches to fail. The robots use concepts from
situation awareness theory, along with image schemas and the logic developed around them. To
evaluate the outcomes of the proposed approach, a set of simulated scenarios were used, which
showed the robots were successful in perceiving the correct information from their environment,
and using it to reason about the problem in each situation. Furthermore, the realized system and
the heuristics applied in the action-selection loop constitute a robotic image-schema detector,
which performs cognitive reasoning based on spatial information, detecting image-schematic
situation from knowledge about the physical static/dynamic status quo of the environment.</p>
      <p>An immediate future expansion will be focused on increasing the robot autonomy by
integrating the current approach with a combination of TAMP in more complex constrained-based
manipulation problems that require a sophisticated reasoning mechanism to allow the robot
to adapt its plan in both symbolic and geometric levels to the current situation. Moreover,
another direction of investigation could be automatic generation of the ontology used by the
image schematic reasoning layer (ISRL), i.e. the automatic creation of situation update and
perception/action selection rules by combining knowledge from higher-level, more expressive
formalisms. A concrete application of this is the analysis of unsafe actions, where “unsafe” is
understood through a simple heuristic: an irreversible action is potentially unsafe. Deciding
about safety then requires accounting for the laws of the environment (physical constraints)
and the structure of the robotic agent (agent’s limited capabilities), and may be formalized as
a series of queries to a PDDL planner to detect irreversible actions. These queries would be
ofline, and their results used to create the ontology used by ISRL for situated, fast reasoning. As
an example of where this could be useful, the awareness of not being able to restore a Support
situation to its original configuration could set a threshold to the amount of “safety” an agent
would pursue, and make it avoid toppling over items. Finally, given the above directions, one
could start investigating the possibilities of automated reasoning regarding the causality of
phenomena, especially related to the actions of robots, as well as, the explainability of the
actions chosen by the robot.
[26] M. Diab, M. Pomarlan, D. Beßler, A. Akbari, J. Rosell, J. Bateman, M. Beetz, Skillman – a
skill-based robotic manipulation framework based on perception and reasoning, Robotics
and Autonomous Systems 134 (2020) 103653.
[27] P. Englert, M. Toussaint, Learning manipulation skills from a single demonstration, The
International Journal of Robotics Research 37 (2018) 137–154. URL: https://doi.org/10.1177/
0278364917743795. doi:Path10.1177/0278364917743795.
[28] M. R. Endsley, Toward a theory of situation awareness in dynamic systems, in: Situational
awareness, Routledge, 2017, pp. 9–42.
[29] C. J. Fillmore, Frame semantics, in: Linguistics in the Morning Calm, Seoul: Hanshin,
1982, pp. 111–138.
[30] C. F. Baker, C. J. Fillmore, J. B. Lowe, The berkeley framenet project, in: Proceedings of
the 17th international conference on Computational linguistics-Volume 1, Association for
Computational Linguistics, 1998, pp. 86–90.
[31] A. Gangemi, P. Mika, Understanding the semantic web through descriptions and situations,
in: OTM Confederated International Conferences" On the Move to Meaningful Internet
Systems", Springer, 2003, pp. 689–706.
[32] D. A. Randell, Z. Cui, A. G. Cohn, A spatial logic based on regions and connection, in:</p>
      <p>Proc. 3rd Int. Conf. on knowledge representation and reasoning, 1992.
[33] N. V. D. Weghe, A. G. Cohn, G. D. Tré, P. D. Maeyer, A qualitative trajectory calculus as a
basis for representing moving objects in geographical information systems, Control and
cybernetics 35 (2006) 97–119.
[34] M. M. Hedblom, O. Kutz, T. Mossakowski, F. Neuhaus, Between contact and support:
Introducing a logic for image schemas and directed movement, in: F. Esposito, R. Basili,
S. Ferilli, F. A. Lisi (Eds.), AI*IA 2017: Advances in Artificial Intelligence, 2017, pp. 256–268.
[35] H. P. Lam, On the derivability of defeasible logic, 2012.
[36] J. A. Bateman, Gum: The generalized upper model, Applied Ontology (2021).</p>
      <p>doi:Path10.3233/AO-210258.
[37] E. Coumans, Y. Bai, Pybullet, a python module for physics simulation for games, robotics
and machine learning, http://pybullet.org, 2016–2022.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>P. E.</given-names>
            <surname>Hart</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N. J.</given-names>
            <surname>Nilsson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Raphael</surname>
          </string-name>
          ,
          <article-title>A formal basis for the heuristic determination of minimum cost paths</article-title>
          ,
          <source>IEEE transactions on Systems Science and Cybernetics</source>
          <volume>4</volume>
          (
          <year>1968</year>
          )
          <fpage>100</fpage>
          -
          <lpage>107</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>A.</given-names>
            <surname>Stentz</surname>
          </string-name>
          ,
          <article-title>Optimal and eficient path planning for unknown and dynamic environments</article-title>
          ,
          <source>INTERNATIONAL JOURNAL OF ROBOTICS AND AUTOMATION 10</source>
          (
          <year>1993</year>
          )
          <fpage>89</fpage>
          -
          <lpage>100</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>J. J.</given-names>
            <surname>Kufner</surname>
          </string-name>
          ,
          <string-name>
            <surname>S. M.</surname>
          </string-name>
          <article-title>LaValle, Rrt-connect: An eficient approach to single-query path planning</article-title>
          ,
          <source>in: Proceedings 2000 ICRA. Millennium Conference. IEEE International Conference on Robotics and Automation</source>
          .
          <source>Symposia Proceedings (Cat. No. 00CH37065)</source>
          , volume
          <volume>2</volume>
          , IEEE,
          <year>2000</year>
          , pp.
          <fpage>995</fpage>
          -
          <lpage>1001</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>C.</given-names>
            <surname>Rösmann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Hofmann</surname>
          </string-name>
          , T. Bertram,
          <article-title>Timed-elastic-bands for time-optimal point-to-point nonlinear model predictive control</article-title>
          ,
          <source>in: 2015 european control conference (ECC)</source>
          , IEEE,
          <year>2015</year>
          , pp.
          <fpage>3352</fpage>
          -
          <lpage>3357</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>M.</given-names>
            <surname>Ghallab</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Nau</surname>
          </string-name>
          , P. Traverso,
          <source>Automated Planning: theory and practice</source>
          , Elsevier,
          <year>2004</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Kootbally</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Schlenof</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Lawler</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Kramer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Gupta</surname>
          </string-name>
          ,
          <article-title>Towards robust assembly with knowledge representation for the planning domain definition language (PDDL), Robot</article-title>
          . Comput.-Integr. Manuf.
          <volume>33</volume>
          (
          <year>2015</year>
          )
          <fpage>42</fpage>
          -
          <lpage>55</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>R. A.</given-names>
            <surname>Brooks</surname>
          </string-name>
          ,
          <article-title>Elephants don't play chess</article-title>
          ,
          <source>Robotics and Autonomous Systems</source>
          <volume>6</volume>
          (
          <year>1990</year>
          )
          <fpage>3</fpage>
          -
          <lpage>15</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <surname>M. M. Hedblom</surname>
          </string-name>
          , Image Schemas and Concept Invention: Cognitive, Logical, and Linguistic Investigations, Cognitive Technologies, Springer Computer Science,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>M.</given-names>
            <surname>Johnson</surname>
          </string-name>
          , The Body in the Mind Metaphors, University of Chicago Press,
          <year>1987</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>G.</given-names>
            <surname>Metta</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Sandini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Vernon</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Natale</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Nori</surname>
          </string-name>
          ,
          <article-title>The icub humanoid robot: an open platform for research in embodied cognition</article-title>
          ,
          <source>in: Proceedings of the 8th workshop on performance metrics for intelligent systems</source>
          ,
          <year>2008</year>
          , pp.
          <fpage>50</fpage>
          -
          <lpage>56</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>L.</given-names>
            <surname>Shapiro</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Embodied</given-names>
            <surname>Cognition</surname>
          </string-name>
          , New problems of philosophy, Routledge, London and New York,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <surname>J. M. Mandler</surname>
          </string-name>
          ,
          <article-title>How to build a baby: II. Conceptual primitives</article-title>
          ,
          <source>Psychological review 99</source>
          (
          <year>1992</year>
          )
          <fpage>587</fpage>
          -
          <lpage>604</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>G.</given-names>
            <surname>Lakof</surname>
          </string-name>
          , M. Johnson, Metaphors we live by, University of Chicago press,
          <year>1980</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <surname>M. M. Hedblom</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          <string-name>
            <surname>Kutz</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Peñaloza</surname>
          </string-name>
          , G. Guizzardi,
          <article-title>Image schema combinations and complex events</article-title>
          ,
          <source>KI-Künstliche Intelligenz</source>
          <volume>33</volume>
          (
          <year>2019</year>
          )
          <fpage>279</fpage>
          -
          <lpage>291</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>M.</given-names>
            <surname>Pomarlan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. A.</given-names>
            <surname>Bateman</surname>
          </string-name>
          ,
          <article-title>Embodied functional relations: A formal account combining abstract logical theory with grounding in simulation</article-title>
          ,
          <source>in: Formal Ontology in Information Systems: Proceedings of the 11th International Conference (FOIS</source>
          <year>2020</year>
          ), volume
          <volume>330</volume>
          , IOS Press,
          <year>2020</year>
          , p.
          <fpage>155</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>M.</given-names>
            <surname>Pomarlan</surname>
          </string-name>
          ,
          <string-name>
            <surname>M. M. Hedblom</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Porzel</surname>
          </string-name>
          ,
          <article-title>Panta rhei: Curiosity-driven exploration to learn the image-schematic afordances of pouring liquids</article-title>
          ,
          <source>in: Proceedings of the 29th Irish Conference on Artificial Intelligence and Cognitive Science</source>
          , Dublin, Ireland,
          <year>2021</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>K.</given-names>
            <surname>Dhanabalachandran</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Hassouna</surname>
          </string-name>
          ,
          <string-name>
            <surname>M. M. Hedblom</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Küempel</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          <string-name>
            <surname>Leusmann</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Beetz</surname>
          </string-name>
          , Cutting events:
          <article-title>Towards autonomous plan adaption by robotic agents through imageschematic event segmentation</article-title>
          ,
          <source>in: Proceedings of the 11th on Knowledge Capture Conference</source>
          ,
          <year>2021</year>
          , pp.
          <fpage>25</fpage>
          -
          <lpage>32</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>W.</given-names>
            <surname>Kuhn</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. U.</given-names>
            <surname>Frank</surname>
          </string-name>
          ,
          <article-title>A formalization of metaphors and image-schemas in user interfaces</article-title>
          ,
          <source>in: Cognitive and linguistic aspects of geographic space</source>
          , Springer,
          <year>1991</year>
          , pp.
          <fpage>419</fpage>
          -
          <lpage>434</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>S.</given-names>
            <surname>De Giorgis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Gangemi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Gromann</surname>
          </string-name>
          , Imageschemanet:
          <article-title>Formalizing embodied commonsense knowledge providing an image-schematic layer to framester, Semantic Web Journal forthcoming (</article-title>
          <year>2022</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <surname>M. M. Hedblom</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Gromann</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          <string-name>
            <surname>Kutz</surname>
          </string-name>
          , In, Out and Through:
          <article-title>Formalising some dynamic aspects of the image schema Containment</article-title>
          ,
          <source>in: SAC '18: Proceedings of the 33rd Annual ACM Symposium on Applied Computing</source>
          , Pau, France,
          <year>2018</year>
          , pp.
          <fpage>918</fpage>
          --
          <lpage>925</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <surname>J. M. Mandler</surname>
          </string-name>
          ,
          <article-title>How to build a baby: Ii. conceptual primitives</article-title>
          .,
          <source>Psychological review 99</source>
          (
          <year>1992</year>
          )
          <fpage>587</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>D.</given-names>
            <surname>McDermott</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Ghallab</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Howe</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Knoblock</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Ram</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Veloso</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Weld</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Wilkins</surname>
          </string-name>
          ,
          <article-title>Pddl-the planning domain definition language (</article-title>
          <year>1998</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <given-names>T.</given-names>
            <surname>Pan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. M.</given-names>
            <surname>Wells</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Shome</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L. E.</given-names>
            <surname>Kavraki</surname>
          </string-name>
          ,
          <article-title>A general task and motion planning framework for multiple manipulators</article-title>
          ,
          <source>in: 2021 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS)</source>
          ,
          <year>2021</year>
          , pp.
          <fpage>3168</fpage>
          -
          <lpage>3174</lpage>
          . doi:Path10.1109/IROS51168.
          <year>2021</year>
          .
          <volume>9636119</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [24]
          <string-name>
            <given-names>M.</given-names>
            <surname>Diab</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Akbari</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. Ud</given-names>
            <surname>Din</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Rosell</surname>
          </string-name>
          ,
          <article-title>Pmk-a knowledge processing framework for autonomous robotics perception and manipulation</article-title>
          ,
          <source>Sensors</source>
          <volume>19</volume>
          (
          <year>2019</year>
          ). URL: https: //www.mdpi.com/1424-8220/19/5/1166. doi:Path10.3390/s19051166.
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          [25]
          <string-name>
            <given-names>T.</given-names>
            <surname>Migimatsu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Bohg</surname>
          </string-name>
          ,
          <article-title>Object-centric task and motion planning in dynamic environments</article-title>
          ,
          <source>IEEE Robotics and Automation Letters</source>
          <volume>5</volume>
          (
          <year>2020</year>
          )
          <fpage>844</fpage>
          -
          <lpage>851</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>