<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Getting On Top of Things: Towards Intelligent Robotic Object Stacking through Image-Schematic Reasoning</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>KaviyaDhanabalachandr</string-name>
          <email>kaviya@uni-bremen.de</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>nMaria M.Hedblom</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>MichaelBeetz</string-name>
          <email>beetz@uni-bremen.de</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Workshop Proceedings</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Institute of Artificial Intelligence, University of Bremen</institution>
          ,
          <country country="DE">Germany</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Jönköping Artificial Intelligence Laboratory, Jönköping University</institution>
          ,
          <country country="SE">Sweden</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>The Sixth Image Schema Day</institution>
          ,
          <addr-line>ISD6</addr-line>
        </aff>
      </contrib-group>
      <abstract>
        <p>In this extended abstract, we present initial work on intelligent object stacking by household robots using a symbolic approach grounded in image schema research. Image schemas represent spatiotemporal relationships that capture objects' afordances and dispositions. Therefore, they ofer the first step to ground semantic information in symbolic descriptions. We hypothesise that for a robot to successfully stack objects of diferent dispositions, these relationships can be used to more intelligently identify both task constraints and relevant event segments.</p>
      </abstract>
      <kwd-group>
        <kwd>image schemas</kwd>
        <kwd>object stacking</kwd>
        <kwd>cognitive robotics</kwd>
        <kwd>commonsense reasoning</kwd>
        <kwd>embodied cognition</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction and problem space</title>
      <p>Designing systems for automated spatial reasoning is one of the most challenging yet most
important components for intelligent sys1t]e.mFo[r cognitive robots, which can be defined as
intelligent agents acting in space and time, an ability for spatial reasoning is a requirement for
almost any activity. Yet, spatial reasoning in uncertain environments remains a complex area to
solve within cognitive robotics. To contribute to this research agenda, this extended abstract
introduces our research on identifying how the ‘physical rules’ of objects can be tied to their
successful stacking.</p>
      <p>Stacking is an important skill for many everyday activities. For instance, rearranging a book
shelf, placing groceries into a pantry and carrying objects from one place to another on a tray.
These are all activities that require the knowledge of the involved objects’ features and the
understanding of the underlying rules of how objects with such features can be stacked.</p>
      <p>From repeated experiences with stacking objects, humans have extracted a lot of implicit (and
explicit) knowledge of how certain objects can be treated. Such rules include the understanding
of object properties. For instance, that a flat, sturdy object like a tray will ofer support to other
objects whereas a flat, flexible object like a sheet of paper will not ofer the same support. Equally
important is the understanding that objects will likely slide of when placed on top of slippery,
nEvelop-O
(M. Beetz)
(M. Beetz)
convex surfaces (unless, of course, executed by an expert in object balancing). Other learnt
rules stem from the distinction of the objects’ afordances. For instance, classic, upward-facing
containers (boxes, cups and bowls) behave diferently from flat supporting objects in a stacking
scenario. In the latter, objects are placed on top of the flat object’s surface. In the former, the
stacked objects are placed inside their hull (given that they are of an appropriate size).</p>
      <p>For any formal system dealing with stacking situations, every feature of each involved object
and every waypoint for executing certain trajectory needs to be represented in perfect detail
for the agent to successfully accomplish the task. Providing this to a system is not only time
consuming and cost ineficient for the engineer, it also reduces the autonomy and adaptability
of such artificial agents. Thus, to implement a more intelligent capacity for object stacking, new
directions need to be investigated.</p>
      <p>
        One such direction that often is used in cognitive robotics is based on the theory of embodied
cognition (e.g.2[
        <xref ref-type="bibr" rid="ref3">, 3</xref>
        ]). Within this theoretical framework, intelligent behaviour is thought to
be based on conceptual patterns of meaning that are extracted from repeated experiences, by
some, called image schemas4[
        <xref ref-type="bibr" rid="ref5">, 5</xref>
        ].
      </p>
      <p>Image schemas represent the underlying rules for how object properties allow certain kinds
of actions and are closely connected to object aford1a[n6c]e.sThese patterns take the form of
object dispositions such as oferinSgupport2 andContainment. They also capture relational
properties like relative object sSiczeal(e) and vertical orientatiVoenr(ticality), as well as
the object afordances related to movemeSnotu(rce_Path_Goal). Conceptual components
like these become essential in understanding how certain objects can be used and how they
behave in diferent situations. For instance, a cup is defined as its ability to contain liquids
(Containment), a tray is defined by the ability to support objects for movemSeunptp(ort +
Source_Path_Goal) but both need to be stacked with the right orientation and the correct order.
Thus, an event like stacking objects can be described as a combination of vertical pick-and-place
tasks V(erticality + Source_Path_Goal) with theSupport andContainment constraints of
any involved objects.</p>
      <p>The mission of this research endeavour (s7e,e8[] for some previous work) is to use the
semantic information found in image-schematic patterns when designing robotic actions descriptions
to generate meaningful event segments that can be reasoned9a].bout [</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related Work on Robots Stacking Objects</title>
      <p>Object stacking is a well-known problem in robotics that still engage researchers. The problem
involves understanding how geometric shape relates to vertical stabBilailtaynacnedand how
material properties relate to sturdinesSsuapnpodrt. The problem has been approached with
methods ranging from task planning, reinforcement learning (RL), and to vision-based learning
techniques that aims to learn the naïve physics of stacking.</p>
      <p>For instance, in10[] the stacking problem is approached by iterative incorporation of motion
constraints at the task le1v1e]l.in[stead use a neural network to perform a stability classification.</p>
      <p>1Afordances are actions that environments and objects allow. For instance, a box ofers the afordance of
Containment.</p>
      <p>2Following convention, image-schemas are written in capitalised upper letters.</p>
      <p>Their system learns geometric afordances of introduced objects and arranges the objects based
on a ‘stackability score.’ Based on this score, they are alsoBaabllaentcoe an unstable stack by
placing an object to counterbalance the composition. Another noteworthy cont12r]ibution, [
introduces a robot that can build a tower of irregular stones by employing a gradient descent
based pose search algorithm to find the best pose for each added stone.</p>
      <p>In more domestic environments1,3[] studies the problem of a robot organising shelves based
on user preferences. Arguably, rearranging objects possess similar characteristics to stacking
but instead of balancing on a vertical axis, complexity is to find appropriate compositions on
the horizontal axi1s.3][ approached the problem by predicting pairwise object preference of a
human user by using collaborative filtering model on crowd-sourced data.</p>
      <p>In -more recent work14[], a robot is tasked to stack objects of diferent colours and complex
shapes such as trapezoids and parallelograms. Through an RL algorithm, the simulation is
tasked to stack two of the objects. The complexity of the task is to extract how the shape of
the objects afects the stack and overcome the distraction of the diferent colours of the objects.
Other RL methods for learning object stacking use probabilistic inference for learning control
(PILCO) to learn a task-dependent parameterised policy that generalises to tasks that difer only
by a reward function15[]. In [16], a hierarchical RL method, integrating planning with RL
is presented in which action planning is performed for high-level decision making and an RL
agent is used for low-level motor control. Object stacking is formalised as a multi-step task
planning problem and solved by a hierarchical RL framewor1k7]i.n [</p>
      <p>The capability to reason about object afordance property is important for a successful
stacking. Research exists (e.g1.1,[18, 19]) on modelling physical properties of scenes of objects
using neural network architectures to enable reasoning about the individual object behaviour
and how they behave in pairwise object interactions. While eficient in extracting features,
many machine learning models might not cater to providing an understanding of how these
features relate to functional properties, and a semantic representation is needed to translate
the information. Likewise, the results of the NetHack challenge organised as part of NeurIPS
2021 [20] showed that the robot systems based on symbolic methods outperformed machine
learning agents. Arguably, this means that for intelligent reasoning about object stacking to
take place, some level of semantic information needs to be introduced into the system.</p>
      <p>In the light of this, we propose using symbolic representations of the semantic components
present in the image schemas to approach object stacking. In the next section, we describe the
preliminary methodology on how we intend to proceed with this problem.</p>
    </sec>
    <sec id="sec-3">
      <title>3. Approaching Intelligent Stacking: Foundation and First Steps</title>
      <p>Using a symbolic approach to robotic object stacking is a challenging task as the system needs to
have access to all the (relevant) objects’ properties. However, as it is based on image-schematic
patterns it also provides a base to do intelligent reasoning to efectively predict which spatial
arrangements will be possible and how it will afect the stability of the stack.</p>
      <p>To demonstrate our proposed method, take a household task of clearing the table from a set
of objects; a spoon, a plate, and a cup, see Fig1u.rIef these are to be stacked appropriately
for transport, it is required information that (in most cases) the stack will be more stable if
small-sized objects like spoons are placed into the cup rather than balanced on the plate, or
even worse, that the spoon is at the bottom. Finding the most optimal constellation is based on
understanding the object afordances and dispositions.</p>
      <sec id="sec-3-1">
        <title>3.1. Foundational Framework</title>
        <p>Our framework relies on KnowRo2b1][, a knowledge representation and reasoning system.
It has two parts. One part is an ontological knowledge base written in the Web Ontology
Language (OWL). The other is a logic-based programming language, Prolog, which is used as a
reasoner over the knowledge stored in the Mongo database. It provides an abstraction from
sub-symbolic, high-dimensional data obtained from robot sensors by using high-level symbolic
representations and thus enabling it to ground semantic information from its perceptions.
Together with their image-schematic dispositions and afordance properties, objects are to
be ontologically modelled in SOMA. This type of representation enables the robots to infer
the functional aspect the objects through their afordance pr2o2p]e.rItnyo[ur framework,
afordances are treated as bidirectional dispositi2o3n].sT[his means that for object having the
trigger disposition of beingcoantainer, there must exist another object (real or hypothetical)
that has the bearer dispositiocnaonfbe contained.</p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Preliminary Rules for Stacking</title>
        <p>For many stacking problems in households, placing smaller objects on the top is a sensible rule
as it provides more stability. Additionally, objects of diferent materials and ‘sturdiness’ greatly
impact the stackability.</p>
        <p>In our working example of stacking tableware, we argued that the cup should ideally go on
top of the saucer and the spoon inside the cup. In our framework, this means that the saucer
possess the disposition propertDyeposition (SOMA’s name forSupport) and acts as abearer
for objects that can be deposited on top of it. In this case, the most suitable object is the cup, for
which it afords the action description tasSktaocfking. Correspondingly, the cup is tthriegger
with the role oDfepositedObject as it can be placed on top of the saucer. Similarly, the given
set of objects can be checked for the disposition properCtoynotfainment (Containment)
and can select a suitable trigger object which can be inserted into or contained within that
object. Here the cup with its concave surface would have this disposition and take on the role
of a container to relatively smaller objects such as the spoon. Establishing such relationships
among the objects reduces the search space for possible stack configurations and eliminates the
possibility of unstable combinations such as placing the cup on top of the spoon and the saucer
on top of the cup.</p>
        <p>Based on such object properties and logical rules an algorithm for stacking order is to be
implemented as part of the rule engine in KnowRob. Below are a few preliminary considerations
that we estimate important for intelligent stacking.</p>
        <p>• Objects with thDeeposition disposition go below objects with tDheepositability
disposition.
• Objects with thDeeposition disposition go below objects with tChoentainment
disposition.
• Finally, objects with tChoentainment disposition go below objects with
tChoentainability disposition.</p>
        <p>The above considerations are not intended as complete, neither in terms of how they relate to
object properties nor the relationships between them. For instance, there exist many categories
in which this reasoning does not apply, regardless of their object dispositions. One example is
objects classified as food, which, as a rule, should always go on top of tableware and never the
other way around.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Discussion and future work</title>
      <p>Stacking is a deceptively simple task for humans, with several underlying complexities when
transferred to artificial and robotic agents. For robots to be able to function in household
environments, intelligent reasoning skills of how object properties and afordances relate to
their stackability need to be formally investigated. While still at an early stage, our system aims
to help with this by enabling the robot to infer the order of objects to be stacked and the motion
constraints that have to be maintained.</p>
      <p>Future work includes integrating the formal representations of stackable objects and their
dispositions such aCsontainment andDeposition. We also intend to develop the algorithmic
rules for determining the order on how to stack particular objects based on their dispositions.
One important aspect to consider is that only a limited number of rules can be defined in
KnowRob as they need to be handcrafted. In an ideal case, the system should instead be able to
learn these rules by itself. tThere exist many methods for this that can be considered, but we
envision a combination of observation data (human demonstration data) to model the relation
between the physical attributes of the object and the stacking stability and allow the system to
extract rules from a curiosity-driven exploration in simul2a4t]iaonnd [[25].</p>
      <p>Another step following this work is to establish a method to infer the motion constraints
for the task to be executed by the robot as it has to handle objects with varying physical
properties. These constraints are intended to be passed onto Gi2s6k]a,ardco[nstraint-based
robot controller. This will enable Giskard to take the predicate goal as input and convert them
to robot control commands. This means that defining image-schematic goals sCucohntaasct
(table, cup) as part of the task description is suficient for the robot to infer that the cup should
be placedOnTopOf the table and correctly execute the sequence of actions necessary to reach
the goal.</p>
    </sec>
    <sec id="sec-5">
      <title>Acknowledgements</title>
      <p>The research reported in this paper has been partially supported by the German Research
Foundation DFG, as part of Collaborative Research Center (Sonderforschungsbereich) 1320
“EASE - Everyday Activity Science and Engineering”, University of Brehmtetnp:(//www.ease-crc.
org/).
intuition for generalised object stacking, in: Proceedings of the European Conference on
Computer Vision (ECCV), 2018, pp. 702–717.
[12] F. Furrer, M. Wermelinger, H. Yoshida, F. Gramazio, M. Kohler, R. Siegwart, M. Hutter,
Autonomous robotic stone stacking with online next best object target pose planning, in:
2017 IEEE international conference on robotics and automation (ICRA), IEEE, 2017, pp.
2350–2356.
[13] N. Abdo, C. Stachniss, L. Spinello, W. Burgard, Robot, organize my shelves! tidying up
objects by predicting user preferences, in: 2015 IEEE international conference on robotics
and automation (ICRA), IEEE, 2015, pp. 1557–1564.
[14] A. X. Lee, C. Devin, Y. Zhou, T. Lampe, K. Bousmalis, J. T. Springenberg, A. Byravan,
A. Abdolmaleki, N. Gileadi, D. Khosid, C. Fantacci, J. E. Chen, A. Raju, R. Jeong, M. Neunert,
A. Laurens, S. Saliceti, F. Casarini, M. Riedmiller, R. Hadsell, F. Nori, Beyond pick-and-place:
Tackling robotic stacking of diverse shapes, 2a0r2X1i. v:2110.06192.
[15] M. P. Deisenroth, P. Englert, J. Peters, D. Fox, Multi-task policy search for robotics, in:
2014 IEEE International Conference on Robotics and Automation (ICRA), IEEE, 2014, pp.
3876–3881.
[16] M. Eppe, P. D. Nguyen, S. Wermter, From semantics to execution: Integrating action
planning with reinforcement learning for robotic causal problem-solving, Frontiers in
Robotics and AI 6 (2019).
[17] X. Yang, Z. Ji, J. Wu, Y.-K. Lai, C. Wei, G. Liu, R. Setchi, Hierarchical reinforcement learning
with universal policies for multistep robotic manipulation, IEEE Transactions on Neural
Networks and Learning Systems (2021) 1–15. dPoait:h10.1109/TNNLS.2021.3059912.
[18] P. W. Battaglia, R. Pascanu, M. Lai, D. Rezende, K. Kavukcuoglu, Interaction networks for
learning about objects, relations and physics, a2r0X1i6v.:1612.00222.
[19] D. Raposo, A. Santoro, D. Barrett, R. Pascanu, T. Lillicrap, P. Battaglia, Discovering objects
and their relations from entangled scene representations, arXiv preprint arXiv:1702.05068
(2017).
[20] E. Hambro, E. Grefenstette, H. Küttler, T. Rocktäscel, The nethack challenge:
Dungeons, dragons, and tourishtst,tps://nethackchallenge.com/report,.h20t2m1l. Accessed on
14.12.2021.
[21] M. Beetz, D. Beßler, A. Haidu, M. Pomarlan, A. K. Bozcuoglu, G. Bartels, KnowRob 2.0 – A
2nd Generation Knowledge Processing Framework for Cognition-Enabled Robotic Agents,
in: 2018 IEEE Int. Conf. on Robotics and Automation, ICRA 2018, Brisbane, Australia, May
21-25, 2018, 2018, pp. 512–519. doi:Path10.1109/ICRA.2018.8460964.
[22] D. Beßler, R. Porzel, M. Pomarlan, M. Beetz, R. Malaka, J. Bateman, A formal model
of afordances for flexible robotic task execution, in: ECAI 2020, IOS Press, 2020, pp.
2425–2432.
[23] M. Turvey, Ecological foundations of cognition: Invariants of perception and action. (1992).
[24] M. Gasse, D. Grasset, G. Gaudron, P.-Y. Oudeyer, Causal reinforcement learning using
observational and interventional data,a2r0X2i1v.:2106.14421.
[25] M. M. Hedblom, M. Pomarlan, R. Porzel, Panta Rhei: curiosity-driven exploration to learn
the image-schematic afordances of pouring liquids, in: Proceedings of the 29th Irish
Conference on Artificial Intelligence and Cognitive Science, Dublin, Ireland, 2021.
[26] Z. Fang, G. Bartels, M. Beetz, Learning models for constraint-based motion
parameterization from interactive physics-based simulation, in: 2016 IEEE/RSJ International Conference
on Intelligent Robots and Systems (IROS), IEEE, 2016, pp. 4005–4012.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>B.</given-names>
            <surname>Bennett</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Cohn</surname>
          </string-name>
          ,
          <article-title>Automated common-sense spatial reasoning: still a huge challenge</article-title>
          , in: S. Muggleton, N. Chater (Eds.),
          <source>Human-Like Machine Intelligence</source>
          , Oxford University Press,
          <year>2021</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>D.</given-names>
            <surname>Batra</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. X.</given-names>
            <surname>Chang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Chernova</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. J.</given-names>
            <surname>Davison</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Deng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Koltun</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Levine</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Malik</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Mordatch</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Mottaghi</surname>
          </string-name>
          , et al.,
          <article-title>Rearrangement: A challenge for embodied ai</article-title>
          , arXiv preprint arXiv:
          <year>2011</year>
          .
          <year>01975</year>
          (
          <year>2020</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>L.</given-names>
            <surname>Smith</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Gasser</surname>
          </string-name>
          ,
          <article-title>The development of embodied cognition: Six lessons from babies</article-title>
          ,
          <source>Artif. Life</source>
          <volume>11</volume>
          (
          <year>2005</year>
          )
          <fpage>13</fpage>
          -
          <lpage>30</lpage>
          . URL:https://doi.org/10.1162/106454605327897.3 doi:Path10.1162/1064546053278973.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>M.</given-names>
            <surname>Johnson</surname>
          </string-name>
          ,
          <article-title>The Body in the Mind: The Bodily Basis of Meaning, Imagination, and</article-title>
          <string-name>
            <surname>Reason</surname>
          </string-name>
          , The University of Chicago Press, Chicago and London,
          <year>1987</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <surname>M. M. Hedblom</surname>
          </string-name>
          , Image Schemas and Concept Invention: Cognitive, Logical, and Linguistic Investigations, Cognitive Technologies, Springer Computer Science,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>J. J.</given-names>
            <surname>Gibson</surname>
          </string-name>
          ,
          <article-title>The theory of afordances</article-title>
          , in: R.
          <string-name>
            <surname>Shaw</surname>
          </string-name>
          , J. Bransford (Eds.), Perceiving, Acting, and
          <article-title>Knowing: Toward an Ecological Psychology</article-title>
          , NJ: Lawrence Erlbaum, Hillsdale,
          <year>1977</year>
          , pp.
          <fpage>67</fpage>
          -
          <lpage>82</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <surname>M. M. Hedblom</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Pomarlan</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Porzel</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Malaka</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Beetz</surname>
          </string-name>
          ,
          <article-title>Dynamic action selection using image schema-based reasoning for robots</article-title>
          ,
          <source>in: Proceedings of the 7th Joint Ontology Workshops</source>
          ,
          <year>2021</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>K.</given-names>
            <surname>Dhanabalachandran</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Hassouna</surname>
          </string-name>
          ,
          <string-name>
            <surname>M. M. Hedblom</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Küempel</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          <string-name>
            <surname>Leusmann</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Beetz</surname>
          </string-name>
          , Cutting events:
          <article-title>Towards autonomous plan adaption by robotic agents through imageschematic event segmentation</article-title>
          ,
          <source>in: Proceedings of the 11th on Knowledge Capture Conference, K-CAP '21</source>
          ,
          <string-name>
            <surname>Association</surname>
          </string-name>
          for Computing Machinery, New York, NY, USA,
          <year>2021</year>
          , p.
          <fpage>25</fpage>
          -
          <lpage>32</lpage>
          . URL: https://doi.org/10.1145/3460210.349358.5doi:
          <fpage>Path10</fpage>
          .1145/3460210.3493585.
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <surname>M. M. Hedblom</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          <string-name>
            <surname>Kutz</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Peñaloza</surname>
          </string-name>
          , G. Guizzardi,
          <article-title>Image schema combinations and complex events</article-title>
          ,
          <source>KI-Künstliche Intelligenz</source>
          <volume>33</volume>
          (
          <year>2019</year>
          )
          <fpage>279</fpage>
          -
          <lpage>291</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>N. T.</given-names>
            <surname>Dantam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z. K.</given-names>
            <surname>Kingston</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Chaudhuri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L. E.</given-names>
            <surname>Kavraki</surname>
          </string-name>
          ,
          <article-title>Incremental task and motion planning: A constraint-based approach</article-title>
          , in: Robotics:
          <article-title>Science and systems</article-title>
          , volume
          <volume>12</volume>
          ,
          <string-name>
            <surname>Ann</surname>
            <given-names>Arbor</given-names>
          </string-name>
          , MI, USA,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>O.</given-names>
            <surname>Groth</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F. B.</given-names>
            <surname>Fuchs</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Posner</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Vedaldi</surname>
          </string-name>
          ,
          <article-title>Shapestacks: Learning vision-based physical</article-title>
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>