<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>A balancing act: Ordering algorithm and image-schematic action descriptors for stacking objects by household robots</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>KaviyaDhanabalachandr</string-name>
          <email>kaviya@uni-bremen.de</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>nMaria M.Hedblom</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>MichaelBeetz</string-name>
          <email>beetz@uni-bremen.de</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Workshop Proceedings</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Institute of Artificial Intelligence, University of Bremen</institution>
          ,
          <country country="DE">Germany</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Jönköping Artificial Intelligence Laboratory, Jönköping University</institution>
          ,
          <country country="SE">Sweden</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>The Eighth Joint Ontology Workshops</institution>
          ,
          <addr-line>JOWO'22</addr-line>
        </aff>
      </contrib-group>
      <abstract>
        <p>Optimising object order in stacking problems remains a hard problem for cognitive robotics research. In this paper, we continue our work on using the spatiotemporal relationships called image schemas to represent afordance spaces founded on object properties. Based on object properties, we introduce a stacking-order algorithm and describe the action descriptors using an image-schematic event segmentation format by describing a small subset using the Image SchemaISLLo gic.</p>
      </abstract>
      <kwd-group>
        <kwd>object stacking</kwd>
        <kwd>cognitive robotics</kwd>
        <kwd>afordances</kwd>
        <kwd>algorithms</kwd>
        <kwd>action representation</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction and Problem Space</title>
      <p>Despite the past decades progress in developing increasingly sophisticated software and
hardware, research in artificial intelligence and cognitive robotics is still struggling to accurately
represent and execute tasks that are learned by a human children in early infancy, a phenomenon
known as Moravec’s parado1x][. One such task is the pick-and-place task. Appearing
deceptively simple, the task includes visual mastery of gaze control, understand spatial depth and
positioning, as well as object recognition and object permanence. It requires object
manipulation abilities for moving the agent’s own body (or body parts), understand the spatial and
force-dynamic relationships of grasping objects. Additionally, it needs to be able to identify
object properties and reason how their afordances behave under particular conditions. In this
paper, we look at a particular complex pick-and-place task by focusing on how to stack object
on the vertical axis. In addition, to the complexities of the pick-and-place task, stacking objects
require a deeper understanding of the properties and afordances of the objects being stacked.
For instance, it is not possible to stack heavy objects on top of flexible objects, nor it is (under
normal circumstances) possible to stack a flat object onto a convex surface.</p>
      <p>The motivation for this research agenda is the common occurrence of stacking objects in
nEvelop-O
(M. Beetz)
(M. Beetz)
household activities. Kitchen utensils are stacked on top of one another in the cupboard and
objects are carried on trays in particular arrangement.</p>
      <p>To provide the household robots with a more intelligent understanding for stacking, we will
use inspiration from cognitive science that demonstrate how human children core down object
properties and action segments into meaningful components. We will then proceed to utilise
these components by formally representing action descriptors involved in stacking and present
an ordering algorithm that takes the object properties into account to provide a stable stack.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Foundation and Related Work</title>
      <sec id="sec-2-1">
        <title>2.1. Theoretical Foundation</title>
        <p>
          Our work is based on the hypothesis that there exists a finite number of conceptual building
blocks that represent the salient features of object relations, including those providing balance
and stackability. These are often referred tiomaasge schemas and encompass spatiotemporal
relationships between objects, agents and their enviro1n[m2e,n3]t.sThe theory stems from
the notion of embodied cognition, a common theoretical framework in cognitive robotics
research (e.g. 4[
          <xref ref-type="bibr" rid="ref5">, 5</xref>
          ]) which assumes that intelligent behaviour, in all its forms, stems from
the body’s sensorimotor experiences in its environment. Within this framework the image
schemas represent the atomic yet salient features that distinguish one particular situation from
another. For instance, the image schCemoantact is present in a situation in which two objects
are physically touching and the image schLeimnak defines objects that (may or may not be)
touching but have a causal relationship to one another. The salient diferenceCiosntthaactt,in
moving one object will not necessarily afect the other object, but if tLhineykeadr,emoving
one object will automatically move the other object as well. Likewise, the salient features of an
object like caup is defined by its ability tCoontain liquids, and objects with flat sturdy surfaces
like atray are defined by their afordance toSupport other objects. In most cases, an event like
stacking objects can be described as a combination of diferent vertical pick-and-place tasks
(Verticality + Source_Path_Goal) with theSupport andContainment constraints of any
involved objects.
        </p>
        <p>Due to the conceptually-rich content of image schemas and their finite number, combinations
of them can be argued to describe all kinds of spatiotemporal situations an6d, 7e]v.eFnotrs [
robotic research, this means that it is possible to formally represent the physical states of both
the initial state and the goal state of particular actions, but also in detail describe the changes
over time that constitute the actions that leads to these changes. In previous work we have
investigated this for diferent scenarios (s8e,e9][) and in Section4 we will demonstrate this for
stacking usingISL .</p>
        <p>1The theory was originally developed in cognitive linguistics to explain the high number of spatial metaphors
in abstract language, but has become a common hypothesis in many research areas dealing with semantic relevance
in relation to spatial reasoning. Not all research fields would agree on defining them as spatiotemporal relationships,
but it is a useful delimitation in cognitive robotics.</p>
      </sec>
      <sec id="sec-2-2">
        <title>2.2. Formal Framework</title>
        <p>
          The representation language ISL : We base our representation on the expressive
combination language the Image Schema LoIgSiLc [
          <xref ref-type="bibr" rid="ref3">3</xref>
          ]. Following a popular temporalisation
strategy (as in10[]), temporal structures are the primary model-theoretic objects, in turn based
on Linear Temporal Logic over the reals (LTL). At each moment of time, we allow for the
employment of secondary semantics to represent complex propositions. These atomic representations
are topological assertions in 3D Euclidean space based on the Region Connection Calculus
(RCC) [
          <xref ref-type="bibr" rid="ref11">11</xref>
          ], relative movement dimension using Qualitative Trajectory Calculus1(2Q] TaCn)d[
relative object position using Ligozat’s Cardinal Direction1s3](C.QDu)a[ntification is used to
separate diferent sortal objects, whilst otherwise the syntax of the language follows a standard
multi-modal logic paradigm. For more details on the logic we re3f,e1r4t].o [
The robotics framework: While preliminary, the robotics framework we utilise is based on
KnowRob [
          <xref ref-type="bibr" rid="ref15">15</xref>
          ], a knowledge representation and reasoning system. Knowrob uses Web ontology
language (OWL) based on description logics to represent the knowledge. Prolog, a logic-based
programming language is used to reason over the knowledge base and to assert new facts that
are computed. We use the formal afordance model defined i1n6][ and represented in SOMA
[
          <xref ref-type="bibr" rid="ref17">17</xref>
          ]. The formal model comprises of concepts including Afordance, Disposition and Role where
Dispositions are properties of objects that takes on the role of a bearer and acts as a description
of Afordance. With the model defined in1[
          <xref ref-type="bibr" rid="ref6">6</xref>
          ], for an event requiring a certain afordance to
execute a task, the task can be executed only if there is a presence of two suitably disposed
objects with a trigger and a bearer role in the environment. This model is used in the algorithm
for stacking.
        </p>
      </sec>
      <sec id="sec-2-3">
        <title>2.3. Previous Work</title>
      </sec>
      <sec id="sec-2-4">
        <title>Image-schematic event representation for robotics: In [7], the authors demonstrate how</title>
        <p>
          it is possible to divide the event of cracking eggs into bowls into image-schematic components.
In previous work, we looked at how robotic action descriptors could be cored down into their
image-schematic relationships. I8n],[we defined image-schematic relations liVkerticality,
Source_Path_Goal, Containment, Link, Contact, Support to describe the action of cutting.
The functional properties of objects and scene description defined using image schema concepts
are used in1[
          <xref ref-type="bibr" rid="ref8">8</xref>
          ] to construct a simulation of the scene and estimate the parameters necessary
for performing the action of pouring. While19in], [the author uses qualitatively described
functional object relations in a spatial arrangement to enable the agent to understand the causal
structure embedded in the scene.
        </p>
        <p>Object stacking: Robotics research on object manipulation comes in many forms, including
work on stacking objects with a focus on the complexity of balancing objects on top of one
another. The major challenge is to learn the physics of how shapes of diferent level of complexity
can be stacked on top of one one another without fa2l0l]inagp.p[roach the problem by using a
neural network to learn the geometric afordance2s1.]I,na[reinforcement learning algorithm
was used to train a system how to stack a series of complex shapes while simultaneously
ignoring irrelevant object features such as colour.</p>
        <p>In more domestic environments, in which household items are to be stacked for more practical
reasons, other challenges presents themselves2.2]I,nt[he authors introduce a method for
stacking objects on shelves based on crowd sourced data.</p>
        <p>Learning functional relationships and afordances: To perform a successful stacking, it
is important to reason about the object afordance property. There are wor2k0s, 2(e3.,g2. 4[])
using machine learning to learn the extracted features representing the physical properties
of objects in a scene to be able to reason about the individual object behaviour and how they
behave in pairwise object interactions. But the problem is the machine learning models does
not enable an understanding of how the object features relate to functional properties. There
needs to be a semantic component as in image schemas to perform object stacking and also be
able to explain in case of failure, when certain objects do not in the stack or failure in action
when a robot executes the task of object stacking.</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Stacking order algorithm</title>
      <p>Computing the stacking order is a search problem to find a stable configuration that can be
stacked from the given set of objects in a scene. The search space of this problem is n!, where n
is the number of objects. Only rigid objects are considered in this work and the stack height is
limited to 5 objects.</p>
      <p>
        Algorithm1 presents the stacking order by computing the disposition properties of the objects.
By Turvey’s [
        <xref ref-type="bibr" rid="ref25">25</xref>
        ] definition of dispositions, there needs to be two objects whose disposition
properties complement each other, for the afordances of the objects to be realised if the situation
demands. For the task of stacking to be performed, one of the objects assumes the role of a
bearer with the ability to support other objects and the second object with a trigger role can be
placed on top of other objects. The roles of the objects can be interchanged, with the object
that is in need of support acting as a bearer and the supporting object acting as a trigger. The
important aspect is the presence of two suitably disposed objects that can enable stacking. The
RuleonTopOf is used to verify if the considered objects satisfy the condition to be stacked and
does not mean that the obj e1ctis placed on top o f2. The rule is defined below,
∀ 1,  2 ∶   (
2,  1) ↔ ∃ 1,  1 ∶ (
1) ∧ (
2) ∧ (
1)∧
(
ℎ(
1) ∧ ℎ(
2,  1) ∧ (
1,  1) ∧ ℎ  (
      </p>
      <p>It comprises of the unary predicaOtebsject, Role andDisposition which are used to assert
tha t 1 and 2 are instances oOfbject,  1 is of typeRole and 1 is aDisposition. The predicate
ℎ relates the object1 with a suitable dispositi o1n, similarlyhasTrigger is a
relation betwe en1 and the role ty pe1 that can act as a trigge r f1oarndhasRole relates
the object2 to role 1 if the object can take on the role. The last precdaincSatatcek is used
to perform geometric reasoning if the objec1tasnd 2 playing the trigger and bearer roles
aforded by the dispositio n 1 can be stacked.</p>
      <p>The rule above explains how to check if two objects can be stacked. When a set of objects
is considered, they are of separated into two groups based on their disposition property. The
separation of the objects is done with the reasoning that any object that can support or contain
other objects with the disposition Deposition and Containment respectively, should be placed
at the bottom of the stack. Likewise, objects without such dispositions such as spoons, as well
as spherical objects that has the property of Rollability, are at the top.</p>
      <p>The objects are modelled with their top facing the poZs-iatxiivse(height) and the length
along theX-axis and the width along tYh-aexis, see Figure1. This is identical to how they are
to be placed on a surface to ensure stability. To avoid comparing each object with the rest of the
objects, we use a simple heuristic of sorting the objects based on their diagonal length along
theXY plane. This is motivated with how it ensures the object with a large bottom surface is at
the bottom of the stack.</p>
      <p>Consider an example scenario with an ordered list of objects comprising of a plate, a spoon
and a bowl as shown in Figu1r.eFirst, we will apply Ru1leto the plate and the bowl. The plate,
having the disposition propertyDoepfosition, needs a trigger role of tyDpeepositedObject, as the
lfat nature of the plate afords tSoupport other objects when placed on top of it. Based on its
physical features, the bowl can play the rolDeeopfosaitedObject and the relatiocannStack holds
as well. Thus, the bowl can be placed on the plate. Continuing, with the bowl and the spoon,
the bowl with the disposition propertCyonotfainment needs a trigger of typCeontainedObject,
as the concave shape alloCwosntainment of objects inside. If the spoon can be contained in
the given bowl then the taskStoacfking can be realise2d. The obtained stack is similar to the
left stack in Figure2. The unstable arrangement of objects is one of the configurations that will
not be considered by the algorithm as the spoon placed at the bottom will not ofer any support
to other objects or contain other objects.</p>
      <p>In the next section, we define the actions necessary for stacking. Stacking is performed by
transporting the objects in the order provided by the algorithm. The defined action sequence
has to be performed recursively until all the objects are stacked.</p>
      <p>2Note that dispositions often correlate with their image-schematic representations, but that we have chosen to
keep their syntactic representation in the text.</p>
    </sec>
    <sec id="sec-4">
      <title>4. Action representation of stacking</title>
      <p>In order to be able to execute successful object stacking, the event of stacking one object on
top of another needs to be cored down and explained in terms of each of the individual actions.
Stacking is a repetition of the task of transporting objects until all objects in our set have been
placed on top of one another following the stacking algorithm.</p>
      <p>
        The actions are classified into atomic and compound as described in the taxonomy of actions,
see [
        <xref ref-type="bibr" rid="ref26">26</xref>
        ]. An atomic action is defined as below,
Definition 1. An action causing a single change in the image schematic relation is an atomic
action.
      </p>
      <p>For example, consider the action of Lifting, before the action is executed an object is in contact
with the robot gripper and is supported by a surface. Once the action is executed, there exists
no contact between the surface and the lifted object and the support is ofered by the gripper.
The Lifting action leads to a changeSuinpport relationship, hencLei,fting is an atomic action.
Compound actions are a composition of atomic actions or compound actions. By that definition,
an action likteransporting is a compound action comprising of two other compound actions:
picking up and placing. Each of the compound action consists of smaller atomic descriptors
that can be described in image-schematic terms.</p>
      <p>Further, the pre- and post scene of the compound action is defined in terms of the spatial
relations between the involved objects. The defined spatial relations and the object roles must
hold for a successful task execution. By defining the pre- and post scene of an action, it is
possible to infer the necessary sequence of actions to execute a task. The representation of the
action with its initial and the terminal scene has the following syntax:
Action () ↔ (InitialScene1() → TerminalScen e(2)), where 1 ,  2 are time points an dis an
interval which begins  a1tand ends at2 during which the action happens.</p>
      <p>This means that to execute an Action, the InitialScene is a prerequisite and once the action is
performed, it results in the TerminalScene, the goal state of the Action. The action takes place
during the time inter varlesulting in the end stat e2a,gtiven the prerequisite is satisfied at
some time point 1. The ’→’ does not point to implication, it rather means that the defined goal
state is achieved if the action is successfully executed under the given context. The initial scene
of an action comprises of objects that can take on suitable roles and relevant spatial relation
Algorithm 1 Rules for Stacking
Input  ← given set of objects in scene
Output  ← Stacking order of given objects
   ← Objects that are not part of the stack
1: procedure ComputesOrderedObjectList
2: OrderedObject←s 
3: Sortedlis←t Objectlist sorted based on their diagonal length in descending order
ignoring the height
/* considers objects that can support other objects */
for object O in do
if hasDispositio n, ( ) and D in [Deposition, Containmentht]en</p>
      <p>Remove O from Sortedlist
if OrderedObjects is empttyhen</p>
      <p>OrderedObject←s OrderedObject∪s O
among objects. Note that when the action is executed, the relationships between the objects
and the roles the objects initially have, can change.</p>
      <p>For a transporting action to happen, the prerequisite is the presencSeuopfpaort relation
between two physical objects taking the roDleespoosfit andDepositedObject at a time point
 1 . When the task is completed successfully, there is agSauipnpoart relation between the
transported object and an object witDhetphoseit role at a time poi n2t. The transporting
predicate is parameterised by 3 objects and a time inter1v, al2,,  3 and  , where 1 is
transported fro m2 to 3 during . The predicate hasInterval relates the TimeIntienrsvtaalnce
with the starting time po i1natnd the end poin t2.
Picking up comprises of several atomic actions: LookingFor, LookingAt, MovingTo, Grasping
and Lifting. LookingFor and LookingAt are movements that are necessary to perceive the object.
MovingTo is a movement towards the object so that the object is reachable. Lifting and Grasping
involve interaction with objects, the gripper of the agent comes in contact with the object in
Grasping and in Lifting, the agent is in control of the object and is being supported by the
gripper. Similar to transporting, there needs tSoubpepoart relation before and after the action
is executed successfully. For picking up o1ffrom  2 to happen, there needs to beSuapport
relation between the object and the surface. There is an3aignevnotlved in picking up and
the object 1 is supported by the agent once the action is complete. The action happens over a
time interval which start s1aatnd ends at2.
supportedBy( 1,  3,  2))
supportedBy( 1,  2,  1)
We provide the definition of the actions involving objects using ISL logic. The o perisator
Until from Temporal Logi↩c, and⇝ are used as in Qualitative Trajectory Calculus to denote
the movement of the object away from, and towards the other object respectively. Grasping
defined below is a combination of the image-schematic aspectLsinokf andContainment.
When the grasping action is performed, the o bj1ecatnd gripper 2 are initially disconnected
and eventually the object becomes a part of the gripper and there exists a contact force between
the gripper and the object.</p>
      <p>∀ 1 ∶ grasping( 1) ↔ ∃ 2 Object( 1) ∧ Gripper( 2)∧
(
1,  2) ( (
Lifting is equivalent to a combination of image schema concepts namUepl,ySource_Path and
Support. During the execution of Lifting, the obje c1tto be lifted is supported by the arm2
while the arm supporting the object, moves along an upwards paawtahy from the support
surface 3. The execution is complete when the arm lies above the support surface and there is
no contact between the object and support surface.</p>
      <p>∀ 1 ∶ lifting ( 1) ↔ ∃ 2,  3,  Object( 1) ∧ Arm( 2) ∧ Object( 3)∧
ℎ() ∧  (
(( 1 ↩  3 ∧ (
(¬((
1, )) ∧ ¬((
2,  3)) ∧ (
1,  3)))
Placing The placing action is performing picking up in reverse order, which comprises of
the following atomic actions: MovingTo, Lowering, Releasing, MovingAway. MovingTo is the
action of an agent moving towards the object to achieve reachability and MovingAway is the act
of moving away from the object once it is placed stable on a surface. Again, we define Lowering
and Releasing actions which involve objects using ISL. The condition necessary for placing to
be executed is the existence oSfuapport that is provided by the ag en2tto an object1 at
some time point 1. When the placing action is performed successfully during the time interval
 , there exists no contact between the agent and the object, and tShueprpoerits arelation
between the placed obje c1tand the objec t3 that can aford to support at time p oi2n. t
∀ 1,  2,  ∶ placing( 1,  2,  ) ↔(∃ 3,  1,  2 Object( 1) ∧ Timepoint( 1) ∧ Timepoint( 2)∧
Timeinterva(l) ∧  2 &gt;  1 ∧ Agent( 3)∧
hasInterva( l, 1,  2) ∧ hasRole( 1, )∧
hasRole( 3, ) ∧
→ Object( 2) ∧ hasRole( 2, )∧
supportedBy( 1,  2,  2))
supportedBy( 1,  3,  1)
Lowering, an atomic action of Placing is a combinatioPnatohf_,Goal andSupport. As in
Lifting, there exists a contact force between the obj1eacntd the ar m 2 during the execution
phase and the arm moves towards the supporting surf3acaelong a pat h. The action is
successfully completed, if at the end, there exists a contact between the object held by the
gripper and the object with a supporting surface.</p>
      <p>∀ 1 ∶ lowering( 1) ↔ ∃ 2,  3,  Object( 1) ∧ Arm( 2) ∧ Object( 3)∧</p>
      <p>Path() ∧  (
(( 1 ⇝  3 ∧ (
(¬((
The next action is Releasing, which begins with a contact between th e g2rainppdetrhe object
 1 and ends with the gripper and the object being disconnected. Once the action is complete,
the object has no contact with the agent.</p>
      <p>∀ 1 ∶ releasin(g 1) ↔ ∃ 2 Object( 1) ∧ Gripper( 2)∧
( (</p>
    </sec>
    <sec id="sec-5">
      <title>5. Discussion and Future Work</title>
      <p>The complexity of designing methods for cognitive robots to successfully stack objects, extends
the problem for eficient manipulation. Performing a successful manipulation, requires an
understanding of how the objects behave in a particular arrangement. In this paper, we utilised
object afordances and introduced an algorithm for intelligent stacking based on individual
object properties, and a rule that acts on objects pairwise to check if on top of relation holds
between them. Taking inspiration from human cognition, we argue that stacking objects of
diferent properties require the agent to understand the underlying rules pertaining to how a
particular object interacts with one another. Our inclination to use object afordances stems
from the fact that they can be described to relate to spatiotemporal relationships, image schemas,
which also constitute the fundamental steps in action execution. Further, we showcase how
image-schematic event representation in the formISaLt of , can be used to describe the action
descriptors. The described action descriptors can then be used for failure monitoring and also
in case of failures in execution, for explaining the reason behind a failure. For example, with the
image-schematic definition of Lifting, there needs to be a support relation between the lifted
object and the gripper involved in lifting and the gripper in contact with the object must be
moving away from the supporting surface in such a way the gripper lies above the supporting
surface. In case of an object slipping while lifting, this can be detected with our definition of
lifting as the support relation between the lifted object and the gripper is violated.</p>
      <p>
        Much work (e.g. 2[
        <xref ref-type="bibr" rid="ref23 ref7">7, 23</xref>
        ]) has used simulation as a means to understand the scene. Considering
an instance of stacking performed only with a physics simulator, the number of random samples
required by the simulator to understand the object properties when they interact with other
objects is huge. Also a part of the generated samples might consist of spatial arrangement
of objects that violates what the object afords. For example, in2F,itguhreere is an object
constellation with spoon at the bottom with a bowl on the top.However, this is an invalid
combination according to our qualitative description while this is a completely valid sample
considered by the simulator. Generating samples that respect our qualitative descriptions of the
scene and object properties can significantly reduce the number of samples needed.
      </p>
      <p>Due to the current theoretical nature of the work, evaluations of the underlying ideas are
still lacking. In the future, we will rectify this, by investigating how the formalised image
schemas can be used by the robot control programs. To use image-schema formalisms along
with robot control programs, it is necessary to represent and quantify the defined spatiotemporal
relations using physics based simulators. The image schematic event segmentation de4fined in
can be used to infer the sequence of actions that the robot has to repeat to perform stacking
until all the objects received from the algo1raitrhemin the stack. To execute the inferred
actions successfully, we need failure handling. With the action descriptors defined in ISL logic,
failure monitoring can be performed. To combine the action sequence for stacking and failure
monitoring in a modular fashion, behaviour trees will be used. For instance, a behaviour tree
can be generated with a parallel node consisting of two children, one for performing the action
execution and the other monitoring the failures based on the action executed. The action
execution node consists of a child node for each atomic action to be executed and will execute
the actions in sequence until all of them succeed. We intend to use
G2i8s]k,aarcdo,n[straintbased robot controller for executing the stacking task. As Giskard already uses behaviour trees
for planning and executing the goals, it is relatively less complex to integrate our approach with
Giskard. And with regards to the rules of stacking, we are interested in a long term research
goal of extracting the rules by letting the agent interact with the environment. For this we want
to use observational data for modelling the relation between the physical attributes of the object
that influence the stacking stability and also collect intervention data by allowing the agent to
interact with the environment similar to a curiosity-driven exploration in si2m9u].lation [</p>
    </sec>
    <sec id="sec-6">
      <title>Acknowledgements</title>
      <p>The research reported in this paper has been partially supported by the Federal Ministry for
Economic Afairs and Energy BMWi within the Knowledge4Retail project, subproject semantic
Digital Twin 01MK20001M (https://knowledge4retail.org).
tion from interactive physics-based simulation, in: 2016 IEEE/RSJ International Conference
on Intelligent Robots and Systems (IROS), IEEE, 2016, pp. 4005–4012.
[29] M. Gasse, D. Grasset, G. Gaudron, P.-Y. Oudeyer, Causal reinforcement learning using
observational and interventional dataa,r2X0i21v.:2106.14421.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>H.</given-names>
            <surname>Moravec</surname>
          </string-name>
          ,
          <article-title>Mind children: The future of robot and human intelligence</article-title>
          , Harvard University Press,
          <year>1988</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>M.</given-names>
            <surname>Johnson</surname>
          </string-name>
          ,
          <article-title>The Body in the Mind: The Bodily Basis of Meaning, Imagination, and</article-title>
          <string-name>
            <surname>Reason</surname>
          </string-name>
          , The University of Chicago Press, Chicago and London,
          <year>1987</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <surname>M. M. Hedblom</surname>
          </string-name>
          , Image Schemas and Concept Invention: Cognitive, Logical, and Linguistic Investigations, Cognitive Technologies, Springer Computer Science,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>D.</given-names>
            <surname>Batra</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. X.</given-names>
            <surname>Chang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Chernova</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. J.</given-names>
            <surname>Davison</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Deng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Koltun</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Levine</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Malik</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Mordatch</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Mottaghi</surname>
          </string-name>
          , et al.,
          <article-title>Rearrangement: A challenge for embodied ai</article-title>
          , arXiv preprint arXiv:
          <year>2011</year>
          .
          <year>01975</year>
          (
          <year>2020</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>L.</given-names>
            <surname>Smith</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Gasser</surname>
          </string-name>
          ,
          <article-title>The development of embodied cognition: Six lessons from babies</article-title>
          ,
          <source>Artif. Life</source>
          <volume>11</volume>
          (
          <year>2005</year>
          )
          <fpage>13</fpage>
          -
          <lpage>30</lpage>
          . URL:https://doi.org/10.1162/10645460532789 7.3 doi:Path10.1162/1064546053278973.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>T.</given-names>
            <surname>Oakley</surname>
          </string-name>
          ,
          <article-title>Image schema</article-title>
          , in: D.
          <string-name>
            <surname>Geeraerts</surname>
          </string-name>
          , H. Cuyckens (Eds.),
          <source>The Oxford Handbook of Cognitive Linguistics</source>
          , Oxford University Press, Oxford,
          <year>2010</year>
          , pp.
          <fpage>214</fpage>
          -
          <lpage>235</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <surname>M. M. Hedblom</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          <string-name>
            <surname>Kutz</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Peñaloza</surname>
          </string-name>
          , G. Guizzardi,
          <article-title>Image schema combinations and complex events</article-title>
          ,
          <source>KI-Künstliche Intelligenz</source>
          <volume>33</volume>
          (
          <year>2019</year>
          )
          <fpage>279</fpage>
          -
          <lpage>291</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>K.</given-names>
            <surname>Dhanabalachandran</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Hassouna</surname>
          </string-name>
          ,
          <string-name>
            <surname>M. M. Hedblom</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Küempel</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          <string-name>
            <surname>Leusmann</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Beetz</surname>
          </string-name>
          , Cutting events:
          <article-title>Towards autonomous plan adaption by robotic agents through imageschematic event segmentation</article-title>
          ,
          <source>in: Proceedings of the 11th on Knowledge Capture Conference, K-CAP '21</source>
          ,
          <string-name>
            <surname>Association</surname>
          </string-name>
          for Computing Machinery, New York, NY, USA,
          <year>2021</year>
          , p.
          <fpage>25</fpage>
          -
          <lpage>32</lpage>
          . URL: https://doi.org/10.1145/3460210.34935 8.5doi:
          <fpage>Path10</fpage>
          .1145/3460210.3493585.
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <surname>M. M. Hedblom</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Pomarlan</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Porzel</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Malaka</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Beetz</surname>
          </string-name>
          ,
          <article-title>Dynamic action selection using image schema-based reasoning for robots</article-title>
          ,
          <source>in: Proc. of the Joint Ontology Workshops</source>
          ,
          <year>2021</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>M.</given-names>
            <surname>Finger</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. M.</given-names>
            <surname>Gabbay</surname>
          </string-name>
          ,
          <article-title>Adding a Temporal Dimension to a Logic System</article-title>
          ,
          <source>Journal of Logic, Language and Information</source>
          <volume>1</volume>
          (
          <year>1993</year>
          )
          <fpage>203</fpage>
          -
          <lpage>233</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>D. A.</given-names>
            <surname>Randell</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Cui</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. G.</given-names>
            <surname>Cohn</surname>
          </string-name>
          ,
          <article-title>A spatial logic based on regions and connection</article-title>
          ,
          <source>in: Proc. 3rd Int. Conf. on Knowledge Rep. and Reas</source>
          .,
          <year>1992</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <surname>N. Van De Weghe</surname>
            ,
            <given-names>A. G.</given-names>
          </string-name>
          <string-name>
            <surname>Cohn</surname>
            , G. De Tré,
            <given-names>P. D.</given-names>
          </string-name>
          <string-name>
            <surname>Maeyer</surname>
          </string-name>
          ,
          <article-title>A qualitative trajectory calculus as a basis for representing moving objects in geographical information systems</article-title>
          ,
          <source>Control and cybernetics 35</source>
          (
          <year>2006</year>
          )
          <fpage>97</fpage>
          -
          <lpage>119</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>G.</given-names>
            <surname>Ligozat</surname>
          </string-name>
          ,
          <article-title>Reasoning about cardinal directions</article-title>
          ,
          <source>J. Vis. Lang. Comput</source>
          .
          <volume>9</volume>
          (
          <year>1998</year>
          )
          <fpage>23</fpage>
          -
          <lpage>44</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <surname>M. M. Hedblom</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          <string-name>
            <surname>Kutz</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Mossakowski</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Neuhaus</surname>
          </string-name>
          ,
          <article-title>Between contact and support: Introducing a logic for image schemas and directed movement</article-title>
          , in: F. Esposito,
          <string-name>
            <given-names>R.</given-names>
            <surname>Basili</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Ferilli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F. A.</given-names>
            <surname>Lisi</surname>
          </string-name>
          (Eds.),
          <source>AI*IA 2017: Advances in Artificial Intelligence</source>
          ,
          <year>2017</year>
          , pp.
          <fpage>256</fpage>
          -
          <lpage>268</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>M.</given-names>
            <surname>Beetz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Beßler</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Haidu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Pomarlan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. K.</given-names>
            <surname>Bozcuoglu</surname>
          </string-name>
          , G. Bartels, KnowRob
          <volume>2</volume>
          .0
          <article-title>- A 2nd Generation Knowledge Processing Framework for Cognition-Enabled Robotic Agents</article-title>
          ,
          <source>in: 2018 IEEE Int. Conf. on Robotics and Automation</source>
          ,
          <string-name>
            <surname>ICRA</surname>
          </string-name>
          <year>2018</year>
          , Brisbane, Australia, May
          <volume>21</volume>
          -25,
          <year>2018</year>
          ,
          <year>2018</year>
          , pp.
          <fpage>512</fpage>
          -
          <lpage>519</lpage>
          . doi:Path10.1109/ICRA.
          <year>2018</year>
          .
          <volume>8460964</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>D.</given-names>
            <surname>Beßler</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Porzel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Pomarlan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Beetz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Malaka</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Bateman</surname>
          </string-name>
          ,
          <article-title>A formal model of afordances for flexible robotic task execution</article-title>
          ,
          <source>in: ECAI</source>
          <year>2020</year>
          , IOS Press,
          <year>2020</year>
          , pp.
          <fpage>2425</fpage>
          -
          <lpage>2432</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>D.</given-names>
            <surname>Beßler</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Porzel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Pomarlan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Vyas</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Höfner</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Beetz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Malaka</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Bateman</surname>
          </string-name>
          ,
          <article-title>Foundations of the socio-physical model of activities (soma) for autonomous robotic agents</article-title>
          , arXiv preprint arXiv:
          <year>2011</year>
          .
          <volume>11972</volume>
          (
          <year>2020</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>M.</given-names>
            <surname>Pomarlan</surname>
          </string-name>
          ,
          <string-name>
            <surname>M. M. Hedblom</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Porzel</surname>
          </string-name>
          , Panta Rhei:
          <article-title>curiosity-driven exploration to learn the image-schematic afordances of pouring liquids</article-title>
          ,
          <source>in: Proceedings of the 29th Irish Conference on Artificial Intelligence and Cognitive Science</source>
          , Dublin, Ireland,
          <year>2021</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>M.</given-names>
            <surname>Pomarlan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. A.</given-names>
            <surname>Bateman</surname>
          </string-name>
          ,
          <article-title>Embodied functional relations: A formal account combining abstract logical theory with grounding in simulation</article-title>
          .,
          <source>in: FOIS</source>
          ,
          <year>2020</year>
          , pp.
          <fpage>155</fpage>
          -
          <lpage>168</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>O.</given-names>
            <surname>Groth</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F. B.</given-names>
            <surname>Fuchs</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Posner</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Vedaldi</surname>
          </string-name>
          ,
          <article-title>Shapestacks: Learning vision-based physical intuition for generalised object stacking</article-title>
          ,
          <source>in: Proceedings of the European Conference on Computer Vision (ECCV)</source>
          ,
          <year>2018</year>
          , pp.
          <fpage>702</fpage>
          -
          <lpage>717</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>A. X.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Devin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Lampe</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Bousmalis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. T.</given-names>
            <surname>Springenberg</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Byravan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Abdolmaleki</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Gileadi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Khosid</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Fantacci</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. E.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Raju</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Jeong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Neunert</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Laurens</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Saliceti</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Casarini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Riedmiller</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Hadsell</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Nori</surname>
          </string-name>
          ,
          <article-title>Beyond pick-and-place: Tackling robotic stacking of diverse shapes, 2a0r2X1i</article-title>
          .v:
          <volume>2110</volume>
          .
          <fpage>06192</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>N.</given-names>
            <surname>Abdo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Stachniss</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Spinello</surname>
          </string-name>
          , W. Burgard,
          <article-title>Robot, organize my shelves! tidying up objects by predicting user preferences</article-title>
          ,
          <source>in: 2015 IEEE international conference on robotics and automation (ICRA)</source>
          , IEEE,
          <year>2015</year>
          , pp.
          <fpage>1557</fpage>
          -
          <lpage>1564</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <given-names>P. W.</given-names>
            <surname>Battaglia</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Pascanu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Lai</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Rezende</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Kavukcuoglu</surname>
          </string-name>
          ,
          <article-title>Interaction networks for learning about objects</article-title>
          ,
          <source>relations and physics,a2r0X1i6v.:1612</source>
          .
          <fpage>00222</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [24]
          <string-name>
            <given-names>D.</given-names>
            <surname>Raposo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Santoro</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Barrett</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Pascanu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Lillicrap</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Battaglia</surname>
          </string-name>
          ,
          <article-title>Discovering objects and their relations from entangled scene representations</article-title>
          ,
          <source>arXiv preprint arXiv:1702.05068</source>
          (
          <year>2017</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          [25]
          <string-name>
            <given-names>M.</given-names>
            <surname>Turvey</surname>
          </string-name>
          ,
          <article-title>Ecological foundations of cognition: Invariants of perception and action</article-title>
          . (
          <year>1992</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          [26]
          <string-name>
            <given-names>P.</given-names>
            <surname>Zech</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Renaudo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Haller</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , J. Piater,
          <article-title>Action representations in robotics: A taxonomy and systematic classification</article-title>
          ,
          <source>The International Journal of Robotics Research</source>
          <volume>38</volume>
          (
          <year>2019</year>
          )
          <fpage>518</fpage>
          -
          <lpage>562</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          [27]
          <string-name>
            <given-names>P. W.</given-names>
            <surname>Battaglia</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. B.</given-names>
            <surname>Hamrick</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. B.</given-names>
            <surname>Tenenbaum</surname>
          </string-name>
          ,
          <article-title>Simulation as an engine of physical scene understanding</article-title>
          ,
          <source>Proceedings of the National Academy of Sciences</source>
          <volume>110</volume>
          (
          <year>2013</year>
          )
          <fpage>18327</fpage>
          -
          <lpage>18332</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          [28]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Fang</surname>
          </string-name>
          , G. Bartels,
          <string-name>
            <given-names>M.</given-names>
            <surname>Beetz</surname>
          </string-name>
          ,
          <article-title>Learning models for constraint-based motion parameteriza-</article-title>
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>