<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>A Framework Inspired by Cognitive Memory to Learn Planning Domains From Demonstrations</article-title>
      </title-group>
      <contrib-group>
        <aff id="aff0">
          <label>0</label>
          <institution>Bioengineering, Robotics and Systems Engineering, University of Genoa</institution>
          ,
          <addr-line>Via Opera Pia 13, 16145, Genoa</addr-line>
          ,
          <country country="IT">Italy</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>We introduce a framework for acquiring structured knowledge from human-lead demonstrations and generate task planning domains for robots. It is based on a novel algorithm that builds symbolic models of environmental states as structured memory items, which are stored and retrieved after reasoning processes. The paper addresses the formalisation of memory items and its management over time through cognitive-like functions, i.e., encoding, storing, retrieving, consolidating and forgetting. Based on the two simple scenarios, we present preliminary results and we discuss the benefits and limitations of our approach.</p>
      </abstract>
      <kwd-group>
        <kwd>Knowledge Acquisition</kwd>
        <kwd>Structured Concepts Learning</kwd>
        <kwd>HumanRobot Collaboration</kwd>
        <kwd>Description Logic</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Introduction</title>
      <p>
        Robots should be able to bootstrap knowledge by observing humans [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ], which
might communicate verbally with supervising purposes. A robot needs to
focus on the important changes of the environment in order to build an agnostic
structure, which will be used for reproducing the observed task in other contexts.
State of the art approaches use imitation learning for mapping observations and
interactions into motion primitives at the trajectory and symbolic levels [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ].
Recurrent Neural Networks [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ], Reinforcement Learning [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ] and Deep Learning
[
        <xref ref-type="bibr" rid="ref15">15</xref>
        ] have been used to actuate a robot based on demonstrations, and the latter
work relies on cognitive aspects to acquire knowledge, i.e., with attention factors.
However, it is challenging to address the issue of storing tasks structures that
are communicable to users, which might improve them through dialogues for
instance. Also, learning black-box like structures strongly limits the integration
of state of the art symbolic task planners and imitation learning techniques.
      </p>
      <p>
        We present an approach to structure models of the observed environment into
a memory, which can be used with reasoning and planning purposes. We used
Description Logic (DL) [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] to manage a general-purpose memory, which contains
items and have functions to store and retrieve observations deduced through
interaction. This paper introduces a formal framework to investigate methods
to acquire communicable knowledge into the robot’s memory for supporting its
Copyright © 2020 for this paper by its authors. Use permitted under Creative
Commons License Attribution 4.0 International (CC BY 4.0).
1
      </p>
      <p>R1
2
3
4</p>
      <p>R2
R2</p>
      <p>Facts
actions. In particular, we consider a scenario where knowledge is memorised
online and stored into a structured tasks representation domain.
2</p>
    </sec>
    <sec id="sec-2">
      <title>Memory Items</title>
      <p>
        We developed the Scene Identification and Tagging (SIT) algorithm [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ], which
generate a symbolic representation of a scene acquired while observing a
humanlead demonstration. The algorithm models scene categories from beliefs about
the environment, which are computed with perception modules that provide
facts. Facts and beliefs describe the environment only for the current instant of
time, while categories are stored in – and retrieved from – the memory. SIT uses
scene beliefs for (i) creating a new category from observation, and (ii) classifying
the current scenes with respect to categories previously learned, if any.
      </p>
      <p>Since SIT is based on symbols in an ontology, it can be defined with a
generalpurpose input interface, based on (i) a set of DL concepts v f 1 : : : ng
describing entities in the environment (e.g., RedBox), and (ii) a set of DL role
R v fR1 : : : Rmg representing relationship among entities of type . Thus, the
input facts are role assertions at a specific time instant, i.e., a role 1; 3: R1,
which relates the DL instances 1 and 3, that are classified in (e.g., 1: 1).
Figure 1 shows a simple 2D example and possible input facts required by SIT.
In this case, a fact is 1; 3: alignedWith, where 1: RedBox and 3: BlueBox.</p>
      <p>For each fact, SIT computes a belief that contributes to the description of
the scene t. Beliefs are computed through reification, which defines a DL role
R j with a symbol deduced from the concatenation of the symbols defining Ri
i
and j , e.g., R1 3 alignedWithBlueBox. With beliefs about t, SIT can create
a new DL concept t that represents a scenes category in the ontology, which is
defined with conjunctions of cardinality restrictions, as shown in the last column
of Figure 1. In the example of Figure 1, the model of the environment at time t
is expressed as a scene category t where: “at least 1 BlueBox is alignedWith
a RedBox, and at least 2 GreenBox are connectedTo a BlueBox”. Remarkably,
each t is defined with respect to the universal scene , which contains all the
possible scenes that can be represented with an input interface h ; Ri.</p>
      <p>SIT checks the consistency among categories restrictions through DL
reasoning, which generates a graph, i.e., the robot’s memory. In the memory, each
(a) A demonstration of an assembly task. (b) A demonstration of objects stacking.
node is an item describing a scene category, while each edge identifies a logic
implication among them (Figure 3). Such a graph does not only represents
relations among sub-scenes (i.e., i v j ), but it can also be used to classify a scene
t with respect to previously generated categories, i.e., t: t i.</p>
      <p>For instance, if at t2 a new block is introduced in the scene of Figure 1
(acquired at t1), SIT will perform one-shot learning to generate a new category
2. Then SIT would classify the new scene 2 in the categories 1 and 2, which
are related to time t1 and t2 respectively. This occurs because 2 has beliefs that
also respect the restrictions learned from 1 and stored in 1. In other words,
SIT infers that the second category implies the first, i.e., 2 v 1, that is related
to an edge in the memory graph (e.g., Figure 3).</p>
      <p>Moreover, SIT provides a normalised similarity value to describe the
classification of t in more categories. This value is low when few beliefs of t satisfy
the restrictions of a category t i, and high otherwise. Based on such a value,
the SIT output interface is a sub-graph of the memory containing each node j
that (i) has all the restrictions satisfies by the beliefs of the current scene t,
and (ii) does not have too many unspecified restrictions for the other beliefs of
t (e.g., the once introduced by the new block when 2: 1 is evaluated).
3</p>
    </sec>
    <sec id="sec-3">
      <title>Memory Capabilities</title>
      <p>Since we want to use SIT when demonstrations hold for a reasonably long interval
of time, and the robot perceives inputs fact about the scenes observed with a
suitable frequency, we define a consolidation score for each node in the memory
graph t, and five functions inspired by cognitive models. Remarkably, since
SIT performs one-shot learning, it might occur that the knowledge in memory
overfits a particular demonstration. In our framework, the consolidating and
forgetting capabilities are used to avoid this issue by implementing an attentive
behaviour that identifies the important items to maintain in memory, e.g., the
once that do not involve the pens in Figure 2.</p>
      <p>More in detail, (i) the encoding functionally generates input fact based on
a contextualisation of sensory data, e.g., to extract spatial relations based on
1
6
2
3
4</p>
      <p>
        5
7
8
9
the centre of mass and shape of objects. During (ii) the storing function, SIT
attempts to classify t and, if it successes, each t i node in the output graph will
increase their consolidating score. Otherwise, a new category t will be derived
from beliefs and added to the memory. (iii) The retrieving function uses DL
queries to classify categories when a scene is requested through beliefs. Similarly
to storing, also retrieving affects the consolidating scores. (iv) The consolidate
function traverses the memory and normalises the scores of each node based
on its neighbours and time trace decay theory [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ]. Whereas (v) the forgetting
function removes the nodes with a low score and restructures the edges of the
graph consistently.
4
      </p>
    </sec>
    <sec id="sec-4">
      <title>Preliminary Results</title>
      <p>The consolidation score does not only rank categories for retrieving purposes, but
it also allows to implement a forgetting function for removing categories that are
not relevant to the demonstrated task. We preliminary tested our system with
the hypothesis that often observed scenes are more important (and might never
be forgotten), than sporadic configurations of facts (which can be neglected). We
tested this approach in two scenarios involving different types of demonstrations.
One consists of assembling the four legs of a table (Figure 2a), while the second
in stacking four objects on top of each other (Figure 2b). For both scenarios
we considered the R v fconnectedTog role assertion, which is estimated from
objects’ centre of mass. Whereas v fSupport; Leg; Peng was considered in the
first scenario, while v fBox; Peng in the second. In each scenario, we consider
an object not related to the task (i.e., Pen), which is used to increase the scenes
variability with configurations not strictly related to the demonstrated task.</p>
      <p>
        Figure 3 shows the memory graph after the observation of the demonstrations
partially shown in Figure 2. We notice that all the categories restricting some
pens in the scene have been forgotten since they were not persistent during the
overall demonstrations. In our scenario, SIT generates a memory that supports
planning techniques because it is possible to find the differences among a ij-th
category pair, and perform the actions required to change the classification of
t from i to j (e.g., with simulations [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]). Without using a consolidating and
forgetting approach, we obtained a memory graph including all the demonstrated
scenes, and many nodes were not directly related to the task. This would not
only strongly limit the performances of SIT over time, but it also requires to
deploy sophisticated reasoning and planning techniques since the graph becomes
more complex. Instead, with the forgetting policy configured for our tests, SIT
represents the tasks in a manner that is effective for planning purposes. However,
to generalise this result for many different scenarios is still an open issue.
      </p>
      <p>
        We implemented the SIT algorithm in a ROS architecture based on the
ARMOR service [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ], and we investigate a scenario where a human could refine the
robot knowledge through dialogues during the demonstration. In [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ], we
addressed this complex human-robot interaction with a relatively simple system
since we exploited the transparent representation that SIT generates. More
generally, we obtained such representation because we based SIT into a symbolic
representation that is familiar to users.
      </p>
      <p>
        Nonetheless, using a symbolic formalism also allows us to design SIT with a
general-purpose input interface, which supports multimodality and that can be
used to generate graphs that contextualise facts differently, e.g., for implementing
semantic and episodic memory types [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ]. On the other hand, our symbolic
input interface also leads to the main drawback of our framework since it does
not allow to use sensory data directly, and it requires a prior symbolic set of
environmental features h ; Ri to be accurately perceived over time, e.g., using
[
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] and [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. Nonetheless, our framework gives a formal platform to investigate the
generation of planning domain through demonstrations also under uncertainties
since the approach presented in this paper is compliant with the SIT extension
based on fuzzy logic [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ].
5
      </p>
    </sec>
    <sec id="sec-5">
      <title>Conclusions</title>
      <p>We presented a framework to acquire knowledge through interaction and produce
a transparent robot memory that can represent planning domains. The memory
allows encoding, storing, retrieving, consolidating and forgetting models of
environmental states based on reasoning and contextualisation. With two proof
of concept scenarios, we discussed a flexible framework for further investigating
memory capabilities. Also, we introduced some open issues and limitations.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Baader</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Calvanese</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>McGuinness</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Patel-Schneider</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Nardi</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          , et al.:
          <article-title>The description logic handbook: Theory, implementation and applications</article-title>
          . Cambridge university press (
          <year>2003</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Billard</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Calinon</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Dillmann</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Schaal</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          :
          <article-title>Robot Programming by Demonstration</article-title>
          . In: Siciliano,
          <string-name>
            <given-names>B.</given-names>
            ,
            <surname>Khatib</surname>
          </string-name>
          ,
          <string-name>
            <surname>O</surname>
          </string-name>
          . (eds.) Springer Handbook of Robotics, pp.
          <fpage>1371</fpage>
          -
          <lpage>1394</lpage>
          . Springer Berlin Heidelberg, Berlin, Heidelberg (
          <year>2008</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Buoncompagni</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Capitanelli</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mastrogiovanni</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          :
          <article-title>A ros multi-ontology references service: Owl reasoners and application prototyping issues</article-title>
          .
          <source>In: Proceedings of the 5th Italian Workshop on Artificial Intelligence and Robotics A workshop of the XVII International Conference of the Italian Association for Artificial Intelligence (AI*IA</source>
          <year>2018</year>
          ). vol.
          <volume>2352</volume>
          , pp.
          <fpage>36</fpage>
          -
          <lpage>41</lpage>
          . CEUR-WS, Trento, Italy (nov
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Buoncompagni</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Carfì</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mastrogiovanni</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          :
          <article-title>A software architecture for multimodal semantic perception fusion</article-title>
          .
          <source>In: Proceedings of the 5th Italian Workshop on Artificial Intelligence and Robotics A workshop of the XVII International Conference of the Italian Association for Artificial Intelligence (AI*IA</source>
          <year>2018</year>
          ). vol.
          <volume>2352</volume>
          , pp.
          <fpage>18</fpage>
          -
          <lpage>23</lpage>
          . CEUR-WS, Trento, Italy (nov
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Buoncompagni</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ghosh</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Moura</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mastrogiovanni</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          :
          <article-title>A Scalable Architecture to Design Multi-modal Interactions for Qualitative Robot Navigation</article-title>
          . In: Ghidini,
          <string-name>
            <given-names>C.</given-names>
            ,
            <surname>Magnini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            ,
            <surname>Passerini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            ,
            <surname>Traverso</surname>
          </string-name>
          , P. (eds.)
          <source>AI*IA 2018 - Advances in Artificial Intelligence</source>
          . pp.
          <fpage>96</fpage>
          -
          <lpage>109</lpage>
          . Lecture Notes in Computer Science, Springer International Publishing, Trento, Italy (nov
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Buoncompagni</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mastrogiovanni</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          :
          <article-title>A software architecture for object perception and semantic representation</article-title>
          .
          <source>In: Proceedings of the 2nd Italian Workshop on Artificial Intelligence and Robotics A workshop of the XIV International Conference of the Italian Association for Artificial Intelligence (AI*IA</source>
          <year>2015</year>
          ). vol.
          <volume>1544</volume>
          , pp.
          <fpage>116</fpage>
          -
          <lpage>124</lpage>
          . CEUR-WS, Ferrara, Italy (sep
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Buoncompagni</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mastrogiovanni</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          :
          <article-title>Teaching a robot how to spatially arrange objects: Representation and recognition issues</article-title>
          .
          <source>In: 2019 28th IEEE International Conference on Robot and Human Interactive Communication (RO-MAN)</source>
          . pp.
          <fpage>1</fpage>
          -
          <lpage>8</lpage>
          . New Delhi, India (Oct
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>Buoncompagni</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mastrogiovanni</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          :
          <article-title>Dialogue-based supervision and explanation of robot spatial beliefs: a software architecture perspective</article-title>
          .
          <source>In: 2018 27th IEEE International Symposium on Robot and Human Interactive Communication (RO-MAN)</source>
          . pp.
          <fpage>977</fpage>
          -
          <lpage>984</lpage>
          . IEEE, Nanjing,
          <source>China (Aug</source>
          <volume>218</volume>
          )
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Buoncompagni</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mastrogiovanni</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Saffiotti</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Scene learning, recognition and similarity detection in a fuzzy ontology via human examples</article-title>
          .
          <source>In: Proceedings of the 4th Italian Workshop on Artificial Intelligence and Robotics A workshop of the XVI International Conference of the Italian Association for Artificial Intelligence (AI*IA</source>
          <year>2017</year>
          ). vol.
          <year>2054</year>
          , pp.
          <fpage>10</fpage>
          -
          <lpage>15</lpage>
          . CEUR-WS, Bari, Italy (nov
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Duan</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Andrychowicz</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Stadie</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ho</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Schneider</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sutskever</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Abbeel</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zaremba</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          :
          <article-title>One-shot Imitation Learning</article-title>
          .
          <source>In: Proceedings of the 31st International Conference on Neural Information Processing Systems</source>
          . pp.
          <fpage>1087</fpage>
          -
          <lpage>1098</lpage>
          . NIPS'
          <volume>17</volume>
          , Curran Associates Inc., New York, USA (dec
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>Jonides</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lewis</surname>
            ,
            <given-names>R.L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Nee</surname>
            ,
            <given-names>D.E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lustig</surname>
            ,
            <given-names>C.A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Berman</surname>
            ,
            <given-names>M.G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Moore</surname>
            ,
            <given-names>K.S.:</given-names>
          </string-name>
          <article-title>The Mind and Brain of Short-Term Memory</article-title>
          .
          <source>Annual review of psychology 59</source>
          ,
          <fpage>193</fpage>
          -
          <lpage>224</lpage>
          (sep
          <year>2008</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <string-name>
            <surname>Kunze</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Burbridge</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hawes</surname>
          </string-name>
          , N.:
          <article-title>Bootstrapping Probabilistic Models of Qualitative Spatial Relations for Active Visual Object Search</article-title>
          . In: 2014 AAAI Spring Symposium Series (Mar
          <year>2014</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13.
          <string-name>
            <surname>Stulp</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Theodorou</surname>
            ,
            <given-names>E.A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Schaal</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          :
          <article-title>Reinforcement Learning With Sequences of Motion Primitives for Robust Manipulation</article-title>
          .
          <source>IEEE Transactions on Robotics</source>
          <volume>28</volume>
          (
          <issue>6</issue>
          ),
          <fpage>1360</fpage>
          -
          <lpage>1370</lpage>
          (
          <year>Dec 2012</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          14.
          <string-name>
            <surname>Vernon</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          :
          <source>Artificial Cognitive Systems: A Primer</source>
          . MIT Press (
          <year>Oct 2014</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          15.
          <string-name>
            <surname>Zhang</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>McCarthy</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jow</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lee</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chen</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Goldberg</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Abbeel</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          :
          <article-title>Deep Imitation Learning for Complex Manipulation Tasks from Virtual Reality Teleoperation</article-title>
          .
          <source>In: 2018 IEEE International Conference on Robotics and Automation (ICRA)</source>
          . pp.
          <fpage>1</fpage>
          -
          <lpage>8</lpage>
          . Brisbane, Australia (May
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>