<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>A Case-Based Framework for Task Demonstration Storage and Adaptation</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Tesca Fitzgerald</string-name>
          <email>tesca.fitzgerald@cc.gatech.edu</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Ashok Goel</string-name>
          <email>goel@cc.gatech.edu</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>School of Interactive Computing, Georgia Institute of Technology</institution>
          ,
          <addr-line>Atlanta, Georgia, 30332</addr-line>
          ,
          <country country="US">USA</country>
        </aff>
      </contrib-group>
      <fpage>53</fpage>
      <lpage>57</lpage>
      <abstract>
        <p>We address the problem of imitation learning in interactive robots which learn from task demonstrations. Many current approaches to interactive robot learning are performed over a set of demonstrations, where the robot observes several demonstrations of the same task and then creates a generalized model. In contrast, we aim to enable a robot to learn from individual demonstrations, each of which are stored in the robot's memory as source cases. When the robot is later tasked with repeating a task in a new environment containing a di↵erent set of objects, features, or a new object configuration, the robot would then use a casebased reasoning framework to retrieve, adapt, and execute the source case demonstration in the new environment. We describe our ongoing work to implement this case-based framework for imitation learning in robotic agents.</p>
      </abstract>
      <kwd-group>
        <kwd>Case-based agents</kwd>
        <kwd>imitation learning</kwd>
        <kwd>robotics</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Introduction</title>
      <p>
        Imitation is an essential process in human social learning and cognition [
        <xref ref-type="bibr" rid="ref10 ref11">11, 10</xref>
        ].
Imitation learning occurs when a learner observes a teacher demonstrating some
action, providing knowledge of (i) how the action was performed and (ii) the
resulting e↵ects of that action. This interaction-guided learning method allows us
to learn quickly and e↵ectively. As a result of its importance in human cognition,
it follows that imitation learning has become an area of increasing focus in
interactive robotics research as well.
      </p>
      <p>
        The goal of Learning from Demonstration is to enable imitation learning in
robots through interactive demonstrations, provided through methods such as
teleoperation or kinesthetic teaching [
        <xref ref-type="bibr" rid="ref1 ref2">1, 2</xref>
        ]. Regardless of which demonstration
method is used, the following process is often used. First, the human teacher
provides several demonstrations of a skill. Between demonstrations, the teacher
may adjust the environment such that the skill is demonstrated in a variety of
initial configurations. The robot then creates an action model which
generalizes over the provided demonstrations. Lastly, the robot applies the generalized
action model to plan a trajectory which is executed in a new environment.
Copyright © 2015 for this paper by its authors. Copying permitted for private and
academic purposes. In Proceedings of the ICCBR 2015 Workshops. Frankfurt, Germany.
      </p>
      <p>However, a challenge of this process is that the resulting action model is
dependent on the number of demonstrations that were provided for that particular
task. We also assume that the robot has been exposed to enough variations of
the initial configuration such that its generalized model can be applied to a wide
range of related initial configurations. As such, the generalized model is restricted
to application in environments which are similar to those demonstrated.</p>
      <p>We describe our preliminary work toward defining an alternate approach to
imitation learning in robotics, one which takes a case-based approach in which
the robot stores demonstrations individually in memory. We define a case-based
framework which enables the full imitation learning process, from observing a
task demonstration to transfer and execution. We also define a case
representation which encodes task demonstrations for storage in source case memory.
2</p>
    </sec>
    <sec id="sec-2">
      <title>Related Work</title>
      <p>
        Case-based reasoning has been used to address the problem of transfer in robotics
domains. Floyd, Esfandiari &amp; Lam [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] describe a CBR approach to learning
strategies for RoboCup soccer by observing spatially distributed soccer team
plays. Their approach represents each case as an encoding of a single agent’s
perception and resulting action at a given time. Thus, they transfer the
behavior of an agent when it perceives a situation similar to that of the observed
agent. More recently, Floyd &amp; Esfandiari [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] describe an approach for case-based
learning by observation in which strategy-level domain-independent knowledge
is separated from low-level, domain-dependent information such as the sensors
and e↵ectors on a physical robot. Ontan˜´on et al. [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] describe their approach to
observational learning for agents in real-time strategy games. They use a
casebased approach to online planning, in which agents adapt action plans which are
observed from game logs of expert demonstrations.
      </p>
      <p>
        While these approaches do address knowledge transfer for robotic and
simulated agents, they primarily operate over input and output represented at a
higher level of abstraction, such as actions at a strategic level. The goal of our
work is to enable transfer to generate action at a lower level of control and in
response to real-world perceptual input, where we transfer the demonstrated
action trajectory used to achieve a task. We expand on our previous work [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]
describing a case-based approach to interpretation and imitation in robotic agents.
We discussed two separate processes: (i) interpreting new skill demonstrations
by comparing it to previously observed demonstrations using a case based
process (further described in [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]), and (ii) a related process for imitating a task
demonstration. This paper expands on the latter process, case-based imitation.
      </p>
      <p>
        We previously provided a general outline for imitation in [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] in which four
steps occur: representation of the task demonstration at multiple levels of
abstraction, retrieval of the most relevant source case from memory, adaptation
of the source case to address the target problem, and execution of the adapted
case in the target problem. In this paper, we describe our more recent work
providing (i) a revised, complete process of imitation beginning with observation
of the task demonstration and ending with task transfer and execution, (ii) a
Mapping step which bridges the gap between the Retrieval and Transfer steps,
and (iii) a revised case representation for storing task demonstrations (iterating
on preliminary work introduced in [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]).
3
      </p>
    </sec>
    <sec id="sec-3">
      <title>Approach Overview</title>
      <p>
        We have revised our case-based approach to transfer (originally summarized
in [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]) to consist of two separate processes, as shown in Figure 1: the Case
Storage process in which the robot receives demonstrations of a task and stores
each demonstration as a case in source memory, and a Case Adaptation process
which is used at a later time when the robot is asked to repeat a task in a target
environment.
3.1
      </p>
      <sec id="sec-3-1">
        <title>Why a CBR approach?</title>
        <p>Our eventual goal is to enable transfer for imitation learning in scenarios such
as the following. A human teacher guides the robot to complete a task such as
scooping the contents of one container into another. During the demonstration,
the robot records the demonstrated trajectories and object features. At a later
time, the robot is asked to repeat the scooping task, but in a new, target
environment. Thus, the robot must use a di↵erent set of object features to parameterize
and execute the scooping task than those observed in the original, source
environment. Next, the robot transfers its representation of the scooping task to
accommodate for the di↵erences between the source and target environments.
The transferred task representation is then executed in the target environment.</p>
        <p>
          Rather than generalize over a set of demonstrations as in current Learning
from Demonstration methods (surveyed in [
          <xref ref-type="bibr" rid="ref1 ref2">1, 2</xref>
          ]), using a case-based approach
allows us to: (1) operate under the assumption that the human teacher will
provide a limited number of demonstrations, (2) represent demonstrations as
individual experiences in the robot’s memory, and (3) utilize a complete
framework for transferring skill demonstrations, which includes the steps of retrieving,
analyzing, transferring, and executing a relevant source case demonstration in
an unfamiliar, target environment.
3.2
        </p>
      </sec>
      <sec id="sec-3-2">
        <title>Case Storage Process</title>
        <p>
          Demonstration and Learning We have implemented the first step in the
Case Storage process, where the robot records and stores each task
demonstration as a source case in memory. We define each case as the tuple C =
&lt;L, D, T, O, Si, Sf &gt;, where:
– L represents the label of the task which was demonstrated, e.g. ”scooping”.
– D represents the set of action models which encode the demonstrated motion,
represented as Dynamic Movement Primitives as defined in [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ].
– T is the set of parameterization functions which relate the set of action
models to the locations of objects in the robot’s environment. For example,
a parameterization function may be used to represent how the robot’s hand
must be located above a bowl prior to completing a pouring action.
– O is the set of salient object IDs which are relevant to the task.
– Si and Sf are the initial and final states, respectively, which represent the
set of objects observed in an overhead view of the robot’s environment.
3.3
        </p>
      </sec>
      <sec id="sec-3-3">
        <title>Case Adaptation Process</title>
        <p>At a later time, the robot may be asked to repeat the task in a new, target
environment. We are currently implementing the Case Adaptation process shown
in Figure 1.</p>
        <p>Observation will begin when the robot is asked to address a target problem.
We assume that the robot has been provided a relevant source case which it
can retrieve from memory to address the given target problem. The robot will
then observe the target environment by viewing the objects located in the
tabletop environment using an overhead camera. This will provide it with the target
case’s initial state Si.</p>
        <p>Retrieval must be performed to select a source case from memory containing
the demonstration that is most relevant to the current target problem. Case
retrieval will prioritize (i) similarity of task goals, (ii) similarity of salient objects,
and finally, (iii) similarity of initial states. Once a relevant source case has been
retrieved, the Mapping step must encode the di↵erences between the source and
target environments. This mapping will be later used to transfer the source case
such that di↵erences in the target environment are addressed.</p>
        <p>Given a source case and mapping which encodes the di↵erences between the
source and target cases, the Transfer step adapts the source case. We take a
similarity-based approach to transfer, where we consider the similarity between
the source case and target environments when defining transfer processes. As we
encounter transfer problems in which the source and target problems become
less similar, the source case is transferred at a di↵erent level of abstraction, such
that only high-level features of that case are transferred. The adapted case is
then executed in the target environment.</p>
        <p>We have implemented three methods which implement the Transfer step,
each of which operates by transferring the source case at a di↵erent level of
abstraction. Once the source case has been transferred, it is used to plan and
execute a new action trajectory. In preliminary experiments, we have evaluated
each method separately such that we selected the level of abstraction at which
transfer occurred in each target problem. These experiments have shown us that
by changing the level of abstraction at which a case is transferred, a robot can
use a single source demonstration to address target environments of varying
similarity to the source environment.
4</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Future Work</title>
      <p>We have implemented the Case Storage process and the last two steps of the
Case Adaptation process, the Transfer and Execution steps. Currently, we
manually provide the robot with the most relevant source case demonstration and
a mapping between objects in the source and target environments. Thus, our
next steps are to identify a method for autonomously determining this object
mapping. Furthermore, future work will involve defining a process for identifying
and retrieving an appropriate source case demonstration that is most applicable
to a given transfer problem.</p>
    </sec>
    <sec id="sec-5">
      <title>Acknowledgments</title>
      <p>This work was supported by NSF Graduate Research Fellowship DGE-1148903.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Argall</surname>
            ,
            <given-names>B.D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chernova</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Veloso</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Browning</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          :
          <article-title>A survey of robot learning from demonstration</article-title>
          .
          <source>Robotics and Autonomous Systems</source>
          <volume>57</volume>
          (
          <issue>5</issue>
          ),
          <fpage>469</fpage>
          -
          <lpage>483</lpage>
          (
          <year>2009</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Chernova</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Thomaz</surname>
            ,
            <given-names>A.L.</given-names>
          </string-name>
          :
          <article-title>Robot learning from human teachers</article-title>
          .
          <source>Synthesis Lectures on Artificial Intelligence and Machine Learning</source>
          <volume>8</volume>
          (
          <issue>3</issue>
          ),
          <fpage>1</fpage>
          -
          <lpage>121</lpage>
          (
          <year>2014</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Fitzgerald</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Goel</surname>
            ,
            <given-names>A.:</given-names>
          </string-name>
          <article-title>A case-based approach to imitation learning in robotic agents</article-title>
          .
          <source>Intl. Conf. on Case-Based Reasoning Workshop on Case-Based Agents</source>
          (
          <year>2014</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Fitzgerald</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Goel</surname>
            ,
            <given-names>A.K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Thomaz</surname>
            ,
            <given-names>A.L.</given-names>
          </string-name>
          :
          <article-title>Representing skill demonstrations for adaptation and transfer</article-title>
          .
          <source>AAAI Symposium on Knowledge, Skill</source>
          , and Behavior Transfer in Autonomous Robots (
          <year>2014</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Fitzgerald</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>McGreggor</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Akgun</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Thomaz</surname>
            ,
            <given-names>A.L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Goel</surname>
            ,
            <given-names>A.K.</given-names>
          </string-name>
          :
          <article-title>Visual case retrieval for interpreting skill demonstrations</article-title>
          .
          <source>International Conference on CaseBased Reasoning</source>
          (
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Floyd</surname>
            ,
            <given-names>M.W.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Esfandiari</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          :
          <article-title>A case-based reasoning framework for developing agents using learning by observation</article-title>
          .
          <source>In: 2011 23rd IEEE International Conference on Tools with Artificial Intelligence (ICTAI)</source>
          . pp.
          <fpage>531</fpage>
          -
          <lpage>538</lpage>
          . IEEE (
          <year>2011</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Floyd</surname>
            ,
            <given-names>M.W.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Esfandiari</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lam</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          :
          <article-title>A case-based reasoning approach to imitating robocup players</article-title>
          .
          <source>In: FLAIRS Conference</source>
          . pp.
          <fpage>251</fpage>
          -
          <lpage>256</lpage>
          (
          <year>2008</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8. Ontan˜o´n,
          <string-name>
            <given-names>S.</given-names>
            ,
            <surname>Mishra</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            ,
            <surname>Sugandh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            ,
            <surname>Ram</surname>
          </string-name>
          ,
          <string-name>
            <surname>A.</surname>
          </string-name>
          :
          <article-title>Case-based planning and execution for real-time strategy games</article-title>
          .
          <source>In: Case-Based Reasoning Research and Development</source>
          , pp.
          <fpage>164</fpage>
          -
          <lpage>178</lpage>
          . Springer (
          <year>2007</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Pastor</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          , Ho↵mann, H.,
          <string-name>
            <surname>Asfour</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Schaal</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          :
          <article-title>Learning and generalization of motor skills by learning from demonstration</article-title>
          .
          <source>In: Robotics and Automation</source>
          ,
          <year>2009</year>
          . ICRA'09. IEEE International Conference on. pp.
          <fpage>763</fpage>
          -
          <lpage>768</lpage>
          . IEEE (
          <year>2009</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Piaget</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cook</surname>
          </string-name>
          , M.T.:
          <article-title>The origins of intelligence in children</article-title>
          . (
          <year>1952</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>Tomasello</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kruger</surname>
            ,
            <given-names>A.C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ratner</surname>
            ,
            <given-names>H.H.</given-names>
          </string-name>
          :
          <article-title>Cultural learning</article-title>
          .
          <source>Behavioral and brain sciences</source>
          <volume>16</volume>
          (
          <issue>03</issue>
          ),
          <fpage>495</fpage>
          -
          <lpage>511</lpage>
          (
          <year>1993</year>
          )
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>