<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Teaching psychomotor skills using machine learning for error detection</article-title>
      </title-group>
      <contrib-group>
        <aff id="aff0">
          <label>0</label>
          <institution>German Research Center for Arti cial Intelligence</institution>
          ,
          <addr-line>10559 Berlin</addr-line>
          ,
          <country country="DE">Germany</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Learning psychomotor skills is challenging because motion is fast, relies strongly on subconscious mechanisms, and instruction typically disrupts the activity. As such, learners would pro t from mechanisms that can react swiftly, raise subtle mistakes to the conscious level, and do not disrupt the activity. In this paper, we sketch a machine learning-supported approach to provide feedback in two example scenarios: running, and interacting with a robot. For the running case, we provide an evaluation how motions can be compared to highlight deviations between student and expert motion.</p>
      </abstract>
      <kwd-group>
        <kwd>Psychomotor skills</kwd>
        <kwd>Running</kwd>
        <kwd>Human robot interaction</kwd>
        <kwd>Dynamic movement primitives</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Introduction</title>
      <p>
        Teaching bene cial psychomotor skills { such as moving healthily or interacting
with a robot companion { is challenging [
        <xref ref-type="bibr" rid="ref3 ref6">3,6</xref>
        ], because coaches need to infer
mistakes from subtle clues in observable behavior, build a hypothesis regarding
the underlying cause of the mistake, and verbalize an instruction that enables the
learner to improve performance, even though the learner may not be conscious
of the mistake or how to correct it [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. Automatic mechanisms may be helpful
to support instruction in such cases. An automatic feedback mechanism can
perceive and analyze psychomotor activity with only split-second delay and,
thus, provide feedback in almost real-time [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. This permits learners to improve
their psychomotor performance in a much faster loop: if they receive feedback,
they can adapt during the activity and receive additional feedback immediately,
whereas classic coaching would require to interrupt an activity, then receive
verbal feedback, discuss it, and re-start the activity to check the improvement.
      </p>
      <p>In this paper, we sketch a machine-learning-supported approach for feedback
for two psychomotor activity scenarios, namely running and interacting with a
robot. Our approach is intended as an inspiration for a general template that can
be applied across a wide range of psychomotor skills. For the case of running,
we provide a rst analysis using dynamic movement primitives which shows that
we can abstract from irrelevant variations in psychomotor data and hone in on
deviations from expert demonstrations that may be indicative of mistakes.
Copyright © 2021 for this paper by its authors. Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0).</p>
    </sec>
    <sec id="sec-2">
      <title>Background and Related Work</title>
      <p>
        To avoid injuries in running, a special wearable assistant was created, using
an electrical muscle stimulation (EMS) device and an insole with force sensing
resistors [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. The results of the conducted study showed that EMS actuation
signi cantly outperforms traditional coaching, which implies that this type of
feedback can be bene cial for the motor learning of complex, repetitive
movements.
      </p>
      <p>
        Training of new skills by means of wearable technologies and augmented
reality is supported by a conceptual reference framework, which enables capturing
the expert's performance and provides various transfer mechanisms [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. To
capture an expert's experience with wearable sensors high-level tasks were mapped
to low-level functions (including body posture, hand/arm gestures, biosignals,
haptic feedback, and user location), which were decomposed to their associated
sensors [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ].
      </p>
      <p>
        The Visual Inspection Tool (VIT) facilitates annotation of multimodal data
as well as the processing and exploitation for learning purposes [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. The VIT
enables 1) triangulating multimodal data with video recordings; 2) segmenting
the multimodal data into time-intervals and adding annotations to the
timeintervals; 3) downloading the annotated dataset and using it for multimodal
data analysis. The tool is part of the Multimodal Learning Analytics Pipeline.
      </p>
      <p>
        To describe running motion, we rely on dynamic movement primitives (DMPs)
[
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. DMPs describe a motion as a combination of two forces: First, a dampened
spring system which counteracts any undesired disturbances over time and,
second, a time-dependent force term which is tted to the data. More speci cally,
the time dynamics of a DMP are described by the following equations.
v_ (t) =
x_ (t) = v(t);
      </p>
      <p>
        x(t) + v(t) + f (t);
where x(t) models the location of a joint at time t, v(t) the velocity of the
joint at time t, 2 R+ is a hyperparameter determining the period length of
the system, 2 R+ and 2 R+ are hyperparameters determining how fast
the dampened spring system counteracts disturbances, and f (t) is the forcing
term which models the speci cs of our motion. In DMPs, this forcing term is
always a linear combination f (t) = PkK=1 k(t) wk= PkK=1 k(t) of (learned)
coe cients wk with nonlinear basis functions k. For this work, we use the
following rhythmically repeating basis functions as suggested in [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ].
k(t) = exp h
cos(2
t
ck)
1 ;
where ck 2 [0; 2 ] is the phase shift of the kth basis function and h regulates the
width of each basis function.
      </p>
      <p>
        The main strengths of DMPs are that we can t the coe cients wk to data
via simple linear regression, and that we can replay a motion at arbitrary speed
(1)
(2)
(by adjusting ) for arbitrary long times (by executing the system in
Equation 1 for longer). DMPs have been particularly popular in robotics to mimic
human demonstrations [
        <xref ref-type="bibr" rid="ref4 ref7 ref8">4,7,8</xref>
        ] but, to our knowledge, have not yet been applied
to provide feedback to human trainees.
3
      </p>
    </sec>
    <sec id="sec-3">
      <title>Methodology</title>
      <p>The aim of our project is to support learners or trainees in developing speci c
psychomotor skills by means of immersive learning environments. The expected
solutions will combine AI approaches processing multimodal data from suitable
sensors by machine learning techniques in order to analyze performance and
detect faults, and nally automatically generating individual feedback.</p>
      <p>We consider two application cases: running and collaboration with a robot.
From the learning perspective, they are di erent. In running (the concrete aim
apparently depends on the target group and the objective, which we currently
set as healthy running for a wide public) the trainee repeats (relatively simple)
rhythmic movements again and again, but the movements of various parts of
the body should be in harmony and follow certain rules, in order not to harm
the body and perform e ectively. It suggests a behavioristic approach of
learning, when each error should be immediately reported to the person, in order to
assign the message to the corresponding movement. So when the deviation in
comparison with an optimal blueprint overcomes a threshold, a suitable
feedback is given. For example, when the person does not lift the feet properly, an
acoustic feedback is provided.</p>
      <p>On the other hand, collaboration between a human and a robot consists of
various actions on both sides, following a common aim. Here, the human actions
are typically performed by hands (e.g. in an assembly process), but in this case
the person needs to evaluate the current context and decide what to do next, e.g.
whether a micro-aim has been achieved and one can proceed with the next step.
Two types of skills are required { the ability to cooperate with the robot (e.g.
learner nudges the robot on its empty arm) and the ability to ful ll the requested
task (e.g. learner puts the lid on the box). This reminds cognitivistic learning
approaches, where formative feedback (we distinguish corrective and reinforcing
type) plays an important role, allowing trial and error. Therefore, more complex
actions need to be assessed, usually considering whether a speci c micro-aim
has been achieved. Nevertheless, immediate feedback cannot be excluded either,
especially in case of dangerous operations.</p>
      <p>What both application scenarios have in common is the necessity of a
summative feedback, which evaluates a whole (training) unit or a (work) session.
Here, di erent phases or sequences of actions can be analyzed, showing which
parts were managed well and where is a potential for improvements.</p>
    </sec>
    <sec id="sec-4">
      <title>Implementation</title>
      <p>The arti cial intelligence in our project essentially has the following tasks that
are important for the learning process:
{ Modeling templates and movement patterns to guide learners: data sets are
collected that represent expert performance in selected psychomotor
processes. With this data, machine learning processes are trained for the use
cases.
{ Detection of mistakes in the execution of movements of the learners
compared to an optimal blueprint. Any deviation above a threshold triggers
feedback.
{ Generating helpful feedback for learners: Detected errors must be processed
in such a way that learners receive starting points for improving their
processes, which they can cognitively process and implement psychomotorically.</p>
      <p>joint angles DMP coe cients</p>
      <p>Let us illustrate these tasks for the running example. We rst need to
collect expert demonstrations that cover a wide range of reasonable and healthy
styles of running. These demonstrations need to be recorded via motion capture
devices to retrieve joint angle information that abstracts from the speci c body
con guration. Further, we need to abstract from running tempo and phase shift.</p>
      <p>
        During a learning episode, we record a learner's current running, convert
it into the same representation, and compare the latest cycle of the learner's
running to the most similar expert demonstration, resulting in a measure of
deviation. If the deviation exceeds a threshold, we provide auditive, tactile, or
visual feedback, e.g. coloring the deviating limb in a virtual avatar of the
runner [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. Figure 1 displays a summary of our pipeline. To represent running, we
use rhythmic dynamic movement primitives over the joint angle representation,
which intrinsically abstracts from body shape and tempo, and is fast enough to
be applied on-the- y for each new cycle of running.
      </p>
    </sec>
    <sec id="sec-5">
      <title>Evaluation</title>
      <p>
        We evaluate our proposed representation for the running case on a data set of 33
runs of three runners recorded in the Carnegie Mellon Motion Capture Lab1. Our
aim in this experiment is to abstract from particularities of a speci c run and
to identify the runner by comparing a run against other runs in the data base.
Each run is represented by Euler angles of the right and left femur as well as the
right and left tibia. We, then, derive a summary representation of each run by
training a rhythmic dynamic movement primitive (DMP). We used K = 12 basis
functions as this was su cient to achieve a local minimum in the reconstruction
error. As hyperparameters we chose ck = 2 Kk , h = log(0:1)=(cos(2 =K) 1),
= 32, and = =4, as recommended by [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. The period length for the
DMP was chosen automatically to minimize the auto-regressive error (i.e. how
similar the signal is to itself after shifting by frames). Finally, we normalized
against phase-shifts by permuting the basis functions such that the distance to
the rst run in the data set was minimized. Figure 2 illustrates the e ect of our
representation on the data. The left plot displays the raw data, the right plot
the reconstructed signal by our DMP representation after normalizing the period
length and the phase shift. Color indicates the runner. We observe that all runs
are much better aligned in the right plot and that it is easy to distinguish the
running style in blue from the running style in red and orange via its peaks at
frames 20 and 80.
      </p>
      <p>Next, we evaluate the accuracy of identifying the runner in a leave-one-out
crossvalidation across all runs. We use a -nearest neighbor classi er as
implemented in sklearn2 with the number of neighbors varying from one to ve. To
select nearest neighbors, we use the Euclidean distance on the DMP coe cients.
As baselines, we also consider the Euclidean distance on the rst 128 frames of</p>
      <sec id="sec-5-1">
        <title>1 http://mocap.cs.cmu.edu/info.php</title>
        <p>2 https://scikit-learn.org/stable/modules/generated/sklearn.neighbors.</p>
        <p>KNeighborsClassifier.html</p>
        <p>40
]
g
ed 20
[
ed 0
t 20
u
i
l
pm 40
a
0
50
frame
100
0
50
frame
100
the raw signal (because all signals where at least 128 frames long) and dynamic
time warping as implemented in the edist package3.
In this paper, we sketched an approach to support learning psychomotor skills
using machine learning. In particular, we proposed a pipeline that compares a
learner's activity to blueprints from experts and provides feedback whenever
deviations are detected that exceed a threshold. To make this approach viable,
our comparison needs to abstract from irrelevant factors such as body shape,
tempo, or phase shifts. We evaluated dynamic movement primitives for the case
of running and did achieve a representation that was abstract enough to identify
the runner from running behavior. In future work, we wish to implement our
pipeline fully for running and for a robot interaction scenario and evaluate its
e ectiveness in supporting human learners.</p>
      </sec>
      <sec id="sec-5-2">
        <title>3 https://pypi.org/project/edist/</title>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <given-names>Di</given-names>
            <surname>Mitri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            ,
            <surname>Schneider</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            ,
            <surname>Klemke</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            ,
            <surname>Specht</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            ,
            <surname>Drachsler</surname>
          </string-name>
          , H.:
          <article-title>Read between the lines: An annotation tool for multimodal data for learning</article-title>
          .
          <source>In: Proceedings of the 9th International Conference on Learning Analytics &amp; Knowledge</source>
          . pp.
          <volume>51</volume>
          {
          <issue>60</issue>
          (
          <year>2019</year>
          ). https://doi.org/10.1145/3303772.3303776
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Hassan</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Daiber</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wiehr</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kosmalla</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          , Kruger, A.:
          <article-title>Footstriker: An ems-based foot strike assistant for running</article-title>
          .
          <source>Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies</source>
          <volume>1</volume>
          (
          <issue>1</issue>
          ),
          <volume>1</volume>
          {
          <fpage>18</fpage>
          (
          <year>2017</year>
          ). https://doi.org/10.1145/3053332
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3. Hulsmann,
          <string-name>
            <given-names>F.</given-names>
            ,
            <surname>Frank</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            ,
            <surname>Senna</surname>
          </string-name>
          ,
          <string-name>
            <given-names>I.</given-names>
            ,
            <surname>Ernst</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.O.</given-names>
            ,
            <surname>Schack</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            ,
            <surname>Botsch</surname>
          </string-name>
          ,
          <string-name>
            <surname>M.</surname>
          </string-name>
          :
          <article-title>Superimposed skilled performance in a virtual mirror improves motor performance and cognitive representation of a full body motor action</article-title>
          .
          <source>Frontiers in Robotics and AI 6</source>
          ,
          <issue>43</issue>
          (
          <year>2019</year>
          ). https://doi.org/10.3389/frobt.
          <year>2019</year>
          .00043
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Ijspeert</surname>
            ,
            <given-names>A.J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Nakanishi</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          , Ho mann, H.,
          <string-name>
            <surname>Pastor</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Schaal</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          :
          <article-title>Dynamical movement primitives: learning attractor models for motor behaviors</article-title>
          .
          <source>Neural computation 25(2)</source>
          ,
          <volume>328</volume>
          {
          <fpage>373</fpage>
          (
          <year>2013</year>
          ). https://doi.org/10.1162/NECO a 00393
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Limbu</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Fominykh</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Klemke</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Specht</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wild</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          :
          <article-title>Supporting training of expertise with wearable technologies: The WEKIT reference framework</article-title>
          .
          <source>In: Mobile and ubiquitous learning</source>
          , pp.
          <volume>157</volume>
          {
          <fpage>175</fpage>
          . Springer (
          <year>2018</year>
          ). https://doi.org/10.1007/
          <fpage>978</fpage>
          - 981-10-6144-8 10
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Magill</surname>
            ,
            <given-names>R.A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Anderson</surname>
            ,
            <given-names>D.I.:</given-names>
          </string-name>
          <article-title>The roles and uses of augmented feedback in motor skill acquisition</article-title>
          , pp.
          <volume>3</volume>
          {
          <fpage>21</fpage>
          .
          <string-name>
            <surname>Routledge</surname>
          </string-name>
          , New York, NY,USA (
          <year>2012</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Nakanishi</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Morimoto</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Endo</surname>
          </string-name>
          , G., Cheng, G.,
          <string-name>
            <surname>Schaal</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kawato</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Learning from demonstration and adaptation of biped locomotion</article-title>
          .
          <source>Robotics and Autonomous Systems</source>
          <volume>47</volume>
          (
          <issue>2</issue>
          ),
          <volume>79</volume>
          {
          <fpage>91</fpage>
          (
          <year>2004</year>
          ). https://doi.org/10.1016/j.robot.
          <year>2004</year>
          .
          <volume>03</volume>
          .003
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>Schaal</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mohajerian</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ijspeert</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Dynamics systems vs. optimal control | a unifying view</article-title>
          . In: Cisek,
          <string-name>
            <given-names>P.</given-names>
            ,
            <surname>Drew</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            ,
            <surname>Kalaska</surname>
          </string-name>
          ,
          <string-name>
            <surname>J.F</surname>
          </string-name>
          . (eds.) Computational Neuroscience:
          <article-title>Theoretical Insights into Brain Function</article-title>
          ,
          <source>Progress in Brain Research</source>
          , vol.
          <volume>165</volume>
          , pp.
          <volume>425</volume>
          {
          <fpage>445</fpage>
          .
          <string-name>
            <surname>Elsevier</surname>
          </string-name>
          (
          <year>2007</year>
          ). https://doi.org/10.1016/S0079-
          <volume>6123</volume>
          (
          <issue>06</issue>
          )
          <fpage>65027</fpage>
          -
          <lpage>9</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Sharma</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Klemke</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wild</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          :
          <article-title>Experience capturing with wearable technology in the WEKIT project</article-title>
          . In: Buchem,
          <string-name>
            <given-names>I.</given-names>
            ,
            <surname>Klamma</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            ,
            <surname>Wild</surname>
          </string-name>
          ,
          <string-name>
            <surname>F</surname>
          </string-name>
          . (eds.)
          <source>Perspectives on Wearable Enhanced Learning (WELL)</source>
          , pp.
          <volume>297</volume>
          {
          <fpage>311</fpage>
          . Springer (
          <year>2019</year>
          ). https://doi.org/10.1007/978-3-
          <fpage>319</fpage>
          -64301-4 14
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>