<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>A Human Multi-Robot Interaction Framework for Search and Rescue in the Alps</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Jonathan Cacace</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Riccardo Caccavale</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Alberto Finzi</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Vincenzo Lippiello</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Universita degli Studi di Napoli Federico II</institution>
        </aff>
      </contrib-group>
      <abstract>
        <p>In this work, we present a framework that allows a single operator to monitor and control the operations of multiple robots during Search &amp; Rescue missions in an alpine environment. This work is framed in the context of the SHERPA project whose goal is to develop a mixed ground and aerial robotic platform to support search and rescue activities in alpine scenario. In this context, the human operator is not fully dedicated to the robot control, but involved in the search and rescue mission, hence only able to partially monitor and command the robotic team. In this paper, we brie y illustrate the overall framework and describe on-the- eld tests with two drones equipped with on-board cameras.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Introduction</title>
      <p>
        We present an architecture suitable for human multi-robot interaction for search
and rescue missions in an alpine environment. This work is framed in the context
of the SHERPA project [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] whose goal is to develop a mixed ground and aerial
robotic platform supporting search and rescue (SAR) activities in an alpine
scenario. In this context, a special rescue operator, called busy genius, can monitor
and interact with a team of robots during the mission operations. In particular,
we focus on the interaction of the operator with a set of drones. In contrast with
typical human-UAVs interaction scenarios [
        <xref ref-type="bibr" rid="ref1 ref6">6, 1</xref>
        ], in place of a fully dedicated
operator, we have a rescuer which can be deeply involved in the SAR mission,
hence only able to provide incomplete and sparse inputs to the robots. In this
setting, the operator should focus his/her cognitive e ort on relevant and
critical activities (e.g. visual inspection, precise maneuvering, etc.), while relying
on the robotic autonomous system for specialized tasks (navigation, scan, etc.).
Moreover, the human should operate in proximity with the drones in hazardous
scenarios (e.g. avalanche), hence the required interaction is substantially
dissimilar to the one considered in other works where the human and co-located UAVs
cooperate in controlled indoor conditions [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]. The interaction framework
illustrated in this paper allows the operator to interact in a natural and e cient way
with the ying robots. We assume the human equipped with light and wearable
devices, such as a headset, a gesture control armband, a health monitor bracelet,
and a mobile device endowed with a touch user interface. In addition, these
devices are used by the operator to get information from the robots, monitoring
their state and retrieving relevant information about the search mission
visualizing the data acquired with the on-board robot sensors. In order to track the
human pose the headset of the operator has been equipped with standard GPS
and IMU sensors. Finally, a bracelet permits to track the operator status (GSR,
heart-rate, temperature, etc.). The proposed framework allows the operator to
monitor and control the team of robots in an adaptive manner, ranging from
a minimally supervised autonomous robotic team, when the human is busy, to
direct and docile teleoperation, when the operator takes the direct control of
one of the robots. In order to test the e ectiveness of the presented architecture,
initial on-the- eld experiments have been performed in which a co-located
human operator controls two drones equipped with an on-board cameras in order
to explore an alpine area.
      </p>
    </sec>
    <sec id="sec-2">
      <title>System Architecture</title>
      <p>The proposed architecture is depicted in Figure 1. The operator is equipped
with light wearable devices to interact with the robotic system. The output of
these devices is continuously sent to the Multimodal Human-Robot-Interaction
(MHRI ) module in order to generate new commands. If the command provided
by the human operator does not explicitly assign tasks to the robots, the
Distributed Multi-Robot Task Allocation (DMRTA) module is to decompose
abstract tasks into primitive action nding a valid allocation for the robotic team
members; when the task is already completely speci ed and allocated this is
validated with respect to resources and mission constraints. In the following we
describe the associated modules.</p>
      <p>
        Multimodal Interaction. The MHRI module interprets the operator commands
integrating inputs from multiple communication channels. For instance, either
speech- or gesture-based commands may be used to stop a drone, while vocal
commands in combination with deictic gestures can be used to specify
navigational commands (e.g. \go-there") to the co-located drones. We mainly focus on
commands suitable for interacting with the set of co-located drones during
navigation and search tasks. More speci cally, we are concerned with multimodal
communication with the drones suitable for the following purposes: robot
selection commands (e.g. \all wasps", \red wasp", \you wasp"); motion commands
(e.g. \go there", \land", etc.); search primitives (to scan an area with a speci c
search pattern); multimedia commands used to acquire data through on-board
sensors (e.g. pictures, video, environment mapping, etc.); nally, switch meta
level commands allow the operator to change the interaction mode (e.g.
highlevel commands, interactive control, teleoperation). We rely on the Julius
framework for continuous speech recognition. As for gesture recognition, we recognize
harm, hand, and tablet-based gestures. We exploit the Thalmic Myo Armband
to detect and distinguish several poses of the hand (from the electrical activity
of the muscles) and the arm (the band is endowed with a 9 DOF IMU that
permits motion capture). The results form these multiple inputs (gestures, voice,
hand, tablet, etc.) are combined into a unique interpretation of the operator
intention exploiting a late-fusion approach: single modalities are rst classi ed
and synchronized, then fused into a single interpretation using the con dence
value associated with each input data (see Figure 2, left).
Implicit Drone Selection. In order to simplify the interaction with the robotic
team, we permit both implicit and explicit robot selection. Namely, we propose
an approach where, each available robot can evaluate the probability to be the
one designated by the human for the execution of a command when the target
is not explicitly assigned by a selection command, but it can be inferred by the
context (see Figure 2, right). The robot evaluation process relies on a
multilayered architecture in which a Dynamic Bayesian Network is deployed to infer
the human intentions form the state of the robots and learned contextual and
geometrical information. More details can be found in [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ].
      </p>
      <p>
        Distributed Multi-Robot Task Allocation. The DMRTA module is responsible for
multi-robot task allocation. Speci cally, given a task to perform, this module
should nd a suitable decomposition and assignment to the robotic team. This
process takes into account di erent constraints, such as resources and capabilities
needed to perform the task (e.g. take-a-picture needs a robot equipped with a
camera), the state of the robotic system and time constraints. Moreover, the
DMRTA is also to validate the feasibility of the operator's request (see [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] for
additional details).
      </p>
      <p>
        Multi-robot Supervision and Adaptive Attentional Interface. A suitable interface
is need to lter and adapt the information presented to the busy genius through
di erent communication channels: head-set (audio), tablet (video), band
(vibrotactile). We assume that information ltering and adaptive presentation of data
are managed by an attentive interface modeled as a supervisory attentional
system that regulates contentions among multiple bottom-up stimuli (alerts,
periodic information, warnings, state changes, etc.) depending on the mission state
and the human constraints. This attentional regulation process [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] takes into
account the operative tasks of the human and the drones, the limited human
workload for each communication mode (visual, audio, vibro-tactile) along with
timing and task switching constraints. In [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] a detailed description of this
framework is provided.
      </p>
      <p>
        Testing Scenario
We now illustrate an initial on-the- eld experimentation with the proposed
framework. In the experimental setting the human operator is located in a
real alpine scenario (i.e. Pordoi Pass in Alps: 2200 meters altitude) endowed
with his/her wearable devices to control two drones (called textitGreen and Red
wasp), each equipped with a standard camera. In this set-up human monitoring
and the adaptive interface were not enabled. The goal of the operator is to
inspect two di erent regions of the area depicted in Figure 3 (Left ). The operator
exploits the on-board camera of the drones to visualize real-time video streaming
and collect pictures of the terrain [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. At the beginning of the mission the drones
are landed on the ground. The user can select the robots either via the user
interface or via vocal commands. Di erent communication channels are used to drive
the Green Wasp towards a desired location and then request the scanning of
an area. During this operation, several pictures of the terrain are autonomously
taken by the robot and provided to the operator via the user interface as show in
Figure 3 (b - e, Right ). At the same time, while the Green Wasp is autonomously
executing the scanning mission, the user can focus on the Red Wasp navigating
in Direct Control mode towards a di erent area not covered by the Green wasp,
while receiving the video streaming from the on-board camera. An example of
the interaction in Direct Control mode is shown in Figure 3 (f - h, Right ), when
this control mode is active the operator can directly generate velocity data from
the orientation of its arm. At the end of the test, all the drones are driven back
to their initial position. In Figure 3 (Left ) we illustrate the trajectories of the
green and red drone along with the associated inspected zones. The duration
of the reported ying mission is about 3 minutes, with a covered area of about
4200 m2. The two zones are adjacent and non-overlapping with a satisfactory
coverage of the area; the red zone is smaller because inspected be the human the
Direct Control mode.
      </p>
      <p>Acknowledgement. The research leading to these results has been supported
by the SHERPA project, which has received funding from the European Research
Council under Advanced Grant agreement number 600958.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Bitton</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Goldberg</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          :
          <article-title>Hydra: A framework and algorithms for mixed-initiative uav-assisted search and rescue</article-title>
          .
          <source>In: Proc. of CASE 2008</source>
          . pp.
          <volume>61</volume>
          {
          <issue>66</issue>
          (
          <year>2008</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Cacace</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Caccavale</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Finzi</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lippiello</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          :
          <article-title>Attentional multimodal interface for multidrone search in the alps</article-title>
          .
          <source>In: Proc. of SMC-16</source>
          (
          <year>2016</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Cacace</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Finzi</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lippiello</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          :
          <article-title>Implicit robot selection for human multi-robot interaction in search and rescue missions</article-title>
          .
          <source>In: Proc. of Ro-Man-16</source>
          (
          <year>2016</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Cacace</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Finzi</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lippiello</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Furci</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mimmo</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Marconi</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          :
          <article-title>A control architecture for multiple drones operated via multimodal interaction in search &amp; rescue mission</article-title>
          .
          <source>In: Proc. of SSRR-16</source>
          (
          <year>2016</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Caccavale</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Finzi</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Flexible task execution and attentional regulations in human-robot interaction</article-title>
          .
          <source>IEEE Trans. Cognitive and Developmental Systems</source>
          <volume>9</volume>
          (
          <issue>1</issue>
          ),
          <volume>68</volume>
          {
          <fpage>79</fpage>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Cummings</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bruni</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mercier</surname>
          </string-name>
          , S., Mitchell, P.J.:
          <article-title>Automation architecture for single operator, multiple uav command and control</article-title>
          .
          <source>The International Command and Control Journal</source>
          <volume>1</volume>
          (
          <issue>2</issue>
          ),
          <volume>1</volume>
          {
          <fpage>24</fpage>
          (
          <year>2007</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Doherty</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Heintz</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          , Kvarnstrom, J.:
          <article-title>High-level Mission Speci cation and Planning for Collaborative Unmanned Aircraft Systems using Delegation</article-title>
          .
          <source>Unmanned Systems</source>
          <volume>1</volume>
          (
          <issue>1</issue>
          ),
          <volume>75</volume>
          {
          <fpage>119</fpage>
          (
          <year>2013</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>Marconi</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Melchiorri</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Beetz</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pangercic</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Siegwart</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Leutenegger</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Carloni</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Stramigioli</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bruyninckx</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Doherty</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kleiner</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lippiello</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Finzi</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Siciliano</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sala</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tomatis</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          :
          <article-title>The sherpa project: Smart collaboration between humans and ground-aerial robots for improving rescuing activities in alpine environments</article-title>
          .
          <source>In: Proceedings of SSRR-2012</source>
          . pp.
          <volume>1</volume>
          {
          <fpage>4</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Sza</surname>
          </string-name>
          r, D.,
          <string-name>
            <surname>Mutlu</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Fong</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          :
          <article-title>Communication of intent in assistive free yers</article-title>
          .
          <source>In: Proc. of HRI '14</source>
          . pp.
          <volume>358</volume>
          {
          <issue>365</issue>
          (
          <year>2014</year>
          )
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>