<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>[32] P. Rashidi and A. Mihailidis. A survey on ambient-assisted living tools for older adults. IEEE Journal of
Biomedial and Health Informatics</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>SmartLobby: A 24/7 Human-Machine-Interaction Space within an O ce Environment</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Stefan Fuchs</string-name>
          <email>stefan.fuchs@honda-ri.de</email>
          <email>stefan.fuchs@honda-ri.de Manuel Muhlig Honda Research Institute Europe O enbach, Germany</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Nils Einecke</string-name>
          <email>nils.einecke@honda-ri.de</email>
          <email>nils.einecke@honda-ri.de Bram Bolder Honda Research Institute Europe O enbach, Germany</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Fabian Eisele</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Honda Research Institute Europe</institution>
          ,
          <addr-line>O enbach</addr-line>
          ,
          <country country="DE">Germany</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Technical University of Darmstadt</institution>
          ,
          <addr-line>Darmstadt</addr-line>
          ,
          <country country="DE">Germany</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2019</year>
      </pub-date>
      <volume>83</volume>
      <fpage>13</fpage>
      <lpage>15</lpage>
      <abstract>
        <p>In this work, we present the SmartLobby, an intelligent environment system integrated into the lobby of a research institute. The SmartLobby is running 24/7, i.e. it can be used any time by anyone without any preparations. The goal of the system is to conduct research in the domain of human machine cooperation. One important rst step towards this goal is a detailed human state modeling and estimation. As the system is built into the lobby area, also people having no scienti c background are interacting with the system which strongly reduces the bias in the collected data. On software side the SmartLobby mainly integrates state-of-the-art algorithms that enable a thorough analysis of human behavior and state. These algorithms constitute the fundamental basis for the development of higher level system components. Here, we present our system with its various hardware and software components. Thereby, we focus on the realization of a non-intrusive head-eye-tracking as a key component to continuously observe persons using the system. A rst multi-week lasting experiment utilizes this capability to investigate a method for content personalization, which demonstrates the e ectiveness of the whole system.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>INTRODUCTION</title>
      <p>This paper presents the SmartLobby, an intelligent environment system, built into the entrance area of a research
institute. It allows to present the institute's ongoing research and to perform research at the same time. The
SmartLobby is designed for the institute's major research topic: cooperative intelligence, as the next level of
arti cial intelligence (AI). That is, AI systems cooperating with humans in an adaptive fashion and having a
sense for responsibilities in the cooperation. To achieve this goal, it is necessary to research on anthropomorphic
interaction mechanisms as well as on innovative interactions specialized to speci c tasks. It is also important
to equip intelligent systems with competences that adapt the interaction with the human depending on the
situation. An essential prerequisite is the ability to estimate the internal and observable human states. Although
this topic list is not comprehensive it highlights that this kind of research cannot be conducted isolated inside a
lab. Especially, online learning, adaptation and personalization rely on a continuous feed with data from human
interaction.</p>
      <p>The SmartLobby presented in this paper comes with an unobtrusive appearance, which makes the ongoing
research a tangible and natural experience for visitors and employees. At the same time, the SmartLobby features
properties that support versatility and exibility as these are requirements that come along with this very agile
research domain.
2</p>
    </sec>
    <sec id="sec-2">
      <title>RELATED</title>
    </sec>
    <sec id="sec-3">
      <title>WORK</title>
      <p>
        The SmartLobby is following the concept of living lab that was coined at the MIT by Mitchell, Larson and
Pentland in 2003 [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ]. The rst living labs have been realized in smart environments to observe humans in
their usage of emerging technologies (microscopic level). Especially with the transformation from ubiquitous
computing [37] towards intelligent environments [
        <xref ref-type="bibr" rid="ref17 ref5">5, 17, 34</xref>
        ] the importance of living labs has grown. Intelligent
environments are considered to support and collaborate with humans in a non-intrusive intelligent way and to
satisfy their experience in the environment. Hence, the development of such systems requires a user-centered
research methodology in order to continuously take into account the needs and wishes of potential users as well as
the evolving real life context. Meanwhile, the concept of living labs has been tailored more generally (macroscopic
level) to enhance innovation and usability of information technology driven applications in the society. This trend
intensi ed even more with the recent progress in the eld of arti cial intelligence where seeing the potential users
as competent partners in transparent development processes may reduce qualms regarding this technology.
      </p>
      <p>
        We see intelligent-cyber-physical-systems (ICPS) [
        <xref ref-type="bibr" rid="ref25">25</xref>
        ] at the intersection of cooperative intelligence, smart
environments and robotics. Augusto et al. [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] describe intelligent environments as a synthesis of ubiquitous
computing, smart environments and arti cial intelligence. Sa otti et al. [34] went a step further and introduced
physically-embedded-intelligent-systems (PEIS-ecology) as the intersection of ubiquitous computing, AI and
robotics. In this context, arti cial intelligence is an essential prerequisite to gain autonomous self-directed
agents that can follow own goals and draw related decisions. We, however, aim at enriching the system with a
sense for the human state as well as a sense for responsibilities in the cooperation with the human. Thus, we
refer to ICPS as an intelligent environment that adheres to the principles of cooperation (see [
        <xref ref-type="bibr" rid="ref22">22</xref>
        ]).
      </p>
      <p>
        ICPS systems target at applications in various domains from health, hospitals over manufacturing and
emergency services to education. A very common application are smart homes equipped with health features also
called: ambient assisted living (AAL) [36, 32]. AAL is mainly focusing on helping elderly people to keep their
self-determination. The basic idea is to augment the home environment of elderly people with intelligent devices
and o ering remote care services that assist the elderly people in their daily live. This eld has grown
substantially due to advancements in smart sensors and devices. A survey [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] of AAL papers between the years 2004 and
2013 found that most work is concentrating on activity recognition [
        <xref ref-type="bibr" rid="ref10 ref23 ref27">10, 23, 27</xref>
        ] and monitoring of the vital status
[
        <xref ref-type="bibr" rid="ref3 ref4">4, 3, 33</xref>
        ]. Further reviews [
        <xref ref-type="bibr" rid="ref9">30, 9</xref>
        ] show that the most potential sensors in current AAL projects are cameras, in
particular RGB-D cameras. Cameras allow for analysis that is di cult to achieve with other sensors like human
behavior analysis, fall detection or emotion recognition. On the downside, the information richness of cameras
also constitute a critical privacy issue. In particular, cameras installed in sensitive areas like bathrooms are very
critical and result in low acceptance. Hence, privacy preservation is also an important focus of research. For
example [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] tackle the privacy issue by replacing the real camera images with a pre-stored background image plus
stick- gures of the current people visible. Furthermore, all necessary data for computing behavior and situation
analysis is stored locally.
      </p>
      <p>
        Another close area to our work is research on improving quality of life in working environments. However, the
setup is often restricted to sensors directly installed at each desktop [
        <xref ref-type="bibr" rid="ref20 ref24 ref3">20, 24, 3</xref>
        ] and not in an ambient fashion as
in AAL. Going more in direction of ambient assisted working is the idea of a virtual secretary presented in [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ].
Several o ces were equipped with cameras and microphones in order to recognize people and their activities. In
case a meeting situation is detected, visitors trying to enter the door or callers trying to phone someone in the
room are automatically informed. Unfortunately, false detection rates are high because of the simple heuristics
used. Furthermore, the system was mainly tailored around detecting presence and availability of persons in the
o ces.
      </p>
      <p>
        Finally, [31] analyzed two prominent AAL middleware platforms and derived a set of recommendations for
future AAL systems from this. Two major insights are that AAL systems should favor GUIs over text-based
interfaces and that the system should have a good balance between standardization and customizability. Our
system [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ] adheres to these suggestions by having a graphical touch interface and by using ROS [
        <xref ref-type="bibr" rid="ref29">29</xref>
        ] as underlying
middle-ware platform.
3
      </p>
    </sec>
    <sec id="sec-4">
      <title>SMARTLOBBY SYSTEM</title>
      <p>The following section describes the requirements and the realization of the the overall system with its structural
concept, selected sensors, e ectors and software.
3.1</p>
      <sec id="sec-4-1">
        <title>General Requirements</title>
        <p>Our goal was to design a room that has both a representative character and an intelligent behavior according to
the concept of ICPS. The intelligence of the SmartLobby is not meant to be a static demonstration of research
results but rather to be continuously developed further by interaction. That is, there are challenges arising
from the users with their heterogeneous background and privacy considerations as well as regarding robustness,
operability and versatility. The daily usage by expert and non-expert (e.g. administration or guests) users in
a standard o ce environment provides the necessary insights to actually generate intelligent systems with a
natural and intuitive human-machine-interaction.</p>
        <p>The overall impression of the room should avoid usual lab impressions such as cables, extensive sensor rigs
or prototypical electronics. A presentation software has to present the institute with its structure, employees,
connections to other institutes and research projects. Additionally, the presentation software should welcome
visitors or announce talks embedded into a weekly press review. The presentation software should be intuitive
and self-explaining in order to encourage a usage by guests and novices.</p>
        <p>One major target is the 24/7 operation of the system's basic functionalities, i.e. the presentation software,
people tracking, live feedback, and head-eye-tracking have to run continuously. Furthermore, a fundamental
requirement is versatility and extensibility of the hardware installation to accommodate to new research
approaches.
3.2</p>
      </sec>
      <sec id="sec-4-2">
        <title>Structural Concept</title>
        <p>The location we selected for the SmartLobby is an essential junction of the institute's traveling paths, because
it connects two building parts with the employee o ces and laboratories. With administration and management
o ces, a co ee kitchen and two meeting rooms directly connected, this area is frequently visited by employees
and visitors alike. Fig. 1 gives an image of the o ce lobby.</p>
        <p>The room has a size of 6 m 5.5 m and features ve doors and two open gateways. A large bar table extends
from the co ee kitchen's service hatch into the room. The bar table invites to meet, interact with the system
and operate the touch screens that are embedded into it.</p>
        <p>The wooden material of the walls generates a comfortable atmosphere while the lamella design maintains a
more technical touch. In the upper part the lamella distance is increased to make them more transparent for
light from the adjacent rooms.</p>
        <p>Sensors and actors are either directly built into the wooden wall or attached to it. Power and network cables
are hidden inside the wall to support a living room atmosphere. In order to keep the room exible, the walls
feature various hidden extension slots that can house additional sensors, e ectors or computation hardware.
However, major computing hardware is installed remotely in a server room.
3.3</p>
      </sec>
      <sec id="sec-4-3">
        <title>Sensors and E ectors</title>
        <p>Fig. 3a shows a top view of the SmartLobby and its sensors. The room is equipped with seven imaging devices.
There are three Kinects (XBox One) for providing depth measurements, three ceiling-mounted RGB cameras
(IDS UI-5250CP-C-HQ, lens 6 mm) in the corners for a complete room overview, and a
pan-tilt-zoom-(PTZ)camera (Axis Q6128, lens 3.9 mm to 46.8 mm) for high delity scans. At a distance of 6 m the PTZ-camera still
captures images with a spatial resolution of 5 pix/mm. The room is also equipped with two microphone arrays
(depicted by small green circles in Fig. 3a) each consisting of eight pressure zone microphones (AKG C 562
CM, Beyerdynamic Classis BM32W). Their arrangement enables sound localization and highly sensitive speech
recognition.</p>
        <p>The most straightforward interaction with the system is the usage of the displays. At the heart of the
interaction is a multi-touch-table (MTT) that comprises two touch displays (scape tangible 55) each having
a UHD resolution (3840 pix 2160 pix) at 55 inch screen diagonal. The left touch display runs a proprietary
software for interactively presenting the research institute. It can be operated not only by nger touches but also
by physical passive and active tokens that trigger certain information. There is one wall mounted display close
to the MTT with a 84 inch screen diagonal. The wall along the main passage features two ultra-wide displays
each with a size of 2.16 m 0.35 m at a resolution of 3840 pix 600 pix. This installation enables the interactive
control of the displays as a feedback for persons walking by. Two speakers beside the large display allow for
playing back audio and video data as well as performing speech-based interaction.</p>
        <p>A strong focus was also put on the light installation. The lamellas in the walls are partially replaced by
multi-color LED stripes. These LED stripes are grouped into eight segments (two for each wall), which can be
controlled independently. The ceiling lights provide white light with controllable intensity. The light is driven
by DMX protocol and accessible via Ethernet by a LANBox-CLX device.
3.4</p>
      </sec>
      <sec id="sec-4-4">
        <title>Mobile Robot</title>
        <p>
          The SmartLobby covers only a small area of the institute. As a straightforward extension further research
projects involve mobile robots that are permanently active in the whole o ce area, e.g. for guiding people.
Together with these robots the SmartLobby connects to a multi-entity system that shares information among
the di erent entities. For example the robots act as mobile extended sensors for the SmartLobby. In return
the SmartLobby feeds user information gathered in user interactions back to the robots and potentially other
entities. A more detailed description of the mobile robots and their usage is described in [
          <xref ref-type="bibr" rid="ref13 ref16">16, 13</xref>
          ].
3.5
        </p>
      </sec>
      <sec id="sec-4-5">
        <title>Software</title>
        <p>
          The large variety of sensors and the generated data cannot be processed on a single computer. Thus, there is the
need for a versatile middleware that allows for distributed processing as well as fast integration of new sensors and
algorithms. We decided to use ROS [
          <xref ref-type="bibr" rid="ref29">29</xref>
          ], which comes with a lot of available sensor packages (e.g. iai_kinect2,
ueye_cam) and deployable applications (e.g. OpenPTrack). Also ROS provides means for web-based usage that
(b) Screen Patterns
(e) Desaturation
(f) Rig Detection
(g) Daily Calibration
(h) Fused Clouds
is independent of the operating systems involved (ros_bridge). Hence, ROS can be easily integrated with any
kind of hardware as long as an HTML5-renderer is available.
        </p>
        <p>
          In order to assure robustness of our multi-entity system, the mobile robots and the smart lobby are organized
in separate ROS systems with separate ROS masters. This ensures that the robots can disconnect and reconnect
at any time without the need of restarting the overall system. To share speci c topics and services across the
multiple ROS systems, the multimaster_fkie package [
          <xref ref-type="bibr" rid="ref1">1</xref>
          ] is used.
4
        </p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>SMARTLOBBY</title>
    </sec>
    <sec id="sec-6">
      <title>MODULES</title>
      <p>The variety of ROS software modules deployed in the SmartLobby can be categorized into: infrastructure modules
to control the sensors and e ectors, maintenance modules to guarantee the 24/7-operation, and application
modules. The following sections describe selected software modules from the latter two categories. Thereby, we
focus on modules that enable and foster accurate human monitoring to realize a personalization application.
4.1</p>
      <sec id="sec-6-1">
        <title>Automatic Camera Calibration</title>
        <p>After the installation of the cameras an extrinsic calibration is necessary to unify the coordinate systems to
merge (see Fig. 2h) or to transfer information amongst the sensors. In a rst attempt, we calibrated the cameras
with a standard checkerboard calibration rig (see Fig. 2a). Unfortunately, the cameras are subject to slight
movements, due to ball joints in the camera mount and lightweight dry-walls which vibrate slightly when a door
is closed. A regular recalibration would be necessary but is very time consuming when using a manually placed
rig, especially because multiple rigs are necessary in order to cover all cameras. Hence, we devised an automatic
calibration, which makes use of the installed monitors.</p>
        <p>The basic idea is displaying the calibration patterns on the screens. Since the screen positions are very stable
their exact position needs to be measured only once. In order to easily distinguish the screens we use di erent
pattern colors as shown in Fig. 2b. There are, however, some details that need to be considered. Firstly, the
calibration patterns shouldn't be occluded. For this reason, we decided to trigger the calibration at 1 am, when
people are rarely present. Secondly, there is a problem with the monitors' re ective properties. The cameras
might see a re ection from the lighting which overcasts the calibration pattern (see Fig. 2c) and thus prevents a
clean calibration.</p>
        <p>Automatically switching o the light before calibration prevents re ections but introduces another problem. As
the room is dark and the screens are bright, the automatic camera exposure will be high and lead to overexposed
pixels (see Fig. 2d). To account for that, we reduce the screen brightness to 15 % during calibration. Finally, we
observed some desaturation e ects when having a camera at a steep angle relative the the calibration screens.
As Fig. 2e shows, this desaturation is not homogeneous. The calibration pattern is green on the left side (far
0
0.0
0.5 1.0 1.5
velocity / m/s
2.0
(a) Cameras and Tracks
(b) Speed Histogram
side) of the display but is almost white on the right (close side). We counteract this mainly by reducing the
pattern extent on the ultra-wide screens.</p>
        <p>
          The actual calibration procedure is done with OpenCV [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ] (see Fig. 2f). Fig. 2g shows the extrinsic parameter
measurement of the center Kinect over the course of 296 days. In this plot a notable drift in orientation is
observable. Furthermore, the plot shows some events of strong perturbations. These were caused by maintenance
work on the pan-tilt-zoom unit which is located on the ceiling above the center Kinect.
        </p>
        <p>In order to measure the accuracy of the automatic calibration we generated a manual Ground Truth calibration
by manually annotating the corners of the screen calibration patterns in the camera images. We found the
automatic calibration error to be between 0.36 and 0.64 for camera orientations and between 0.1 m and 0.4 m
for camera positions. The eligibility of the manual Ground Truth was judged with a simple experiment: We
assumed that humans are able to localize corner points with an accuracy of at least 1 pix. By jittering the human
screen corner point annotations by 1 pix we found that the average accuracy of the manual calibration is about
0.01 m in position and 0.13 in orientation. Since the average error caused by 1 pix jitter is much smaller than
the di erence between automatic and manual calibration, we can regard the manual calibration as a reasonable
ground truth.
4.2</p>
      </sec>
      <sec id="sec-6-2">
        <title>People Tracking</title>
        <p>
          For tracking people in the SmartLobby, we employ OpenPTrack [
          <xref ref-type="bibr" rid="ref26">26</xref>
          ]. One important aspect of OpenPTrack is
its support for distributed computing. In our setup, we use one ZBox (Zotac Magnus-1070k, i5-7500T CPU @
2.70GHz Quadcore) for each Kinect. OpenPTrack allows to run point cloud generation and detection on the
ZBox while fusion and tracking is done on one of the central SmartLobby computers. Hence, the computational
load is highly distributed and data bandwidth is reduced as the data-heavy point clouds are not sent over the
network.
        </p>
        <p>Fig. 3a shows that the arrangement of Kinects allows to track humans in the lobby on all reasonable paths.
During one morning about 120 persons pass the SmartLobby. The histogram of average walking speeds
corresponds to the typical speed of 1.4 m/s (see red line in Fig. 3b).</p>
        <sec id="sec-6-2-1">
          <title>OpenPTrack</title>
        </sec>
        <sec id="sec-6-2-2">
          <title>Head-Eye-Tracking</title>
          <p>gaze,
head pose,
ROI</p>
        </sec>
        <sec id="sec-6-2-3">
          <title>Face</title>
        </sec>
        <sec id="sec-6-2-4">
          <title>Recognition</title>
          <p>id
gaze,
head pose</p>
        </sec>
        <sec id="sec-6-2-5">
          <title>Gaze</title>
        </sec>
        <sec id="sec-6-2-6">
          <title>Calibration</title>
          <p>calibrated</p>
          <p>gaze,
head pose
id AOI</p>
        </sec>
        <sec id="sec-6-2-7">
          <title>Analysis</title>
          <p>AoI
Fixations</p>
        </sec>
        <sec id="sec-6-2-8">
          <title>Slide Control / Statistics</title>
          <p>tracks, track ids
stathee,atrdacpkosied
image, intrinsics,
extrinsics</p>
          <p>gaze,
head pose</p>
        </sec>
        <sec id="sec-6-2-9">
          <title>Image</title>
        </sec>
        <sec id="sec-6-2-10">
          <title>Server</title>
          <p>track from</p>
          <p>OpenPTrack
Adapt Content</p>
          <p>
            velocities
pan, tilt, zoom
pan, tilt, zoom,
image
Head pose and eye gaze are regarded as a major cue to infer the internal human state [
            <xref ref-type="bibr" rid="ref2">2</xref>
            ]. In the context of smart
environments usually vision based remote eye tracking systems are applied, because of their non-intrusiveness and
their possibility to promptly interact with the system. Gaze estimation relies on qualitative images, which results
in a con ned tracking box. The enhancement of the tracking box is an integral challenge of such systems. There
are systems with movable cameras (rays, drone), systems that try to cope with less resolution and pan-tilt-zoom
cameras.
          </p>
          <p>We decided to integrate an Axis PTZ-camera, which is placed above the large display as it is the most
prominent information hub in the room (see Fig. 4). Thus, it is possible to track all humans that enter the area
and capture their glances while watching the screen. Two threaded processes retrieve the images and forward
them as ROS messages at 40 Hz and a latency of 75 ms. We developed a simple kinematic model of the
PTZcamera given the assumption that the image plane is moving on a sphere located in camera center. Here, we used
a common checkerboard pattern to retrieve the anchor point of the camera, the radius of the sphere, and the
pan o set. Given the possibility to change the camera's focal length from 3.9 mm to 46.8 mm with a resolution
of 10000 steps there could be the same amount of intrinsic calibrations. The high quality optics allows us to
neglect the distortion and we set the principal point to the center of the image. Furthermore, we estimated the
actual focal length (camera scale) for ve operation points and generated a look-up-table.</p>
          <p>The PTZ-camera provides an extensive REST-API (VAPIX) to monitor, parametrize and control the device.
We implemented a closed-loop PI-controller, which controls the camera orientation and its zoom by adjusting
the velocities of pan, tilt, and zoom. The controller runs at 20 Hz and its major parameters (Kp and Tn) have
been optimized to capture an entering human within maximal 1.5 s. The maximum speed of the PTZ-camera is
set to 90 =s.</p>
          <p>
            The head-eye-tracking-(HET) uses tracking messages from OpenPTrack as an initial guess for a head position
in order to follow the people's faces with the PTZ camera. Head pose and eye gaze are estimated by means
of the dlib-library [
            <xref ref-type="bibr" rid="ref18 ref19">18, 19</xref>
            ] and GazeML [
            <xref ref-type="bibr" rid="ref28">28</xref>
            ] with an average accuracy of 10 . This accuracy is su cient to
estimate grid-wise areas-of-interest (AoIs), which are used by other applications, such as the personalization (see
Fig. 7c).
automatic
enrollment
face
recognition
          </p>
          <p>EnFcaocdeing EnFcaocdeing EnFcaocdeing
approoprireiantteatsihonar,pgnaezses,qburailgithytness
-1 2 2 2
Query Each Single Encoding</p>
          <p>EnFcaocdeing EnFcaocdeing EnFcaocdeing EnFcaocdeing
OPT0 OPT0</p>
          <p>OPT1 OPT1 OPT1 OPT1 OPT1 OPT1
(a) Enrollment Process</p>
          <p>EncEondFcinaocgdeing 2</p>
          <p>Face
1Face</p>
          <p>Face
EncEondciEnongdFcinaocgdeing</p>
          <p>EncoFdaicnFegEaFncaecoeding3
EncEondcinogdFinacge
FaceEncoding
OPT0
already
known
create new
face model
augment
face model
200
150
e
l
p
o
e
100 p
50
0.0MTWTFMTWTFMTWTFMTWTFMTWTFMTWTFMTWTFMTWTFMTWTFMTWTFMTWTFMTWTF0</p>
          <p>week day
(b) Unknown faces vs. unfolding of face-id-database</p>
        </sec>
      </sec>
      <sec id="sec-6-3">
        <title>Face Recognition</title>
        <p>Recognizing individuals entering the SmartLobby is an inevitable capability to implement personalized behavior.
The most common and non-intrusive approach is face recognition. Usually, face recognition is encompassed by
an enrollment process wherein a user deliberately provides one or more images of his face together with his name.
In spirit of tighter data protection regulations and its uncertain reading, we decided to implement an anonymized
face recognition. The system learns new faces in an unsupervised fashion without associating the actual names
of the users and still providing unique id numbers.</p>
        <p>
          Our face recognition system is based on the dlib-library [
          <xref ref-type="bibr" rid="ref19">19</xref>
          ] and the HET-module outcomes. Two processes
are running in parallel (see Fig. 5a): First, every time the HET-module gives a head pose, the system tries to
recognize a face. To this end it passes an image cutout of the head pose to the face recognition of the dlib-library,
which transforms these into a 128-dimensional face-feature-vector. This measured feature vector is used to search
for the most similar face model in the database. The system computes the L2-norm of the measured face feature
and the stored ones. A match is found for the pair with the minimal norm that is above a certain threshold
(1 L2-norm). Second, all face RoIs of a coherent head tracking are ltered according to particular requirements
(brightness, orientation, sharpness). The remaining RoIs are assumed to relate to a single individual and the
facial-feature-vector is computed for each RoI. By using the aforementioned search method a matching face
feature is added to the database (augmentation) or, if not available, a new face model is created (enrollment).
        </p>
        <p>
          These two processes rely on di erent con dence thresholds that have been decided on the basis of face
recognition performance measures. According to [
          <xref ref-type="bibr" rid="ref15">15</xref>
          ] and [35] we identi ed the false non-match rates (FNMR), false
match rates (FMR), false acceptance rates (FAR), and false rejection rates (FRR). Fig. 6a shows the comparison
of the dlib-face recognition performance with the best approach from a recent NIST evaluation. At a FMR of
0.1 % genuine facial features are classi ed as unknown for 27 % of the queries (FNMR). This is a magnitude higher
than state-of-the-art (2.5 % FNMR). It is representative only for our setup and potentially caused by imperfect
lighting conditions and unfavorable viewing angles. Fig. 6b is used to decide on the con dence thresholds for the
recognition and unsupervised augmentation processes. Usually, a threshold is selected that gives equal FARs and
FRRs. The developers of the dlib-face recognition software recommend a threshold of 0.4 (light green dots). In
our setup, we used for the regular face recognition and the initial enrollment a threshold of 0.475 , which yields
a FRR and FAR of 9 %. For the unsupervised augmentation, we used a more strict threshold of 0.6 . Thus, we
can prevent the face model database from adding incorrect facial features and mixing up individuals.
        </p>
        <p>Fig. 5b shows the face database unfolding over a period of 12 weeks. Currently, approximately 100 researchers
are with the institute and should regularly pass-by the SmartLobby. The automatic enrollment prefers to enrich
the database with persons that are captured by the HET-system, i.e. persons that either are present in the
lobby or interact with the screens. Over time, the number of enrolled faces continuously increases, whereas
the percentage of unknown faces reduces continuously. It can be observed that the system generates more IDs
than the expected 100 researchers. There are two main reasons: First, the system also captures guests and
maintenance sta . Second, the system sometimes creates more than one instance for a person, which needs to
be improved in the future.
0.7
0.6
A major task of the SmartLobby is to represent the institute and to visualize ongoing research activities with
their achievements. We targeted for a system that can express the creativity and innovative strength of the
institute by using state-of-the-art devices and recent technology. As a result, we decided to use multi-touch
displays (MTD) as interaction hubs, which not only react on touch gestures but also recognize physical tokens
placed on the displays.</p>
        <p>The latter capability is a basic feature of the presentation software easire (developed by interactive scape),
which govern the left MTD and the large display. For example, easire can present the structure of the institute,
CVs of the people working at the institute or give information about running and nished projects. Fig. 7a
depicts a view when standing in front of the left MTD. The large, upright screen can either show a duplicate
of the content of the left MTD or additionally display videos, documents and charts. This feature is especially
useful when presenting information to a larger audience, because the horizontal touch screen is hardly visibly
from farther away due to an unfavorable view perspective. We found that people enjoyed this kind of presentation
much more than dusty old slides. Furthermore, the token concept allows them to easily explore the things they
are interested in. The right MTD is set aside for RViz, which visualizes ROS topics, such as point clouds (see
Fig. 2h), robot trajectories or results of our algorithms.
4.6</p>
      </sec>
      <sec id="sec-6-4">
        <title>Personalization Experiment</title>
        <p>During idle times a slide show welcomes visitors, announces talks and comprises a weekly updated press review.
Typically a dozen pages are repeatedly shown, each for 2 min. Given the face recognition and the AoI analysis,
the system can memorize the slides a person has already seen. A rst multi-week lasting experiment investigates
a basic content personalization method targeted at all kinds of visitors to the SmartLobby. We divided the
people into two groups, one where the system intervenes in the repetitive slide process and one where it does
not. The groups are split randomly by intervening only for people with an even ID. For the intervention group,
the system skips the slide show to pages that have not been regarded by the person yet. This means, if the
current slide, when entering the room, has not been seen by the person yet, the system will also not switch to
another slide for people in the intervention group. Fig. 7b shows the impact of the slide-skipping intervention.
The analysis of the intervention group is split into events where no intervention happened (slide was not yet
known anyway) and actual skipping events. As one can expect, there is no signi cant di erence between the
control group (no intervention) and the events of the intervention group where no intervention happened (slide</p>
        <p>Control Exp. Group Exp. Group
Group w/o inter- with
intervention vention
(a) Tokens trigger the presentation of various contents
(b) Personalization
(c) Glance statistics of two particular slides
unknown yet). The median viewing time is similar for both groups with about 1 s. This seems rather short,
but it includes also people passing by without reading the screen. In contrast, skipping the pages attracts the
attention of the visitors and increases the mean viewing time tremendously to 4 s.</p>
        <p>
          Furthermore, we investigated the reading behavior for di erent content categories: tech news and publication
announcements. Fig. 7c shows the glance frequency distribution for the publication slides. The gaze transition
entropy [
          <xref ref-type="bibr" rid="ref21">21</xref>
          ] is a promising criteria to measure interest in content: Though the median entropy is similar for the
two categories with 2.8 , the entropy variance for the news slides is more than twice as big. This demonstrates,
that di erent news topics cause a larger spread in the gaze transition entropy values of the readers.
5
        </p>
      </sec>
    </sec>
    <sec id="sec-7">
      <title>CONCLUSION AND OUTLOOK</title>
      <p>We have presented an intelligent environment integrated into the lobby of a research institute. We described
how to equip a specialized public space with ubiquitious technologies, so as to showcase research results as well
as to serve as a workbench for collecting data for research in the domain of human machine cooperation. This
so-called SmartLobby is available 24/7. The interior design, the hardware and software infrastructure allow for
versatility and extensibility. Thus, we can quickly set up experiments and conduct unobtrusive user studies in
an every-day environment.</p>
      <p>For a start, we developed a head-eye-tracking system, that covers the whole lobby, and an anonymous face
recognition to enable long-term user monitoring in the SmartLobby. On this basis, we equipped the lobby with
a simple intelligent behavior and personalized the news feed shown in the lobby. Our results show the positive
personalization impact by increasing the median viewing time. Furthermore, we investigated gaze transition
entropy as a measure for interest in the news feed slides. As a next step, we will enhance the personalization by
also considering the personal interest.</p>
    </sec>
    <sec id="sec-8">
      <title>ACKNOWLEDGMENT</title>
      <p>We would like to thank studio klv GmbH for designing and managing the reconstruction of the lobby and
Interactive Scape GmbH for implementing the presentation software as well as the token and touch user-interface.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>[1] http://wiki.ros.org/multimaster_fkie.</mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>A.</given-names>
            <surname>Al-Rahayfeh</surname>
          </string-name>
          and
          <string-name>
            <given-names>M.</given-names>
            <surname>Faezipour</surname>
          </string-name>
          .
          <article-title>Eye tracking and head movement detection: A state-of-art survey</article-title>
          .
          <source>IEEE Journal of Translational Engineering in Health and Medicine</source>
          ,
          <volume>1</volume>
          ,
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>A.</given-names>
            <surname>Alberdi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Aztiria</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Basarab</surname>
          </string-name>
          , and
          <string-name>
            <given-names>D. J.</given-names>
            <surname>Cook</surname>
          </string-name>
          .
          <article-title>Using smart o ces to predict occupational stress</article-title>
          . In
          <source>International Journal of Industrial Ergonomics</source>
          , volume
          <volume>67</volume>
          , pages
          <fpage>60</fpage>
          {
          <fpage>66</fpage>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Andreu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Chiarugi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Colantonio</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Giannakaki</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Giorgi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Henriquez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Kazantzaki</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Manousos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Maria</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B. J.</given-names>
            <surname>Matuszewskia</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. A.</given-names>
            <surname>Pascali</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Pediaditis</surname>
          </string-name>
          , G. Raccichini, and
          <string-name>
            <given-names>M.</given-names>
            <surname>Tsiknakis</surname>
          </string-name>
          .
          <article-title>Wize Mirror - a smart, multisensory cardio-metabolic risk monitoring system</article-title>
          .
          <source>In CVIU</source>
          , pages
          <volume>3</volume>
          {
          <fpage>22</fpage>
          ,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>J. C.</given-names>
            <surname>Augusto</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Callaghan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Cook</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Kameas</surname>
          </string-name>
          ,
          <string-name>
            <surname>and I. Satoh.</surname>
          </string-name>
          <article-title>Intelligent Environments: a manifesto</article-title>
          .
          <source>Human-centric Computing and Information Sciences</source>
          ,
          <volume>3</volume>
          :
          <fpage>12</fpage>
          ,
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>G.</given-names>
            <surname>Bradski. The OpenCV Library. Dr</surname>
          </string-name>
          .
          <source>Dobb's Journal of Software Tools</source>
          ,
          <year>2000</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>D.</given-names>
            <surname>Calvaresi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Cesarini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Sernani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Marinoni</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. F.</given-names>
            <surname>Dragoni</surname>
          </string-name>
          ,
          <article-title>and</article-title>
          <string-name>
            <given-names>A.</given-names>
            <surname>Sturm</surname>
          </string-name>
          .
          <article-title>Exploring the ambient assisted living domain: a systematic review</article-title>
          .
          <source>Journal of Ambient Intelligence and Humanized Computing</source>
          ,
          <volume>8</volume>
          (
          <issue>2</issue>
          ):
          <volume>239</volume>
          {
          <fpage>257</fpage>
          ,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8] CherryHome. Cherryhome.
          <source>ai website</source>
          ,
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>S.</given-names>
            <surname>Colantonio</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Coppini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Giorgi</surname>
          </string-name>
          , M.
          <article-title>-</article-title>
          <string-name>
            <surname>A. Morales</surname>
            , and
            <given-names>M. A.</given-names>
          </string-name>
          <string-name>
            <surname>Pascali</surname>
          </string-name>
          .
          <article-title>Computer vision for ambient assisted living: Monitoring systems for personalized healthcare and wellness that are robust in the real world and accepted by users, carers, and society</article-title>
          . In Computer Vision for Assistive Healthcare, pages
          <volume>319</volume>
          {
          <fpage>336</fpage>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>R.</given-names>
            <surname>Cucchiara</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Prati</surname>
          </string-name>
          , and
          <string-name>
            <given-names>R.</given-names>
            <surname>Vezzani</surname>
          </string-name>
          .
          <article-title>A multi-camera vision system for fall detection and alarm generation</article-title>
          .
          <source>Expert Systems</source>
          ,
          <volume>24</volume>
          (
          <issue>5</issue>
          ):
          <volume>334</volume>
          {
          <fpage>345</fpage>
          ,
          <year>2007</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>M.</given-names>
            <surname>Danninger</surname>
          </string-name>
          and
          <string-name>
            <given-names>R.</given-names>
            <surname>Stiefelhagen</surname>
          </string-name>
          .
          <article-title>A context-aware virtual secretary in a smart o ce environment</article-title>
          .
          <source>In Proceedings of the 16th ACM International Conference on Multimedia</source>
          , pages
          <volume>529</volume>
          {
          <fpage>538</fpage>
          ,
          <year>2008</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>M.</given-names>
            <surname>Eriksson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Niitamo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Kulkki</surname>
          </string-name>
          , and
          <string-name>
            <given-names>K.</given-names>
            <surname>Hribernik</surname>
          </string-name>
          .
          <article-title>Living labs as a multi-contextual R&amp;D methodology</article-title>
          .
          <source>The 12th ITM Conference</source>
          ,
          <year>January 2006</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>L.</given-names>
            <surname>Fischer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Hasler</surname>
          </string-name>
          , J. Deigmoller, T. Schnurer, M. Redert,
          <string-name>
            <given-names>U.</given-names>
            <surname>Pluntke</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Nagel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Senzel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Richter</surname>
          </string-name>
          , and
          <string-name>
            <given-names>J.</given-names>
            <surname>Eggert</surname>
          </string-name>
          .
          <article-title>Where is the tool? - grounded reasoning in everyday environment with a robot</article-title>
          .
          <source>In International Cognitive Robotics Workshop (CogRob)</source>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>S.</given-names>
            <surname>Fuchs</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Einecke</surname>
          </string-name>
          , and
          <string-name>
            <given-names>F.</given-names>
            <surname>Eisele</surname>
          </string-name>
          .
          <article-title>Smart-lobby: Using a 24/7 remote head-eye-tracking for content personalization</article-title>
          .
          <source>In Proceedings of UbiComp/ISWC</source>
          ,
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>P.</given-names>
            <surname>Grother</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Ngan</surname>
          </string-name>
          , and
          <string-name>
            <given-names>K.</given-names>
            <surname>Hanaoka</surname>
          </string-name>
          .
          <article-title>Ongoing face recognitionvendor test (frvt)</article-title>
          .
          <source>Technical report, NIST National Institute of Standards and Technology</source>
          ,
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>S.</given-names>
            <surname>Hasler</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Kreger</surname>
          </string-name>
          , and
          <string-name>
            <given-names>U.</given-names>
            <surname>Bauer-Wersing</surname>
          </string-name>
          .
          <article-title>Interactive incremental online learning of objects onboard of a cooperative autonomous mobile robot</article-title>
          .
          <source>In Neural Information Processing</source>
          , pages
          <volume>279</volume>
          {
          <fpage>290</fpage>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <surname>M. B. James Dooley</surname>
            and
            <given-names>M. R.</given-names>
          </string-name>
          <string-name>
            <surname>Al-Mulla</surname>
          </string-name>
          .
          <article-title>Beyond four walls: Towards large-scale intelligent environments</article-title>
          .
          <source>pages 287{294</source>
          ,
          <year>2012</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>V.</given-names>
            <surname>Kazemi</surname>
          </string-name>
          and
          <string-name>
            <given-names>J.</given-names>
            <surname>Sullivan</surname>
          </string-name>
          .
          <article-title>One millisecond face alignment with an ensemble of regression trees</article-title>
          .
          <source>June</source>
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>D. E.</given-names>
            <surname>King</surname>
          </string-name>
          .
          <article-title>Dlib-ml: A machine learning toolkit</article-title>
          .
          <source>Journal of Machine Learning Research</source>
          ,
          <volume>10</volume>
          :
          <fpage>1755</fpage>
          {
          <fpage>1758</fpage>
          ,
          <year>2009</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>K.</given-names>
            <surname>Kiyokawa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Hatanaka</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Hosoda</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Okada</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Shigeta</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Ishihara</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Ooshita</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Kakugawa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Kurihara</surname>
          </string-name>
          , and
          <string-name>
            <given-names>K.</given-names>
            <surname>Moriyam</surname>
          </string-name>
          .
          <article-title>Owens luis a context-aware multi-modal smart o ce chair in an ambient environmen</article-title>
          .
          <source>In IEEE VR Workshops (VRW)</source>
          , pages
          <fpage>1</fpage>
          <issue>{4</issue>
          ,
          <year>2012</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>K.</given-names>
            <surname>Krejtz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Duchowski</surname>
          </string-name>
          , and
          <string-name>
            <given-names>N.</given-names>
            <surname>Villalobos</surname>
          </string-name>
          .
          <article-title>Gaze transition entropy</article-title>
          .
          <source>ACM Trans. Appl</source>
          . Percept.,
          <volume>13</volume>
          (
          <issue>1</issue>
          ),
          <year>December 2015</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>M.</given-names>
            <surname>Kru</surname>
          </string-name>
          <article-title>ger, C. B</article-title>
          .
          <string-name>
            <surname>Wiebel</surname>
            , and
            <given-names>H.</given-names>
          </string-name>
          <string-name>
            <surname>Wersing</surname>
          </string-name>
          .
          <article-title>From tools towards cooperative assistants</article-title>
          .
          <source>In Proceedings of the 5th International Conference on Human Agent Interaction</source>
          , pages
          <volume>287</volume>
          {
          <fpage>294</fpage>
          ,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <given-names>D.</given-names>
            <surname>Lymberopoulos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Bamis</surname>
          </string-name>
          ,
          <article-title>and</article-title>
          <string-name>
            <given-names>A.</given-names>
            <surname>Savvides</surname>
          </string-name>
          .
          <article-title>Extracting spatiotemporal human activity patterns in assisted living using a home sensor network</article-title>
          .
          <source>Universal Access in the Information Society</source>
          ,
          <volume>10</volume>
          (
          <issue>2</issue>
          ):
          <volume>125</volume>
          {
          <fpage>13</fpage>
          ,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [24]
          <string-name>
            <given-names>D.</given-names>
            <surname>McDu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Karlson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Kapoor</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Roseway</surname>
          </string-name>
          , and
          <string-name>
            <given-names>M.</given-names>
            <surname>Czerwinski</surname>
          </string-name>
          .
          <article-title>A ectaura: An intelligent system for emotional memor</article-title>
          .
          <source>In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems</source>
          , pages
          <fpage>849</fpage>
          {
          <fpage>858</fpage>
          ,
          <year>2012</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          [25]
          <string-name>
            <given-names>H. A.</given-names>
            <surname>Mu</surname>
          </string-name>
          <article-title>ller. The rise of intelligent cyber-physical systems</article-title>
          .
          <source>Computer</source>
          ,
          <volume>50</volume>
          (
          <issue>12</issue>
          ):
          <volume>7</volume>
          {
          <issue>9</issue>
          ,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          [26]
          <string-name>
            <given-names>M.</given-names>
            <surname>Munaro</surname>
          </string-name>
          and
          <string-name>
            <given-names>E.</given-names>
            <surname>Menegatti</surname>
          </string-name>
          .
          <article-title>Fast RGB-D people tracking for service robots</article-title>
          .
          <source>Journal on Autonomous Robots</source>
          ,
          <volume>37</volume>
          (
          <issue>3</issue>
          ):
          <volume>227</volume>
          {
          <fpage>242</fpage>
          ,
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          [27]
          <string-name>
            <given-names>S.</given-names>
            <surname>Pandya</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Ghayva</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Kotecha</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Awais</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Akbarzadeh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Gope</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. C.</given-names>
            <surname>Mukhopadhyay</surname>
          </string-name>
          , and
          <string-name>
            <given-names>W.</given-names>
            <surname>Chen</surname>
          </string-name>
          .
          <article-title>Smart Home Anti-Theft System: A Novel Approach for Near Real-Time Monitoring and Smart Home Security for Wellness Protocol</article-title>
          .
          <source>In Applied System Innovations</source>
          , volume
          <volume>1</volume>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          [28]
          <string-name>
            <given-names>S.</given-names>
            <surname>Park</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Spurr</surname>
          </string-name>
          , and
          <string-name>
            <given-names>O.</given-names>
            <surname>Hilliges</surname>
          </string-name>
          .
          <article-title>Deep pictorial gaze estimation</article-title>
          .
          <source>In European Conference on Computer Vision</source>
          (ECCV),
          <source>ECCV '18</source>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          [29]
          <string-name>
            <given-names>M.</given-names>
            <surname>Quigley</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Gerkey</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Conley</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Faust</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Foote</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Leibs</surname>
          </string-name>
          , E. Berger,
          <string-name>
            <given-names>R.</given-names>
            <surname>Wheeler</surname>
          </string-name>
          ,
          <article-title>and</article-title>
          <string-name>
            <surname>A. Ng. ROS:</surname>
          </string-name>
          <article-title>An open-source Robot Operating System</article-title>
          .
          <source>ICRA Workshop on Open Source SW</source>
          ,
          <volume>3</volume>
          :
          <issue>1</issue>
          {
          <issue>6</issue>
          ,
          <year>2009</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>