<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Driver Monitoring Systems in Automated Interactions: A Realtime, Thermographic-based Algorithm</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Saifeddine Aloui</string-name>
          <email>saifeddine.aloui@cea.fr</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Raphaël Morvillier</string-name>
          <email>raphael.morvillier@cea.fr</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Christophe Prat</string-name>
          <email>christophe.prat@cea.fr</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Jaka Sodnik</string-name>
          <email>jaka.sodnik@fe.uni-lj.si</email>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Carolina Diaz- Piedra</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Francesco Angioi</string-name>
          <email>frangioi@ugr.es</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Leandro L. Di Stasi</string-name>
          <email>distasi@ugr.es</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Mind, Brain, and Behavior Research Center - CIMCYC, University of Granada</institution>
          ,
          <addr-line>Campus de Cartuja s/n, Granada 18011</addr-line>
          ,
          <country country="ES">Spain</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Univ. Grenoble Alpes, CEA, Leti</institution>
          ,
          <addr-line>F-38000 Grenoble</addr-line>
          ,
          <country country="FR">France</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>University of Ljubljana. Faculty of Electrical Engineering</institution>
          ,
          <addr-line>Tržaška c. 25, 1000 Ljubljana</addr-line>
          ,
          <country country="SI">Slovenia</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Due to the progressive shift of responsibility from the driver to the vehicle itself in automated vehicle technologies, driver-centered innovations represent a key point for its advance. The socalled Driver Monitoring Systems (DMS) are therefore increasingly gaining importance in this context. One of the main aims of DMS is to estimate the driver's arousal levels in order to infer their cognitive state and capabilities. Even though the scientific literature is riddled with useful psychophysiological indices to estimate arousal levels [1], nowadays, arousal estimation is based on broad, mostly blink/gaze-related, indices. The reason is that actual implementation of reliable sensors in a feasible system able to collect, analyze, and interpret measurements in real-life conditions is still an open challenge. One of the alternatives to signal different cognitive states is facial skin temperature [2][3]. Infrared sensors that monitor heat loss have been shown useful to track facial skin temperature that indicate arousal modulations while driving [2][3]. Such intensive, laborious work to extract and analyze temperature changes in some facial landmarks is not reasonable in real-life applications [2]. Here, we present the preliminary results obtained with a new software able to track, in real-time, drivers' facial-skin temperature changes.</p>
      </abstract>
      <kwd-group>
        <kwd>1 Driver state</kwd>
        <kwd>Workload</kwd>
        <kwd>Facial thermography</kwd>
        <kwd>Real-time algorithm</kwd>
        <kwd>Automated vehicle</kwd>
        <kwd>Sensoring and real-time information</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Due to the progressive shift of responsibility from the driver to the vehicle itself in automated vehicle
technologies, driver-centered innovations represent a key point for its advance. The so-called Driver
Monitoring Systems (DMS) are therefore increasingly gaining importance in this context. One of the
main aims of DMS is to estimate the driver’s arousal levels in order to infer their cognitive state and
capabilities. Even though the scientific literature is riddled with useful psychophysiological indices to
estimate arousal levels [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], nowadays, arousal estimation is based on broad, mostly blink/gaze-related,
indices. The reason is that actual implementation of reliable sensors in a feasible system able to collect,
analyze, and interpret measurements in real-life conditions is still an open challenge. One of the
alternatives to signal different cognitive states is facial skin temperature [
        <xref ref-type="bibr" rid="ref2 ref3">2,3</xref>
        ]. Infrared sensors that
monitor heat loss have been shown useful to track facial skin temperature that indicate arousal
modulations while driving. Such intensive, laborious work to extract and analyze temperature changes
in some facial landmarks is not reasonable in real-life applications [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. Face landmarks extraction using
color images has become of common use [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] thanks to several libraries (e.g., Google's MediaPipe
library [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]). However, when applied to thermographic images, these libraries produce unsatisfactory
results: the face is either not detected or the landmarks are not correctly aligned with the real face.
Therefore, two main methods have been developed to perform landmark detection on thermographic
images. The first method is to develop a dedicated system trained on annotated thermographic images
[see 6]. This approach is still limited due to the lack of large thermographic databases. For example,
Kopaczka and colleagues used a database containing 2,935 images [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]. A database of this kind would
not be useful for our data. In the present study, the drivers had to wear transparent face masks due to
the COVID19 pandemic. This made it harder to apply landmark detection on thermographic images
where the mask was visible. Indeed, although the masks were transparent to visible light, they were not
in the wavelength used to measure the temperature, therefore hiding part of the driver’s face. The second
method uses an additional color camera to detect the facial landmarks and transfers them on the
thermographic image (this process of aligning images from different sources is often referred to as
“image registration”). In previous studies, authors detected the edges in both color and thermal images
and match them to align the images [
        <xref ref-type="bibr" rid="ref8 ref9">8, 9</xref>
        ]. A simpler method is described in another work [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ], based
on an initial optical calibration between the two cameras. Goulart and colleagues used the same
principle and add a post-processing step to enhance the transferred landmark position, based on a trained
expert manual annotation [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ]. Here, we present the preliminary results obtained with a system based
on this second method, able to track drivers’ facial-skin temperature changes automatically after an
initial calibration. It is a first step towards a fully automatic system, which could run in real-time in
future vehicles. We present the principle of the system and analyze its performance. In a future work,
we intend to show the usefulness of extracting the face temperature in an automated driving condition.
      </p>
    </sec>
    <sec id="sec-2">
      <title>2. Material and methods</title>
    </sec>
    <sec id="sec-3">
      <title>2.1. Instruments</title>
      <p>We used a sensorized driving simulator (Nervtech™ solution, see Figure 1) running a SCANeR
studio software (AVSimulation, v.DT2.5). Participants’ facial skin temperature was constantly
monitored with a thermographic camera (FLIR A325sc, with a resolution of 320 × 240, a NETD &lt;
50mK and an accuracy of ±2°C or ±2% of reading) synchronized with a color camera (infrared color
camera, Intel® Realsense).
2.2.</p>
    </sec>
    <sec id="sec-4">
      <title>Face temperature extraction algorithm</title>
      <p>To extract participants’ facial-skin temperature at specific locations, we developed an algorithm
(Figure 2) able to identify two facial landmarks (Points of Interest [POIs]), the tip of the nose and the
forehead, as well as the background in a thermographic image. The solution was based on a dual camera
setup (i.e., color camera and thermographic camera), with a spatial correspondence between the two.</p>
      <p>ℎ</p>
      <p>=  
→ ℎ
= [0
1
0
( )
( )
0


0
= [ 0</p>
      <p>→ ℎ
=  
0 
1   ]
0


1
−

0

0


1
0
1
2.3.</p>
    </sec>
    <sec id="sec-5">
      <title>Calibration</title>
      <p>(1)
(2)
(3)
(4)
(5)
multiplication of three matrices (2). The first describes a translation with coordinates [  ,  ] (3), the
second describes a rotation around the center of the screen with angle  (4) and the third describes a
scaling with parameters [  ,  ] (5). Once the positions of the landmarks in the thermographic image
space were found, the POIs temperature values were read in the image. Finally, we multiplied the result
by the skin emissivity (0.98) to obtain the skin temperature.</p>
      <p>The described system first needed to be calibrated to determine the parameters of the transformation
matrix  
→ ℎ</p>
      <p>
        :   ,   ,  ,   and   . Filippini and colleagues used a similar set-up and performed
the calibration using a custom checkerboard [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ], a method we found to be less precise in our situation.
We therefore developed a dedicated calibration software. It allows an operator to visualize
simultaneously the color and the thermographic camera outputs, as shown in Figure 3.
thanks to MediaPipe [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. On the right, the thermographic image with the corresponding landmarks
that the operator has to translate, rotate, and scale to correspond to the driver’s face.
On the color image, the operator can inspect the landmark detection performed by MediaPipe. A
thermographic image shows if these landmarks are transferred correctly. If the result is not satisfying,
additional translations, rotations and scaling of the landmarks “mask” can be done manually with the
mouse. These transformations are recorded by the calibration software to compute the matrix
  → ℎ . The calibration software finally saves the conversion parameters in a dedicated file
which is used by the extraction software to automatically detect the POIs on the thermographic output.
In our experiment, we repeated the calibration procedure for each driver to compensate for slight
differences in the positions of the cameras and head among different drivers.
2.4.
      </p>
    </sec>
    <sec id="sec-6">
      <title>Experimental design</title>
      <p>To test our algorithm, we designed a 2 (traffic vs. low-traffic Traffic density) × 2 (automated
[ADL4] vs. manual [MD] Driving modality) within-participants experiment. Thirty-five expert drivers
(mean age = 41.61 years, standard deviation = 6.26 years) drove along two virtual scenarios (∼ 20
minutes [min] each) with varying traffic density. In both scenarios, the participants performed 10 min
in MD and 10 min in ADL4. The order of the traffic density and driving mode was randomly balanced
across drivers. During ADL4, they were instructed to supervise the system. We expect the arousal level
of the drivers to be modulated by these conditions, as the manual driving mode and the high traffic
condition are more demanding than the autonomous one and the low traffic condition respectively.</p>
      <p>In order to validate the proposed algorithm, we selected randomly one of the two 20-min recordings
(high or low traffic) for each driver. Then, we extracted one pair of color and thermographic images
each 20 seconds. We obtained 65 images per driver and 2,340 in total. We then developed an annotation
software to manually extract the temperature on these images. For each image, we pointed at two
landmarks: the driver’s forehead and the driver’s nose tip. Four trained annotators performed the same
procedure on the 2,340 images. Then, we computed on each image the mean of the four annotated
positions of the nose tip and the forehead to establish the reference location of the nose tip and the
forehead. Finally, we extracted the temperature at these locations to define the reference temperature.
2.6.</p>
    </sec>
    <sec id="sec-7">
      <title>Statistical study on the obtained data</title>
      <p>After removing the drivers on who the algorithm performed worse (see section 3.1.2), we used the
algorithm described in Section 2.2 to extract the face temperature of the remaining drivers (n = 28). In
order to obtain more measurement points, the algorithm at this stage ran at a higher frequency compared
to the validation phase: one each 2 sec instead of one each 20 sec. We were therefore able to remove
extreme values (lower than 25°C and higher than 37°C) as well as the outliers by applying a moving
median thresholding procedure. We finally took the mean of the remaining points on each of the four
segments: High traffic – Manual driving, High traffic – Automated driving, Low traffic – Manual
driving, Low traffic – Automated driving. This gave us four data points per driver that we later used in
our statistical analysis.</p>
    </sec>
    <sec id="sec-8">
      <title>3. Results</title>
    </sec>
    <sec id="sec-9">
      <title>3.1. Validation of the algorithm</title>
      <p>We first analyzed the algorithm performance in terms of position error in the thermographic image,
measured in pixels. We computed the position error between the algorithm’s output and the mean values
provided by the four annotators (see 2.5). We also compared the position error of the mean position
error of each annotator with respect to the overall mean value. Then, we analyzed the consequences of
the algorithm position error in terms of temperature error, measured in degrees Celsius (°C).</p>
    </sec>
    <sec id="sec-10">
      <title>3.1.1. Position error</title>
      <p>In Figure 6, we present the errors’ distributions of the algorithm and the annotators. As a reference,
in our setup the nose tip measures approximately 10 x 10 pixels. When pointing at the nose, the
algorithm performed worse than the annotators with respect to the mean of the annotators. The two
main causes for high mismatches were landmarks estimation errors of MediaPipe and spatial
correspondence errors due to head movements (head turning or bending). Surprisingly, the algorithm
outperformed slightly the annotators on the forehead with respect to the mean of the annotators. Our
interpretation is that for a human, it could be hard to define a precise location on a large area with no
points of reference such as the forehead.</p>
    </sec>
    <sec id="sec-11">
      <title>3.1.2. Temperature error</title>
      <p>Figure 7 shows the errors of the final temperature values computed by the algorithm. On the
forehead, the temperature gradient was low, so the temperature error resulting from the position error
was small. On the nose, however, the temperature gradient was higher, so the temperature error was
also much higher compared to the forehead. Interestingly, the temperatures computed at the positions
annotated by one annotator are consistently smaller than the temperatures computed at the mean of the
annotated positions. This is because the face temperature exhibits a local peak on the nose and one
individual annotator is further from this peak than the mean position of the four annotators.</p>
      <p>Looking at Figure 8, we see that the mean absolute error of the nose temperature highly depends on the
driver (it goes up to 1.6 °C for some drivers). For the statistical study, we excluded the 6 participants
with an absolute error higher than 0.8 °C.</p>
    </sec>
    <sec id="sec-12">
      <title>4. Conclusion and future works</title>
      <p>
        The present work describes the first results obtained with an algorithm for tracking a driver’s facial
skin temperature during driving interactions. The algorithm consistently and effectively tracked
participants’ facial-skin temperature without interfering with their driving tasks. We have analyzed the
position and temperature errors and for some drivers, tracking the nose tip temperature remains a
challenge. Future systems should improve both the initial landmarks detection and the landmark
transfer. The later could be achieved by measuring the distance between the cameras and the driver’s
face like previous studies [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] or considering the face as a 3D shape. Also, a calibration-less process
should be developed to be implemented in a real car. More analysis should be conducted before
publishing the results of a statistical study based on this work.
      </p>
    </sec>
    <sec id="sec-13">
      <title>5. Acknowledgements</title>
      <p>This study was funded by the European Union's Horizon 2020 research and innovation programme
under grant agreement No. 875597 - HADRIAN (Holistic Approach for Driver Role Integration and
Automation Allocation for European Mobility Needs) project. This document reflects only the authors'
view, the European Climate, Infrastructure and Environment Executive Agency (CINEA) is not
responsible for any use that may be made of the information it contains. We thank Leila Maboudi
(Polytechnic University of Turin, Italy) for her comments and assistance in language edition.</p>
    </sec>
    <sec id="sec-14">
      <title>6. References</title>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>L.L. Di</given-names>
            <surname>Stasi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Gianfranchi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Diaz-Piedra</surname>
          </string-name>
          ,
          <article-title>Hand-skin temperature response to driving fatigue: an exploratory study</article-title>
          , in: Krömker, H. (Eds), HCI in Mobility, Transport, and
          <string-name>
            <given-names>Automotive</given-names>
            <surname>Systems</surname>
          </string-name>
          . Driving Behavior, Urban and
          <string-name>
            <given-names>Smart</given-names>
            <surname>Mobility</surname>
          </string-name>
          ,
          <string-name>
            <surname>HCII</surname>
          </string-name>
          <year>2020</year>
          , vol
          <volume>12213</volume>
          of Lecture Notes in Computer Science, Springer, Cham,
          <year>2020</year>
          , pp.
          <fpage>3</fpage>
          -
          <lpage>14</lpage>
          . doi:
          <volume>10</volume>
          .1007/978-3-
          <fpage>030</fpage>
          -50537-
          <issue>0</issue>
          _
          <fpage>1</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>C.</given-names>
            <surname>Diaz-Piedra</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Gomez-Milan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.L. Di</given-names>
            <surname>Stasi</surname>
          </string-name>
          .
          <article-title>Nasal skin temperature reveals changes in arousal levels due to time on task: An experimental thermal infrared imaging study</article-title>
          ,
          <source>Applied Ergonomics</source>
          <volume>81</volume>
          (
          <year>2019</year>
          ). doi:
          <volume>10</volume>
          .1016/j.apergo.
          <year>2019</year>
          .
          <volume>06</volume>
          .001.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>Panasonic</given-names>
            <surname>Corporation</surname>
          </string-name>
          ,
          <article-title>Panasonic develops drowsiness-control technology by detecting and predicting driver's level of drowsiness, 2017</article-title>
          . URL: https://news.panasonic.com/global/press/data/2017/07/en170727-3/en170727-
          <fpage>3</fpage>
          .html
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>M.</given-names>
            <surname>Bodini</surname>
          </string-name>
          .
          <article-title>A review of facial landmark extraction in 2D images and videos using deep learning</article-title>
          ,
          <source>big data and cognitive computing 3</source>
          ,
          <issue>14</issue>
          (
          <year>2019</year>
          ). doi:
          <volume>10</volume>
          .3390/bdcc3010014.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>C.</given-names>
            <surname>Lugaresi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Tang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Nash</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>McClanahan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Uboweja</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Hays</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <surname>C.-L. Chang</surname>
            ,
            <given-names>M.G.</given-names>
          </string-name>
          <string-name>
            <surname>Yong</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Lee</surname>
            ,
            <given-names>W.-T.</given-names>
          </string-name>
          <string-name>
            <surname>Chang</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          <string-name>
            <surname>Hua</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Georg</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Grundmann</surname>
          </string-name>
          ,
          <year>2019</year>
          .
          <article-title>MediaPipe: A framework for building perception pipelines</article-title>
          . arXiv:
          <year>1906</year>
          .08172.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>W.-T.</given-names>
            <surname>Chu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.-H.</given-names>
            <surname>Liu</surname>
          </string-name>
          .
          <article-title>Thermal facial landmark detection by deep multi-task learning</article-title>
          ,
          <source>in 2019 IEEE 21st International Workshop on Multimedia Signal Processing (MMSP)</source>
          ,
          <year>2019</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>6</lpage>
          . doi:
          <volume>10</volume>
          .1109/MMSP.
          <year>2019</year>
          .
          <volume>8901710</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>M.</given-names>
            <surname>Kopaczka</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Kolk</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Merhof</surname>
          </string-name>
          .
          <article-title>A fully annotated thermal face database and its application for thermal facial expression recognition</article-title>
          ,
          <source>in 2018 IEEE International Instrumentation and Measurement Technology Conference (I2MTC)</source>
          ,
          <year>2018</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>6</lpage>
          . doi:
          <volume>10</volume>
          .1109/I2MTC.
          <year>2018</year>
          .
          <volume>8409768</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>H.</given-names>
            <surname>Yoshikawa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Uchiyama</surname>
          </string-name>
          , T. Higashino. ThermalWrist:
          <article-title>Smartphone thermal camera correction using a wristband sensor †</article-title>
          ,
          <source>Sensors</source>
          <volume>19</volume>
          (
          <year>2019</year>
          ). doi:
          <volume>10</volume>
          .3390/s19183826.
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>L.</given-names>
            <surname>Sun</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Zheng</surname>
          </string-name>
          .
          <article-title>Thermal-to-visible face alignment on edge map</article-title>
          ,
          <source>IEEE Access 5</source>
          (
          <year>2017</year>
          )
          <fpage>11215</fpage>
          -
          <lpage>11227</lpage>
          . doi:
          <volume>10</volume>
          .1109/ACCESS.
          <year>2017</year>
          .
          <volume>2712159</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>C.</given-names>
            <surname>Filippini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Spadolini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Cardone</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Bianchi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Preziuso</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Sciaretta</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Del Cimmuto</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Lisciani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Merla</surname>
          </string-name>
          .
          <article-title>Facilitating the child-robot interaction by endowing the robot with the capability of understanding the child engagement: The case of mio amico robot</article-title>
          ,
          <source>International Journal of Social Robotics</source>
          <volume>13</volume>
          (
          <year>2019</year>
          )
          <fpage>677</fpage>
          -
          <lpage>689</lpage>
          . doi:
          <volume>10</volume>
          .1007/s12369-020-00661-w.
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>C.</given-names>
            <surname>Goulart</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Valadão</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Delisle-Rodriguez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Funayama</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Favarato</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Baldo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Binotte</surname>
          </string-name>
          , E. Caldeira,
          <string-name>
            <surname>T.</surname>
          </string-name>
          Bastos-Filho.
          <article-title>Visual and thermal image processing for facial specific landmark detection to infer emotions in a child-robot interaction</article-title>
          ,
          <source>Sensors</source>
          (Basel)
          <volume>19</volume>
          (
          <year>2019</year>
          ). doi:
          <volume>10</volume>
          .3390/s19132844.
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>R.</given-names>
            <surname>Artzy</surname>
          </string-name>
          , Linear Geometry, Dover Publications, New York, NY,
          <year>1993</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>