<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Comparison of Machine Learning approaches for Stress Detection from Wearable Sensors Data</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Michela Quadrini</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Denise Falcone</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Gianluca Gerard</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>School of Science and Technology, University of Camerino</institution>
          ,
          <addr-line>Via Madonna delle Carceri, 9, Camerino, 62032</addr-line>
          ,
          <country country="IT">Italy</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Sorint.Tek</institution>
          ,
          <addr-line>17 Zanica Grassobbio, BG, 24050</addr-line>
          <country country="IT">Italy</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Stress is a prevalent and growing phenomenon in the modern world potentially leading to significant repercussions on both physical and mental health. The analysis of physiological signals, collected from wearable sensors, has emerged as a promising approach to predicting and managing stress. Methods based on machine learning techniques have been defined in the literature and achieved promising results by using handcrafted features extracted from the signal. However, there is no consensus on the list of features, while deep learning approaches that overcomes the problem require significant computational power and a large amount of data. In this paper, we present a comprehensive view of the most common representative machine learning algorithms applied to the stress detection domain by giving a reference point for both academia and industry professionals in this application field. This study considers fragments of signals without extracting any features and uses a public dataset, WESAD, that contains high-resolution physiological, including blood volume pulse, electrocardiogram and electromyogram. The data collected from 15 subjects during a lab study are heterogeneous and characterized by diferent frequencies and noises due to some devices. After preprocessing, we assess the performance of ten machine learning algorithms belonging to four models (tree, ensemble, linear and neighbours) on the WESAD by facing the problem as binary (stress/no-stress) and multiclass (baseline, stress, and amusement) classifications. Our results, evaluated in terms of classical metrics, show that Random Forest outperforms the others in binary and multi-class approaches.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Physiological Signals</kwd>
        <kwd>Binary and multi-class classification</kwd>
        <kwd>Wearable Sensor Data</kwd>
        <kwd>time series</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        tonomic Nervous System, allow us to detect and monitor
stress. Hovsepian et al. [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] pioneered the stress detection
Stress is a non-specific body reaction to any demand by using physiological signals. Both faced the problem
upon it. Its efects influence overall behaviour, well-being, as a binary classification problem, whereas Gjoreski et
and potential personal and professional successes [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. al. [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] aimed at distinguishing diferent levels of stress
Chronic stress may give rise to significant physical and (no stress versus low stress versus high stress). Such
mental health issues, such as cancer, cardiovascular dis- bioignals can be captured non-invasively by wearable
ease, depression, and diabetes. It is an increasingly preva- devices, such as smartphones and smartwatches,
comlent and pervasive phenomenon in the modern world: monly used among people. Such devices can monitor
more than 50% of all work-related ill health cases in some physiological parameters, such as Blood Volume
2020/21 are due to stress [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. Assessments based on psy- Pulse (BVP), Electrodermal Activity (EDA), temperature
chologically designed questions, such as the Perceived (TEMP), and heart rate (HR) etc. In the scenario of stress
Stress Scale (PSS) [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ], are frequently used to detect stress. detection, machine learning and deep learning
methodHowever, these methods may be time-consuming, psy- ologies achieve promising results by analyzing these data.
chologically invasive and lack reliability. Therefore, the These approaches include support vector machines,
randefinition of non-invasive approaches for rapid and accu- dom forest and k-nearest neighbours and use handcrafted
rate stress detection influences the quality and wellness features extracted from the pre-processed signal in order
of people’s lives: managing stress before it causes health to reduce the data noises [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. Moreover, no consensus
issues is fundamental. In the literature, it has been demon- on the list of features to extract from physiological data
strated that physiological signals, a response to the Au- has been reached [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]. To solve the problem, advanced
Ital-IA 2024: 4th National Conference on Artificial Intelligence, orga- deep learning approaches have been applied since they
nized by CINI, May 29-30, 2024, Naples, Italy have the ability to automatically comprehend patterns
* Corresponding author. and, thus extract features. Nevertheless, these require
† These authors contributed equally. significant computational power and a large amount of
$ michela.quadrini@unicam.it (M. Quadrini); data. The appropriate machine learning algorithm choice
denise.facone@studenti.unicam.it (D. Falcone); for a particular problem task is not trivial: no single
clasmic0h0e0l0a-.q00u0a3d-r0in53i@9-u0n29ic0a(mM.i.t Q(Gu.adGreinrai)r;d0)000-0003-0539-0290 sifier works best across all possible scenarios, as stated
(G. Gerard) by no free lunch theorem states [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]. To the best of our
© 2024 Copyright for this paper by its authors. Use permitted under Creative Commons License knowledge, no scientific work compares machine
learnAttribution 4.0 International (CC BY 4.0).
ing methods for stress detection on the same datasets
without feature extraction or dimensionality reduction.
      </p>
      <p>
        In this paper, we present a comprehensive view of
the most common representative machine learning
algorithms applied to the stress detection domain by giving Figure 1: The two protocol versions used to collect data
a reference point for both academia and industry
professionals in this application field. In the analysis, we
consider fragments of signals without extracting any fea- lar disorders. Furthermore, the females subjects were
tures due to the nature of the problem: stress determines not pregnant. The dataset includes blood volume pulse
nonspecific human responses and the feature selection (BVP), electrocardiogram (ECG), electrodermal activity
depends on the subject and do not can be generalized. (EDA), electromyogram (EMG), respiration (RESP), body
Such signal fragments contain samples of all the physio- temperature (TEMP), and three-axis acceleration (ACC).
logical parameters measured. After appropriate resam- ECG, EDA, EMG, RESP, TEMP and ACC were recorded
pling and noise reduction, these values are linearized by a chest-worn device (RespiBan) and sampled at 700
and constitute the input of the considered ML model by Hz, whereas a wrist-worn device (Empatica E4) recorded
following the neural network approach. This study uses BVP (sampled at 64 Hz), EDA (at 4 Hz), TEMP (at 4 Hz),
the WESAD [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] dataset that is public and stores 12 phys- and ACC (at 32 Hz). The dataset comprises 14 time series,
iological signals, such as blood volume pulse and electro- each spanning approximately 2 hours, total experimental
cardiogram, collected from 15 subjects during a lab study. duration. The experiments were conducted to capture
After preprocessing (consisting of resampling, outlier re- three distinct afective states: baseline, stress, and
amusemoval, and normalization), we determine a dataset of ment with durations of 20 minutes, 392 seconds and 7
samples that are signal fragments obtained using the slid- minutes, respectively. They also included two meditation
ing window approach. Over these entries, we evaluate periods. To capture the data during the experiment, a
the most common and popular methods widely in various particular protocol, depicted in Figure 1, has been used. It
application areas. We consider eight machine learning consists of two diferent versions, where amusement and
algorithms, i.e, Decision Tree (DT), Random Forest (RF), stressful conditions are interchanged between diferent
Adaboost (AB), Extratree (ExT), Passive Aggressive Clas- subjects to avoid the efects of order.
sifier (PA), Logistic Regression (LR), K-kneighbors (NKE)
and Nearest Centrod (NC). We face the binary (stress/no- 2.2. Preprocessing
stress) and multi-class (baseline, stress, and amusement)
problem classifications. The results, evaluated in terms
of classical metrics, show that RF outperforms the others
in binary and multi-class approach. We also compare the
results obtained with the ones in the literature [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ].
      </p>
      <p>The paper is organized as follows. Section 2 describes
the materials and the methods used in this study. The
pipeline of the approach used in the study with the main
results are described in Section 3. The paper ends with
some conclusion and future work, Section 4.</p>
      <sec id="sec-1-1">
        <title>The varied sampling frequencies in WESAD, as detailed</title>
        <p>in Section 2.1, necessitated a harmonization step. We
resampled all data to match the 700Hz frequency of the
RespiBAN. Therefore, the resampling is applied only to
the time series detected by Empatica E4 using Fourier
method as an unsampling technique.</p>
        <p>
          After the resampling, we remove the outliers due to
occasional anomalous peaks in some signals, which may be
attributed to instrumental errors or measurement noise.
We removed the anomalies from each time series by
using a Hampel filter, discussed in [
          <xref ref-type="bibr" rid="ref10">10</xref>
          ]. Such a filter uses
1-minute sliding windows as input and calculates the
mean ( ) and standard deviation ( ) of the values within
the corresponding interval. Observations higher than the
threshold of 3 from the mean within the respective
window are classified as outliers (following Pearson’s rule)
and are substituted with the nearest chronological value.
This strategy ensures that outlier substitution doesn’t
introduce significant high-frequency variations.
        </p>
        <p>
          After outliers removal, we normalize all signals in
the interval [
          <xref ref-type="bibr" rid="ref1">− 1, 1</xref>
          ] to treat all inputs equally.Let  =
{1, 2, . . . , } be the considered time series with 
components, where each component corresponds to a
biophysical signal. Each of them are rescaled to the
in
        </p>
      </sec>
    </sec>
    <sec id="sec-2">
      <title>2. MATERIALS AND METHODS</title>
      <sec id="sec-2-1">
        <title>This work proposes a comparative evaluation of ML approaches to understand the best approach for real-time analytics. For this study, we consider the WESAD dataset.</title>
        <sec id="sec-2-1-1">
          <title>2.1. Dataset</title>
          <p>
            WESAD is a public dataset designed for stress and
afective detection. It is a high-quality multimodal dataset
storing physiological and movement data of 15 subjects
(12 male and 3 female) during a controlled lab
experiment [
            <xref ref-type="bibr" rid="ref9">9</xref>
            ]. All the participants were not heavy smokers
and did not sufer from chronic mental or
cardiovascurepresents a classification or decision. The root of the tree
corresponds to the best predictor. Usually, a DT is pruned
by combining the adjacent nodes to avoid overfitting.
2.4.2. Ensemble models
terval [
            <xref ref-type="bibr" rid="ref1">− 1, 1</xref>
            ] by applying the mean normalization: Random Forest Random Forest is an ensemble model
by Breiman [
            <xref ref-type="bibr" rid="ref14">14</xref>
            ] for both classification and regression.
          </p>
          <p>
            ( − ()) + (( − ()) It constructs a set of decision trees during training and
˜ = () − () determines the prediction by selecting the most
common class in the classification problem or calculating the
where () and () is the maximum and mean/average prediction in the regression problem of
minimum value among each component of , respec- the classes output by individual trees. This model
comtively. Therefore, the input is a the scaled time series, bines the bagging approach with the random selection
˜ = {˜1, ˜2, . . . , ˜}. of features to ensure the uncorrelation among the
decision trees of the forest. Feature randomness generates a
2.3. Dataset Entry random subset of features by ensuring low correlation
among decision trees. In bagging, the decision trees
deAfter the data preprocessing phase, we create two pend on trees created from a diferent bootstrap sample,
datasets: one for binary classification and the other for i.e., samples that may appear more than once in the
enmulticlass. All entries are obtained by applying the slid- tries of the training dataset. Diferently from decision
ing window technique to preprocessed signals. Specifi- trees that consider all the possible feature splits, random
cally, the entries consist of time series fragments charac- forests only select a subset of those features.
terized by only an emotional state (or label) obtained by
a slide of 60 seconds and a stride of 30 seconds, according
to the study in [
            <xref ref-type="bibr" rid="ref11">11</xref>
            ]. To create the multiclass dataset, we
consider parts of the time series associated with stress,
Baseline and Amusement, as described in Section 2.1. For
the binary classification, both the Baseline and
Amusement states were aggregated under a single ’non-stress’
label. The labels distribution of the two datasets are
shown in Fig. 2.
          </p>
          <p>
            AdaBoost AdaBoost, Adaptive Boosting, is an
ensemble models developed by Yoav Freund et al. [
            <xref ref-type="bibr" rid="ref15">15</xref>
            ]. It
employs an iterative approach to improve poor classifiers
by learning from their errors. Unlike the random forest
that uses parallel ensembling, Adaboost uses “sequential
ensembling”. Therefore, it is not possible to parallelize
jobs on a multiprocessor machine like Random Forest. It
creates a classifier by combining many poorly
performing classifiers to obtain a good classifier of high accuracy.
2.4. Machine Learning Algorithms Such resulting classifier is accomplished with
sequenIn this section, we describe some machine learning clas- tial weight adjustments, individual voting powers and a
sification techniques. Interested readers can refer to [
            <xref ref-type="bibr" rid="ref12">12</xref>
            ] weighted sum of the final algorithm classifiers.
for a complete treatment of machine learning approaches.
2.4.1. Decision Tree
A DT is a non-parametric supervised learning algorithm
for classification and regression in the form of a tree
structure [
            <xref ref-type="bibr" rid="ref13">13</xref>
            ]. It predicts the value of a target variable
by learning simple decision rules inferred from the data
features. The method exploits the “divide et impera”
approach to learning: it learns from data with a set of
if-then-else decision rules. The depth directly correlates
with the complexity of these decision rules. The output
is a tree comprising decision nodes and leaf nodes: a
decision node has two or more branches, and a leaf node
Extremely Randomized Trees Extremely
Randomized Trees, introduced in [
            <xref ref-type="bibr" rid="ref16">16</xref>
            ], are ensembling methods
that perform regression or classification. It creates a
large number of unpruned decision trees from the
training dataset and uses majority voting to select the decision
trees for the classification. Diferent from Random Forest,
it uses the entire dataset to train decision trees. Moreover,
it randomly selects the values at which to split a feature
and create child nodes to ensure suficient diferences
between individual decision trees.
2.4.3. Linear Models
Logistic Regression Logistic Regression, introduced
in [
            <xref ref-type="bibr" rid="ref17">17</xref>
            ], is a supervised learning algorithm mainly used
for classification tasks where the aim is to estimate the
probability of an instance belonging to a specific class
based on the values of the input features. The method
uses the sigmoid function to map any real-valued
number into a value between 0 and 1. More specifically, it
calculates a weighted sum of the input features, applies
the logistic function to this sum, and then classifies the
input as belonging to one of the two classes based on a
chosen threshold.
          </p>
          <p>
            Passive Aggressive The passive-aggressive algorithm,
introduced in [
            <xref ref-type="bibr" rid="ref18">18</xref>
            ], is one of the few "online learning
algorithms": the input data comes in sequential order,
and the model is updated step-by-step. It is useful in
applications that receive data as a continuous flow and
need to adapt to change rapidly or autonomously or if
you have limited computing resources. The algorithm
is based on based on Passive and Aggressive approches.
          </p>
          <p>If the prediction is correct, keep the model and do not
make any changes (passive), while If the prediction is
incorrect, make changes to the model.
2.4.4. Neighbors-based Models
Supervised neighbors-based models can be applied for
classification and regression. The principle behind
nearest neighbor methods is to find a predefined number of
training samples closest in distance to the new point, and
predict the label from these.
each class (target label). The training data is divided into
clusters based on their class labels, and then the centroid
is computed for each data cluster. Each centroid is simply
the mean value of each of the input variables. Such a
centroid represents the "model": given new examples, the
algorithm assigns the label by computing the distance
between a given data and each centroid.</p>
        </sec>
        <sec id="sec-2-1-2">
          <title>2.5. Metrics</title>
          <p>We evaluate the performance and efectiveness of the
approaches by using Accuracy (), Precision ( ), Recall
(), and F-measure (1), defined as follows
 =</p>
          <p>+  
  +   +   +  
 =
 =</p>
          <p>+</p>
          <p>+  
1 = 2</p>
          <p>· 
·  + 
where   represents the number of true positive,  
denotes the number of false negative,   represents the
number of false positive,   denotes the number of true
negative.</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. RESULTS</title>
      <sec id="sec-3-1">
        <title>The work aims to compare various machine learning algo</title>
        <p>
          rithms to detect stress from signals captured by wearable
devices. The workflow is described in Section 3.1, while
K-Nearest Neighbors The k-nearest neighbours al- the results of the experiments are described in Section 3.2.
gorithm, introduced by Fix and Hodges in 1951 [
          <xref ref-type="bibr" rid="ref19">19</xref>
          ] and
expanded by [
          <xref ref-type="bibr" rid="ref20">20</xref>
          ], is a non-parametric supervised
learning method for classification and regression. K-nearest 3.1. Methodology
neighbours algorithm exploits proximity to make classifi- Our pipeline, depicted in Fig. 3, is implemented in Python
cations or predictions about the grouping of an individual using the scikit-learn package for the machine learning
data point. KNN searches for the k-nearest labelled train- approaches and SciPy for data manipulation and
analying data by using the distance metric and attributes the sis. In particular, some methods of the SciPy library is
label which appears the most to the new observation. In used in the data preprocessing phase. The method
resamour study, we use the Minkowski distance as a metric. ple permits the resampling of signals. In our approach,
The input consists of the k closest training examples in a all signals are resampled at 700 Hz. About the outlier
data set, whereas the output depends on the task, classi- remotion, the Hampel filter is implemented using the
ifcation or regression. Such output is a class membership ‘rolling‘, ‘mean‘, ‘std‘, ‘fillna‘, ‘mask‘, and ‘interpolate‘
or the property value for the entry, respectively. methods from the Pandas library. The ‘MinMaxScaler‘
class of the scikit-learn package is used to perform data
Nearest Centroid Nearest Centroids, defined in [
          <xref ref-type="bibr" rid="ref21">21</xref>
          ], normalization. The machine learning methods Decision
is arguably the simplest classifier. It operates on an intu- Tree, Random Forest , K-Nearest Neighbors and
Logisitive principle: it takes data samples as input and classifies tic Regression are implemented via the tree, ensemble,
them into the class of training examples whose centroid neighbors and linear model modules, respectively. The
(a geometric centre of a data distribution) is closest to it. method K-Folds is used to split the dataset into 
conThe algorithm assumes that the centroids are distinct for secutive folds without shufling and then each fold is
DT
RF
KNN
DT
RF
KNN
        </p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. CONCLUSIONS AND FUTURE</title>
    </sec>
    <sec id="sec-5">
      <title>WORK</title>
      <p>then used once as a validation while the  − 1
remaining folds form the training set. The code used in this
manuscript are available from the corresponding author
upon reasonable request.</p>
      <sec id="sec-5-1">
        <title>3.2. Experiments</title>
        <sec id="sec-5-1-1">
          <title>Given the small number of subjects involved in the ex</title>
          <p>periment, we consider the Leave-One-Subject-Out
CrossValidation (LOSOCV), i.e., an approach that utilizes each
subject as a “test” set and the remaining 14 as a “training”
set. The experiments have been performed considering
the decision tree, random forest, K-Nearest Neighbors
and logistic regress as machine learning methods. For all
experiments, we use the default parameters.</p>
          <p>We evaluate such experiments by considering
Accuracy, Precision, Recall and F1-Score as metrics. Tables
1 shows the average values with the standard deviation
of the considered metrics obtained for binary and
multiclass classification, respectively. Appendix A reports the
values for each experiment.</p>
        </sec>
        <sec id="sec-5-1-2">
          <title>In this work, we have compared various classical ma</title>
          <p>chine learning algorithms. We have used a public dataset,
WESAD, to perform our study. Analyzing the results, we
have noted the best results have been archived by the
random forest algorithm. This evidence is in line with the</p>
          <p>
            Binary Classification results proposed in the literature [
            <xref ref-type="bibr" rid="ref9">9</xref>
            ]. We have observed
RDFT 00..9826A09cc±±ur00a..11c05y30 00..9942P44r e±±ci00si..o1100n05 00..984658R±±ec00a..l02l9029 00..9848F021-±±Sc00o..01re7660 tohnaetsctlhaastsicficoantisoidnesrbsaisgendalofneatthueressig.nal values outcome
EAxBT 00..980496 ±± 00..110594 00..984833 ±± 00,,018493 00..991057 ±± 00..111298 00..982855 ±± 00..019122 In future work, we intend to conduct additional
experLR 0.822 ± 0.232 0.843 ± 0.208 0.925 ± 0.199 0.871 ± 0.186 iments to discern the most relevant physiological signals.
PA 0.823 ± 0.225 0.842 ± 0.200 0.934 ± 0.187 0.874 ± 0.173 It represents another fundamental aspect of detecting
KNNCN 00..982495 ±± 00..110903 00..995339 ±± 00..019178 00..984196 ±± 00..028531 00.,895415±± 00..200735 stress for real-time analysis using wearable sensors and
          </p>
          <p>Multiclass Classification smartphones. In this case, the aim is to store the
min</p>
          <p>
            Accuracy Precision Recall F1-Score imum information to be non-invasive and reduce the
RDFT 00..760279 ±± 00..127212 00..666538 ±± 00..115975 00..760279 ±± 00..127212 00..656949 ±± 00..127333 space while maintaining high model performance. We
LEKRxNTNe 000..,656278307±±± 000...221436196 000...766014352 ±±± 000...122504289 000...656278307 ±±± 000...221436295 000...556864835 ±±± 000...221446925 aplrsooacinhetesn,sdutcoh caosngsriadpehr caonndveomluptiloony
ndeetewpolrekasronrinrgecaupr-NC 0.680 ± 0.219 0.685 ± 0.242 0,680 ± 0.220 0.662 ± 0.228 rent neural networks, motivated by the results obtained
Table 1 in other scenarios [
            <xref ref-type="bibr" rid="ref22 ref23">22, 23</xref>
            ]. Moreover, we also intend to
Average value with metrics with their standard deviation re- study the role of the length of the sliding windows from
lated to the binary and multiclass classification a theoretical perspective by taking into account various
entropy-based methods that have produced evaluable
out
          </p>
          <p>
            The Random Forest model outpaces its counterparts in comes in the scenario of protein-protein interaction site
both binary and multiclass classification scenarios. For prediction [
            <xref ref-type="bibr" rid="ref24">24</xref>
            ]. Another crucial future investigation is
the RF model, the obtained accuracy stands at 92% (bi- to explore and define approaches to extract and describe
nary) and 70% (multiclass). Corresponding F1-scores are the correlation that sliding windows represent. Other
88.2% and 60% , respectively. While multiclass classifica- representations, like arc-annotated sequences, strings
tion ofers insights for emotion detection via wearables, and simplicial complexes, will be explored. We will
exthere remains room for improvement. Comparing re- plore other representations like arc-annotated sequences
sults from Schmidt et al.’s benchmark on the WESAD for the analysis and comparison of time utilizing tools
dataset [
            <xref ref-type="bibr" rid="ref9">9</xref>
            ], which utilized standardized machine learn- like [
            <xref ref-type="bibr" rid="ref25">25</xref>
            ] and strings or simplicial complexes, which
aling techniques and features, our study finds that the RF low applying techniques from formal methods to identify
patterns [
            <xref ref-type="bibr" rid="ref26">26</xref>
            ] or verify properties [
            <xref ref-type="bibr" rid="ref27">27</xref>
            ].
          </p>
        </sec>
        <sec id="sec-5-1-3">
          <title>Acknowledgements. This work has been funded by</title>
          <p>the European Union - NextGenerationEU under the
Italian Ministry of University and Research (MUR) National
Innovation Ecosystem grant ECS00000041 - VITALITY
CUP J13C22000430001</p>
        </sec>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <surname>B. S. McEwen</surname>
          </string-name>
          ,
          <article-title>Protective and damaging efects of stress mediators</article-title>
          ,
          <source>New England journal of medicine 338</source>
          (
          <year>1998</year>
          )
          <fpage>171</fpage>
          -
          <lpage>179</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          <article-title>[2] Health and Safety Executive, HSE on work-related stress</article-title>
          .
          <year>2021</year>
          , http://www.hse.gov.uk/statistics/ causdis/-fstress/index.htm, ????
          <source>Accessed on March 7</source>
          ,
          <year>2022</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>E.-H.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <article-title>Review of the psychometric evidence of the perceived stress scale</article-title>
          ,
          <source>Asian nursing research 6</source>
          (
          <year>2012</year>
          )
          <fpage>121</fpage>
          -
          <lpage>127</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>K.</given-names>
            <surname>Hovsepian</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Al'Absi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Ertin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Kamarck</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Nakajima</surname>
          </string-name>
          ,
          <string-name>
            <surname>S.</surname>
          </string-name>
          <article-title>Kumar, cstress: towards a gold standard for continuous stress assessment in the mobile environment</article-title>
          ,
          <source>in: Proceedings of the 2015 ACM international joint conference on pervasive and ubiquitous computing</source>
          ,
          <year>2015</year>
          , pp.
          <fpage>493</fpage>
          -
          <lpage>504</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>M.</given-names>
            <surname>Gjoreski</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Luštrek</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Gams</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Gjoreski</surname>
          </string-name>
          ,
          <article-title>Monitoring stress with a wrist device using context</article-title>
          ,
          <source>Journal of biomedical informatics 73</source>
          (
          <year>2017</year>
          )
          <fpage>159</fpage>
          -
          <lpage>170</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>L.</given-names>
            <surname>Shu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Xie</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Liao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <article-title>A review of emotion recognition using physiological signals</article-title>
          ,
          <source>Sensors</source>
          <volume>18</volume>
          (
          <year>2018</year>
          )
          <year>2074</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>R.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <article-title>Stress detection using deep neural networks</article-title>
          ,
          <source>BMC Medical Informatics and Decision Making</source>
          <volume>20</volume>
          (
          <year>2020</year>
          )
          <fpage>1</fpage>
          -
          <lpage>10</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>D. H.</given-names>
            <surname>Wolpert</surname>
          </string-name>
          ,
          <article-title>The lack of a priori distinctions between learning algorithms</article-title>
          ,
          <source>Neural computation 8</source>
          (
          <year>1996</year>
          )
          <fpage>1341</fpage>
          -
          <lpage>1390</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>P.</given-names>
            <surname>Schmidt</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Reiss</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Duerichen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Marberger</surname>
          </string-name>
          ,
          <string-name>
            <surname>K. Van Laerhoven</surname>
          </string-name>
          ,
          <article-title>Introducing wesad, a multimodal dataset for wearable stress and afect detection</article-title>
          ,
          <source>in: Proceedings of the 20th ACM international conference on multimodal interaction</source>
          ,
          <year>2018</year>
          , pp.
          <fpage>400</fpage>
          -
          <lpage>408</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>J.</given-names>
            <surname>Astola</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Kuosmanen</surname>
          </string-name>
          ,
          <article-title>Fundamentals of nonlinear digital filtering</article-title>
          , CRC press,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>M.</given-names>
            <surname>Quadrini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Daberdaku</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Blanda</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Capuccio</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Bellanova</surname>
          </string-name>
          , G. Gerard,
          <article-title>Stress detection from wearable sensor data using gramian angular fields and cnn</article-title>
          , in: International Conference on Discovery Science, Springer,
          <year>2022</year>
          , pp.
          <fpage>173</fpage>
          -
          <lpage>183</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>S.</given-names>
            <surname>Shalev-Shwartz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Ben-David</surname>
          </string-name>
          ,
          <article-title>Understanding machine learning:</article-title>
          <source>From theory to algorithms</source>
          , Cambridge university press,
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>J. R.</given-names>
            <surname>Quinlan</surname>
          </string-name>
          ,
          <article-title>Induction of decision trees</article-title>
          ,
          <source>Machine learning 1</source>
          (
          <year>1986</year>
          )
          <fpage>81</fpage>
          -
          <lpage>106</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>L.</given-names>
            <surname>Breiman</surname>
          </string-name>
          , Random forests,
          <source>Machine learning 45</source>
          (
          <year>2001</year>
          )
          <fpage>5</fpage>
          -
          <lpage>32</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Freund</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. E.</given-names>
            <surname>Schapire</surname>
          </string-name>
          , et al.,
          <article-title>Experiments with a new boosting algorithm</article-title>
          , in: icml, volume
          <volume>96</volume>
          ,
          <string-name>
            <surname>Citeseer</surname>
          </string-name>
          ,
          <year>1996</year>
          , pp.
          <fpage>148</fpage>
          -
          <lpage>156</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>P.</given-names>
            <surname>Geurts</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Ernst</surname>
          </string-name>
          , L. Wehenkel,
          <article-title>Extremely randomized trees</article-title>
          ,
          <source>Machine learning 63</source>
          (
          <year>2006</year>
          )
          <fpage>3</fpage>
          -
          <lpage>42</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>D. R.</given-names>
            <surname>Cox</surname>
          </string-name>
          ,
          <article-title>The regression analysis of binary sequences</article-title>
          ,
          <source>Journal of the Royal Statistical Society Series B: Statistical Methodology</source>
          <volume>20</volume>
          (
          <year>1958</year>
          )
          <fpage>215</fpage>
          -
          <lpage>232</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>K.</given-names>
            <surname>Crammer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Dekel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Keshet</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Shalev-Shwartz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Singer</surname>
          </string-name>
          ,
          <article-title>Online passive aggressive algorithms (</article-title>
          <year>2006</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>E.</given-names>
            <surname>Fix</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. L.</given-names>
            <surname>Hodges</surname>
          </string-name>
          ,
          <article-title>Discriminatory analysis. nonparametric discrimination: Consistency properties</article-title>
          , International Statistical Review/Revue Internationale de Statistique 57 (
          <year>1989</year>
          )
          <fpage>238</fpage>
          -
          <lpage>247</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>T.</given-names>
            <surname>Cover</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Hart</surname>
          </string-name>
          ,
          <article-title>Nearest neighbor pattern classification</article-title>
          ,
          <source>IEEE transactions on information theory 13</source>
          (
          <year>1967</year>
          )
          <fpage>21</fpage>
          -
          <lpage>27</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>R.</given-names>
            <surname>Tibshirani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Hastie</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Narasimhan</surname>
          </string-name>
          , G. Chu,
          <article-title>Diagnosis of multiple cancer types by shrunken centroids of gene expression</article-title>
          ,
          <source>Proceedings of the National Academy of Sciences</source>
          <volume>99</volume>
          (
          <year>2002</year>
          )
          <fpage>6567</fpage>
          -
          <lpage>6572</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>M.</given-names>
            <surname>Quadrini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Daberdaku</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Ferrari</surname>
          </string-name>
          ,
          <article-title>Hierarchical representation and graph convolutional networks for the prediction of protein-protein interaction sites</article-title>
          ,
          <source>in: Machine Learning, Optimization, and Data Science: 6th International Conference, LOD</source>
          <year>2020</year>
          , Siena, Italy,
          <source>July 19-23</source>
          ,
          <year>2020</year>
          ,
          <string-name>
            <given-names>Revised</given-names>
            <surname>Selected</surname>
          </string-name>
          <string-name>
            <surname>Papers</surname>
          </string-name>
          ,
          <source>Part II 6</source>
          , Springer,
          <year>2020</year>
          , pp.
          <fpage>409</fpage>
          -
          <lpage>420</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <given-names>M.</given-names>
            <surname>Quadrini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Daberdaku</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Ferrari</surname>
          </string-name>
          ,
          <article-title>Hierarchical representation for ppi sites prediction</article-title>
          ,
          <source>BMC bioinformatics 23</source>
          (
          <year>2022</year>
          )
          <fpage>96</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [24]
          <string-name>
            <given-names>M.</given-names>
            <surname>Quadrini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Cavallin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Daberdaku</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Ferrari</surname>
          </string-name>
          ,
          <article-title>Prosps: protein sites prediction based on sequence fragments</article-title>
          ,
          <source>in: International Conference on Machine Learning, Optimization, and Data Science</source>
          , Springer,
          <year>2021</year>
          , pp.
          <fpage>568</fpage>
          -
          <lpage>580</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          [25]
          <string-name>
            <given-names>M.</given-names>
            <surname>Quadrini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Tesei</surname>
          </string-name>
          , E. Merelli,
          <article-title>Aspralign: a tool for the alignment of rna secondary structures with arbitrary pseudoknots</article-title>
          ,
          <source>Bioinformatics</source>
          <volume>36</volume>
          (
          <year>2020</year>
          )
          <fpage>3578</fpage>
          -
          <lpage>3579</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          [26]
          <string-name>
            <given-names>M.</given-names>
            <surname>Quadrini</surname>
          </string-name>
          , E. Merelli,
          <string-name>
            <given-names>R.</given-names>
            <surname>Piergallini</surname>
          </string-name>
          ,
          <article-title>Loop grammars to identify rna structural patterns</article-title>
          ., in: Bioinformatics,
          <year>2019</year>
          , pp.
          <fpage>302</fpage>
          -
          <lpage>309</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          [27]
          <string-name>
            <given-names>M.</given-names>
            <surname>Loreti</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Quadrini</surname>
          </string-name>
          ,
          <article-title>A spatial logic for simplicial models</article-title>
          ,
          <source>Logical Methods in Computer Science</source>
          <volume>19</volume>
          (
          <year>2023</year>
          ).
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>