<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Micro-environment Recognition in the context of Environmental Crowdsensing</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Mohammad Abboud</string-name>
          <email>mohammad.abboud.2496@gmail</email>
          <email>mohammad.abboud.2496@gmail. com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Hafsa El Hafyani</string-name>
          <email>hafsa.el-hafyani@uvsq.fr</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Jingwei Zuo</string-name>
          <email>jingwei.zuo@uvsq.fr</email>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Karine Zeitouni</string-name>
          <email>karine.zeitouni@uvsq.fr</email>
          <xref ref-type="aff" rid="aff3">3</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Yehia Taher</string-name>
          <email>yehia.taher@uvsq.fr</email>
          <xref ref-type="aff" rid="aff4">4</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Activity Recognition, Multivariate Time Series Classification,</string-name>
          <xref ref-type="aff" rid="aff5">5</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>DAVID Lab, UVSQ - Université Paris-Saclay</institution>
          ,
          <addr-line>Versailles</addr-line>
          ,
          <country country="FR">France</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>DAVID Lab, UVSQ - Université Paris-Saclay</institution>
          ,
          <addr-line>Versailles</addr-line>
          ,
          <country country="FR">France</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>DAVID Lab, UVSQ - Université Paris-Saclay</institution>
          ,
          <addr-line>Versailles</addr-line>
          ,
          <country country="FR">France</country>
        </aff>
        <aff id="aff3">
          <label>3</label>
          <institution>DAVID Lab, UVSQ - Université Paris-Saclay</institution>
          ,
          <addr-line>Versailles</addr-line>
          ,
          <country country="FR">France</country>
        </aff>
        <aff id="aff4">
          <label>4</label>
          <institution>DAVID Lab, UVSQ - Université Paris-Saclay</institution>
          ,
          <addr-line>Versailles</addr-line>
          ,
          <country country="FR">France</country>
        </aff>
        <aff id="aff5">
          <label>5</label>
          <institution>Multi-view Learning, Mobile Crowd Sensing</institution>
          ,
          <addr-line>Air Quality Monitoring</addr-line>
        </aff>
      </contrib-group>
      <abstract>
        <p>With the rapid advancements of sensor technologies and mobile computing, Mobile Crowd-Sensing (MCS) has emerged as a new paradigm to collect massive-scale rich trajectory data. Nomadic sensors empower people and objects with the capability of reporting and sharing observations on their state, their behavior and/or their surrounding environments. Processing and mining multi-source sensor data in MCS raise several challenges due to their multi-dimensional nature where the measured parameters (i.e., dimensions) may difer in terms of quality, variabilty, and time scale. We consider the context of air quality MCS, and focus on the task of mining the context from the MCS data. Relating the measures to their context is crucial to interpret them and analyse the participant's exposure. This paper investigates the feasibility of recognizing the human's context (called herein micro-environment) in an environmental MCS scenario. We put forward a multi-view learning approach, that we adapt to our context, and implement it along with other time series classification approaches. The experimental results, applied to real MCS data, not only confirm the power of MCS data in characterizing the micro-environment, but also show a moderate impact of the integration of mobility data in this recognition. Furthermore, multi-view learning shows similar performance as the reference deep learning algorithm, without requiring specific hardware.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>INTRODUCTION</title>
      <p>
        Nowadays, the Internet of Things (IoT) basically relies on
advanced sensor technologies to bridge the physical world and
information systems. In particular, along with the widespread
use of GPS, various mobile sensors bring rich information
collected from both the surrounding environment and human
activities, which are generally represented as Geo-referenced Time
Series (GTS). Mobile Crowed Sensing (MCS) [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ] emerges as a
new paradigm, which empowers volunteers to contribute data
(i.e., GTS) acquired by their personal sensor-enhanced mobile
devices. Polluscope1, a French project deployed in Île-de-France
(i.e., Paris region), is a typical use case study on MCS. It aims at
getting insights constantly on individual exposure to pollution
everywhere (indoor and outdoor), while enriching the traditional
monitoring system with the collected data by the crowd. The
recruited participants, on a voluntary basis, collect air quality
measurements. Each participant is equipped with a sensor kit, and
a mobile device which allows the transmission of collected
measurements together with the GPS coordinates as a geo-referenced
data stream containing (timestamp, longitude, and latitude). In
addition, the participants are asked to annotate their
environment type through a custom mobile application. This will allow
participants to have personalized insights about their exposure
to pollution everywhere, either in indoor and outdoor
environments (e.g., Home, Work, Transportation, Streets, Park, etc.), and
at a higher resolution along their trajectories, thereby, allowing
to capture local variability and peaks of pollution, depending on
participants’ whereabouts, i.e., micro-environments.
      </p>
      <p>It is worth mentioning that air quality strongly depends on
the context2, and so is the individual exposure to pollution. For
this reason, there is a great interest of making exposure analysis
context-aware. However, the context annotation is by far the most
dificult information to collect in a real-life application setting,
since a very few participants thoroughly annotate their
microenvironment. Therefore, there is a great interest in unburdening
the participants by automatically detecting the context.</p>
      <p>When exploring visually the data, we noticed that
microenvironments preserve a certain pattern. Besides, we observe
the existence of an inter-sensor correlation and with the
context. Figure 1 shows the evolution of three dimensions (i.e. Black
Carbon (BC), NO2 and Particulate Matters (PM)) with
microenvironments identification. As shown in Figure 1, BC and NO2
preserve the same shapes and statistical characteristics in the
micro-environment “Car”. Plus, we note that PM values keep
the same statistical characteristics in the micro-environment
“Indoor”. Moreover, we can observe the existence of a correlation
between the three dimensions during the whole timeline.</p>
      <p>The idea we promote in this paper is to utilize a wisely chosen
annotated dataset, in order to train a model on all the combination
1http://polluscope.uvsq.fr
2In this paper, the terms "context" and "micro-environment" are used
interchangeably.
of air quality, and mobility dimensions as predictors of the
microenvironment. We hypothesize that the multi-variate time series
collected by the MCS campaigns not only depends on the
microenvironment but could be a proxy of it.</p>
      <p>
        The question that arises itself now is how to combine all these
diferent aspects of the data (geo-location, sensors) to identify
the user’s context automatically? And how much a model can
discriminate the observations in diferent micro-environments? To
this end, we envision a holistic approach of activity recognition,
as depicted in [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ].
      </p>
      <p>The micro-environment recognition is a crucial figure to
exposure interpretation. Once the data are correctly annotated, our
ultimate goal is to get insight into all the dimensions (spatial,
temporal, individual, and contextual) of the exposure to pollution.</p>
      <p>In this paper, we evaluate diferent approaches and provide a
framework dedicated to the preparation, the application and the
comparison of diferent machine learning algorithms.</p>
      <p>The rest of this paper is organized as follows. We introduce
the related work in Section 2. The formal presentation of our
micro-environment recognition model is discussed in Section 3.
Section 4 presents the experimental results and evaluation of the
micro-environment recognition model in the context of
environmental crowd sensing. Section 5 gives the extensive discussion
for the perspectives of this work. In section 6, we summarize our
conclusions and provide directions for future work.
2</p>
    </sec>
    <sec id="sec-2">
      <title>RELATED WORK</title>
      <p>
        Human activity recognition involves a wide range of applications
from smart homes activities [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] to daily human activities [
        <xref ref-type="bibr" rid="ref17 ref28 ref4">4,
17, 28</xref>
        ], to human mobility [
        <xref ref-type="bibr" rid="ref30 ref6">6, 30</xref>
        ] to cite a few. It represents a
typical scenario of machine learning, and some public datasets
are widely used in the benchmarks. In this section, we introduce
a summary of two main topics of related work to our approach.
We focus mainly on multi-variate time series (MTS) classification
and multi-view learning.
2.1
      </p>
    </sec>
    <sec id="sec-3">
      <title>Multi-Variate Time Series Classification</title>
      <p>
        Human activity recognition falls in the problem of labelling data
segments with the type of activity, which leads to a multi-variate
time series classification (MTSC) problem based on data collected
by multiple wearable sensors. There is a wide range of time series
classification approaches that can be classified into four
categories: distance-based methods [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ], feature-based methods [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ],
ensemble methods [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] and deep learning models [
        <xref ref-type="bibr" rid="ref24 ref3 ref9">3, 9, 24</xref>
        ]. The
One-Nearest Neighbor (1-NN) classifier with diferent distance
measures, such as Euclidean Distance (ED) or Dynamic Time
Wrapping (DTW) [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ], is always considered as the benchmark to
give a preliminary evaluation in the MTSC problem.
      </p>
      <p>
        Considering the real-life scenarios, where it is dificult or
expensive to obtain a large amount of labeled data for training,
some studies used both labeled and unlabeled data to learn the
human activity, that is Semi-Supervised Learning (SSL) [
        <xref ref-type="bibr" rid="ref25">25</xref>
        ] on
MTSC. The pioneering work by [
        <xref ref-type="bibr" rid="ref25">25</xref>
        ] propose a semi-supervised
technique for time series classification. The authors demonstrated
that semi-supervised learning requires less human efort and
generally achieves higher accuracy than training on limited labels.
The semi-supervised model [
        <xref ref-type="bibr" rid="ref25">25</xref>
        ] is based on the Self-Learning
concept with the One-Nearest-Neighbor (1-NN) classifier. First,
the labeled set, denoted by  (as positively labeled) is applied to
train the 1-NN classifier . Then, the unlabeled samples  are
given the pseudo labels progressively based on their distance to
the samples in  . Thereafter, the enriched labeled set  allows
iteratively repeating the previous step and improving the classifier.
More recently, the deep learning-based models on MTSC show
promising performance under weak supervision. For instance,
Zhang et al. [
        <xref ref-type="bibr" rid="ref29">29</xref>
        ] propose a novel semi-supervised MTSC model
named Time series attentional prototype network (TapNet), to
explore the valuable information in the unlabeled samples. TapNet
projects the raw MTS data into a low-dimensional representation
space. The unlabeled samples approach themselves to the class
prototype in the representation space, where the distance-based
probability and the labeled samples allow training the model
progressively. Moreover, the hybrid Convolutional Neural Network
(CNN) and Long Short-Term Memory (LSTM) structure adopted
in TapNet allows modeling, respectively, the variable interactions
and the temporal features of MTS.
2.2
      </p>
    </sec>
    <sec id="sec-4">
      <title>Multi-View Learning</title>
      <p>
        Another line of studies propose multi-view learning to classify
time series data originated from multiple sensors to recognize
users activities. Garcia-Ceja et al. [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] propose a method based on
multi-view learning and stacked generalization for fusing audio
and accelerometer sensor data for human activity recognition
using wearable devices. Each sensor’s data is seen as a diferent
“view”, and they are combined using stacked generalization [
        <xref ref-type="bibr" rid="ref26">26</xref>
        ].
The approach trains a specific classification model over each view
and an extra meta-learner using the view models as input. The
general idea of the authors is to combine data from heterogeneous
types of sensors to complement each other and thus, increase
recognition accuracy.
      </p>
      <p>
        Wang et al. [
        <xref ref-type="bibr" rid="ref23">23</xref>
        ] propose a framework based on deep learning
to learn features from diferent aspects of the data based on
features of sequence and visualization. In order to imitate the
human brain, which can classify data based on visualization,
the authors transform the time series into an Area Graph. They
use well-trained LSTM-A neural networks and CNN-A neural
networks to extract the features of time series data. LSTM-A is
used to extract sequence features, while CNN-A is used to extract
visual features from the time series. Then, based on the fusion
of features, the authors carry out the time series classification
task. Although the approach gained promising results, it did not
outperform deep learning methods such as InceptionTime [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ].
      </p>
      <p>
        Li et al. [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ] propose a Multi-view Discriminative Bilinear
Projections (MDBP) for multi-view MTSC. The proposed approach
is a multi-view dimensionality reduction method for time series
classification which aims to extract discriminative features from
multi-view MTS data. MDBP mainly projects multi-view data to
a shared subspace through view-specific bilinear projections that
preserve the temporal structure of MTS, and learns discriminative
features by incorporating a novel supervised regularization.
3
      </p>
    </sec>
    <sec id="sec-5">
      <title>MICRO-ENVIRONMENT RECOGNITION</title>
    </sec>
    <sec id="sec-6">
      <title>MODEL</title>
      <p>In this section, we provide an overview of our proposed
framework for micro-environment recognition in the context of MCS.
Our proposed approach contains six steps as shown in Figure 2.
3.1</p>
    </sec>
    <sec id="sec-7">
      <title>Data Collection</title>
      <p>The first step of our micro-environment recognition process is
the data collection. During three campaigns, around one hundred
participants have been recruited to collect ambient air
measurements along with geo-location for one week 24 hours a day,
while performing their daily activities. Each participant carries
a multi sensor box and a tablet empowered with GPS chipset.
The sensors collect time annotated measurements of Particulate
Matter (PM1.0, PM10, PM2.5), nitrogen dioxide ( 2), Black
Carbon (BC), Temperature and Relative Humidity, and the tablet
records participants’ geo-locations and allows them to annotate
their context by using a self-reporting mobile app. They report
every transition to a micro-environment (e.g., Home, Ofice, Park,
Restaurant, etc.), as well as events, which are temporary activities
for a brief period (e.g., Start cooking, Open a window, Close a
window, Smoking, Turn on a chimney, etc.).
3.2</p>
    </sec>
    <sec id="sec-8">
      <title>Data Preparation</title>
      <p>
        The second step consists of pre-processing the data which is two
folds. On the one hand, most sensor data are noisy, and require
a prepossessing phase to clean them from irrelevant
measurements. We have observed this especially in the GPS (due to signal
loss), and in air quality data even though the sensor data quality
is a permanent preoccupation of the project, by careful
evaluation before their selection, and periodic qualification during
the campaign [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ]. The sensors for climatic parameters do not
show such defects. Therefore, a de-noising process is applied to
clean the data. On the other hand, the highest quality sample of
annotated data is selected as a baseline to validate the process
of micro-environment recognition. The idea is to generalize the
micro-environment recognition to all participants’ data, by using
the model derived from a good-quality dataset.
3.3
      </p>
    </sec>
    <sec id="sec-9">
      <title>Multi-View Learning Model</title>
      <p>
        We were interested in the stack generalization approach
proposed in [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ], but we have adapted it to best fit for solving our
problem. We propose to learn the micro-environment of
participants from multi-variate time series through a two-stage model
based on multi-view learning. Our multi-view classification
approach consists of training a first-level learner on each view
(i.e. step 3 in Figure 2), and then train a second-level learner or
meta-learner (i.e. step 5 in Figure 2) to combine the output of
each view and enhance the global accuracy of the classification.
We assume that  is a dimension of the n-dimenstional time
series  = (1 , 2 , ...,  , ...,  ). In our model, each view  ,
where  = (1, 2, ..., , ...,  ) is the set of views, represents a
dimension  of the multi-variate time series  . Thus, we have
as many views as dimensions.
      </p>
      <p>In step 3, the first-level learner takes as input the time series
data coming from each view. Then, each view will generate its
own predicted labels with associated prediction probabilities
with the form [, 1, 2, ...,   , ...,  , ], where  is the predicted
label of the first-level learner ,   is the associated prediction
probability for each class  of the  possible classes, and  is the
true label.</p>
      <p>
        One of the advantages of multi-view learning is its versatility
on first and second level learners’ choices. One can flexibly
substitute classifier choices as wished. We opt for k-nearest-neighbor
(kNN) classifier coupled with the Dynamic Time Warping (DTW)
distance as first-level learner to be trained on each view of the
data. kNN is one of the most popular and traditional TSC
approaches. kNN with DTW metric was considered for a long time
the state-of-the-art in the time series classification problem [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ].
Furthermore, most classification approaches require parameters
settings, whereas kNN with DTW is parameter free. kNN
classiifer has shown to be a strong baseline [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ], and there is no single
distance measure that significantly outperforms DTW [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]. Hence,
recent research has focused on developing ensemble methods
that significantly outperforms the NN coupled with DTW [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ].
      </p>
      <p>In step 4, we aimed at giving a weight for each learner, thus
a new dataset  ′ is generated by joining the first-level learner
predictions and the probability of each prediction, Table 1 shows
the feature vector in this dataset, where  is the predicted label
of the first-level learner ,  is the probability of this prediction,
and  is the true label.</p>
      <p>
        In step 5, after generating a new dataset  ′, a second-level
classifier, or meta-learner, is trained over  ′ through ensemble
learning [
        <xref ref-type="bibr" rid="ref31">31</xref>
        ]. This approach allows to preserve the statistical
properties of each view and learn the classes of the MTS instances
with a significant improvement in the classification accuracy.
      </p>
      <p>
        Many ensemble methods [
        <xref ref-type="bibr" rid="ref31">31</xref>
        ] have been proposed to further
enhance the algorithm’s accuracy by combining learners rather
than trying to find the best single learner. Due to their versatility
and flexibility, ensemble methods attract many researchers and
can be applied in diferent domains, for example, but not limited
to, time series classification [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] and time series segmentation
[
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]. In a previous work [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ], we used a multi-view approach for
segmenting MCS data, where we employed an unsupervised
learning for change detection on each view.
      </p>
      <p>
        In this work, we conduct our experiments using Random Forest
classifier since it has shown high performance when it is applied
in the human activity recognition domain [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ].
4
      </p>
    </sec>
    <sec id="sec-10">
      <title>EXPERIMENTS AND RESULTS</title>
      <p>
        The experiments are carried out on diferent environments. The
multi-view learning model was implemented in Python 3.6 using
scikit-learn 0.23.2 and tslearn [
        <xref ref-type="bibr" rid="ref21">21</xref>
        ]. The deep-learning models
(MLSTM-FCN [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ], TapNet [
        <xref ref-type="bibr" rid="ref29">29</xref>
        ]) were trained on a single Tesla
V100 GPU of 32 Go memory with CUDA 10.2, using respectively
Keras 2.2.4 and PyTorch 1.2.0.
      </p>
    </sec>
    <sec id="sec-11">
      <title>Experimental Settings</title>
      <p>We evaluated the used models in these experiments using real
life data collected within the scope of Polluscope project. For
the experiments we have used the participants’ ambient air data
containing (PM10, PM1.0, PM2.5, NO2, BC, Temperature, and
Relative Humidity), in addition to the speed dimension derived
from the geo-locational data. Moreover, we have 8 classes (i.e.,
micro-environment to recognize), which can be divided into two
categories indoor such as (“Home”, “Ofice”, “Restaurant”, and
"Store"), and outdoor such as (“Street”, “Bus”, “Car”, and “Train”).</p>
      <p>We have selected the data of six participants who have
thoroughly annotated their activities within the campaign. Data were
split into two third for training and one third for testing, with
care taken to keep the data of each participant grouped either
in the training or in the testing sets. We used the cross
validation score with "Repeated Stratified K-fold" in order to split the
training set into training and validation, while we test the overall
accuracy using the test dataset.</p>
      <p>To account for the temporal feature of the data, we segment
them into samples of 10 minutes’ length at maximum. Usually,
people spend most of their time indoors, thus we should take into
consideration other outdoor activities that have a short period of
time compared to indoor activities. For example, the average time
spent in "Bus" is around 8 minutes, for "Car" the average time
is 20 minutes, etc. Globally, the distribution of data samples is
highly imbalanced over the diferent classes, as shown in Figure
3a, which reflects the imbalance of time spent in diferent
microenvironments. Imbalanced classes usually cause low performance
classification for the minority classes. To cope with this problem,
we apply a resampling strategy. Figure 3 shows the distribution
of the data in both the original and the re-sampled dataset. We
have used the random over/under-sampler in order to balance
our dataset.</p>
      <p>To consider the valuable variable of the mobility information,
we carry out our experiments on the datasets while considering
or not the speed variable. We also compare the classifiers’
performance on both resampled and original (i.e., un-resampled) data.
Finally, we introduce and evaluate a two-steps approach, by first
discriminating indoor from outdoor environments, followed by
a refinement step to learn a more specific class.
4.2</p>
    </sec>
    <sec id="sec-12">
      <title>Classification Results</title>
      <p>
        This section details the experimental results. The micro-environment
recognition is formulated as a MTSC problem. We used as a
baseline a basic kNN classifier with DTW as distance metric. To
compare our multi-view learning approach (2NN-DTW for the
ifrst-level learners and Random Forest as a meta-learner) with
state-of-the-art techniques, we implemented the MLSTM-FCN
[
        <xref ref-type="bibr" rid="ref13">13</xref>
        ] and run it on a GPU, since it requires more computational
resources than the multi-view approach. Both kNN-DTW and
MLSTM-FCN were applied on an aggregated feature vector
containing all the dimensions together.
      </p>
      <p>As shown in Table 2, kNN-DTW classifier is not able to
discriminate correctly between the micro-environments, while the
accuracy improves when applying the multi-view approach that
treats each view independently, thus preserving the statistical
features of each view. For the third experiment, a Long Short
Term Memory network is generated in order to learn a
mapping between the input vector and the classes. MLSTM-FCN has
shown promising results in the experiments. Table 2 shows the
accuracy of experiments carried out with diferent conditions.</p>
      <p>Next, we focus on the comparison between the proposed
approach and MLSTM-FCN. We also study the impact of using or
not the mobility data, as well as the learning from the original
or the re-sampled data. We report the performances in terms of
recall, and F1 score. These results are grouped in Table 2 to Table
7 and Figure 4 to Figure 6. Table 2 represents the overall accuracy
of the diferent classifiers used in these experiments while or not
using the Speed data, and with or without re-sampling. Table
3 and 4 reports the precision, recall, and the F1-Score metrics
of the Multi-view learner for raw data and re-sampled data
respectively with and without Speed. Table 5 and 6 reports the
precision, recall, and F1-Score metrics of the MLSTM-FCN for
raw data and re-sampled data respectively with and without
Speed. Figure 4 shows the accuracy among diferent views used
within the experiment of the Multi-view approach. Moreover,
figure 5a and 5b shows the confusion matrix when applying the
Multi-view approach on the re-sampled data with/out the speed
respectively. Figure 6 shows the procedure used for the Grouping
step approach which is also based on the multi-view approach,
and table 7 reports the precision, recall, and F1-Score metrics for
this approach.</p>
      <p>The multi-view learner proposed in these experiments
employs the stacked generalization approach, which combines the
predictions of each independent view in order to get the final
classification result. As shown in figure 4 although the first level
learners may have a low accuracy but the combination of their
predictions, by generating a new dataset D’ and feeding it to
train, the meta-learner can improve the accuracy a lot.</p>
      <p>We observe an improvement of the accuracy of the overall
classification when adding the speed dimension to the ambient air
dimensions. We also notice from the confusion matrix in figures
class</p>
      <p>class
5a and 5b and the recall and F1 score metrics in table 4 that the
model can easily discriminate between the “indoor” and “outdoor”
activities, but it cannot perfectly distinguish between the
microenvironments inside each category. For example, even though
most of the samples in the “Train” micro-environment is falsely
predicted as “Car”, both “Car” and “Train” micro-environments
can be classified as outdoor. Based on this observation, we
introduced a grouping step before recognizing the micro-environment.
In this step we classify the sample into either an “Indoor” or
“Outdoor” environment. Based on the classification result, a model will
be specialized for each indoor or outdoor micro-environments.
Figure 6 shows the added step and the procedure for the
classification.</p>
      <p>The accuracy of the classifier in the grouping phase (“Indoor”
or “Outdoor”) showed a good result when using the resampled
data. It reaches 0.82 for data without speed dimension, while for
data with the speed dimension it reaches around 0.83. Table7
shows the recall and F1 score for both models trained on
resampled data with and without speed. These experiments did not
consider the case with original data due to its low performance,
in particular, for the minority classes.
5</p>
    </sec>
    <sec id="sec-13">
      <title>DISCUSSIONS &amp; PERSPECTIVES</title>
      <p>In this section, we discuss the perspectives for improving our
multi-view learning model and the possibility for tackling the
practical label issue in the context of Polluscope.
The multi-view learner adopted in this paper is composed by the
base learner (i.e., kNN-DTW) and the meta-learner (i.e., Random
Forest), which has greatly improved the performance compared to
the single kNN-DTW classifier. The objective of this paper is not
to propose the best classifier for MTS classicfiation, but to provide
an insight that the multi-view learner is capable of coordinating
efectively the information from diferent variables and achieving
more reliable performance than a single base learner. Moreover,
the results of the grouping approach which is based on the
multiview approach confirms that there is a clear signature for each
micro-environment, thus we can have an efective prediction
with this approach.</p>
      <p>
        Nevertheless, the kNN-DTW is considered as the baseline for
MTS classification and is widely outpaced by the advanced
approaches such as Shapelets [
        <xref ref-type="bibr" rid="ref27 ref32 ref33">27, 32, 33</xref>
        ] or the frequent patterns
[
        <xref ref-type="bibr" rid="ref18">18</xref>
        ]. Essentially, the kNN-DTW captures the global feature based
on the distance measure between the entire sequences, while
the local features (e.g., the frequent patterns [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ], the interval
features [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ], Shapelets [
        <xref ref-type="bibr" rid="ref27">27</xref>
        ], etc.) are more appropriate when a
specific pattern characterizes a class. More specifically, a
combination of features extracted from diferent domains may improve
dramatically the performance of the base learner [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ]. Therefore,
one of the perspectives consists in the optimization of the base
learner and the exploration of the explainability of the
multiview learner on both the feature interpretation and the variable
importance for building the classifier. The visual representation
of Shapelets make them good candidates for such improvement.
5.2
      </p>
    </sec>
    <sec id="sec-14">
      <title>Label Shortage Issue</title>
      <p>The label shortage is a practical issue when building the learning
model. In particular, in the context of Polluscope, post-labelling
for time series sensor data is much more costly than classic data
(e.g., image, text, etc.) due to the low interpretability over the
realvalued sequence. Therefore, the data need to be annotated during
the data collection process. However, certain practical factors
limit the availability of labels. For instance, the participants are
not always conscious in annotating their micro-environment.
Therefore, for certain time periods, no annotations were marked.
(a) With Speed
(b) Without Speed</p>
      <p>
        In order to give an insight for the consistency between the
labeled and unlabeled data, and to see if the unlabeled data are
valuable for improving the classifier’s performance in our context,
we conduct a preliminary test on the Polluscope data with the
newly proposed semi-supervised MTSC model TapNet [
        <xref ref-type="bibr" rid="ref29">29</xref>
        ].
      </p>
      <p>
        TapNet [
        <xref ref-type="bibr" rid="ref29">29</xref>
        ] is a deep learning based approach designed for
multivariate time series classification. By adopting the
prototypical network [
        <xref ref-type="bibr" rid="ref20">20</xref>
        ], TapNet allows learning a low-dimensional
embeddings for the input MTS where the unlabelled samples
help adjusting the class prototype (i.e., class centroid), which
leads to a better classifier than using only the labelled samples.
Table 8 shows the semi-supervised learning results on
Polluscope data considering or not the speed variable. We evaluate the
performance of TapNet under diferent supervision ratios in the
training set. The results show that the unlabeled samples and
the speed variable do improve the performance of the classifier.
Besides, the accuracy didn’t drop a lot when eliminating the
annotations in training set (from ratio=1 for fully labelled to 0.5,
and even for 0.2 when only 20% data in labelled), indicating that
the collected data within each class is not sparsely distributed,
thus learning under weak supervision is reliable with the aid of
the unlabeled samples.
      </p>
      <p>
        Giving the promising results on the data distribution
consistency, another avenue worth exploring is to consider and
integrate a semi-supervised model into our multi-view learner.
Various semi-supervised frameworks are applicable to our model,
such as applying self-learning [
        <xref ref-type="bibr" rid="ref25">25</xref>
        ] to produce the pseudo labels
on the multi-view learner, or adopting the label propagation and
manifold regularization techniques [
        <xref ref-type="bibr" rid="ref22">22</xref>
        ] on the base learner.
6
      </p>
    </sec>
    <sec id="sec-15">
      <title>CONCLUSION</title>
      <p>Activity recognition has gained the interest of many researchers
nowadays, due to the widespread use of mobility sensors.
Microenvironment recognition is essential in MCS projects such as
Polluscope, in order to be able to analyse the individual’s
exposure to air pollution and to relate it to her context. The major
ifnding of our study is to show to some extent that the ambient air
can characterize the micro-environment. Moreover, the accuracy
of the model is high enough to consider an automatic detection
of the micro-environment without burdening the participants
with self-reporting. By using the mobility feature, the accuracy
improves slightly though the gain is moderate. Therefore, we can
keep characterizing the micro-environment even in the absence
of the speed dimension.</p>
      <p>We employed diferent approaches and learners, and
conducted a thorough experimental study, which shows the
eficiency of MLSTM-FCN and the multi-view approach for time
series classification. We have also compared the results with the
kNN-DTW classifier which was considered as the baseline.</p>
      <p>We have also identified several perspectives of this work, and
explored the application of semi-supervised learning to cope
with the lack of labels for some classes. In future work, we can
use various algorithms for the first level learner and the
metalearner, as multi-view learning is flexible. Finally, we intend to
improve the performance of the learned classes by integrating
some a priori rules, like the unlikelihood of being in some
microenvironment at some time of day, or of transitions between some
micro-environments.</p>
    </sec>
    <sec id="sec-16">
      <title>ACKNOWLEDGMENTS</title>
      <p>This work has supported by the French National Research Agency
(ANR) project Polluscope, funded under the grant agreement
ANR-15-CE22-0018, by the H2020 EU GO GREEN ROUTES funded
under the research and innovation programme H2020- EU.3.5.2
grant agreement No 869764, and by the DATAIA convergence
institute project StreamOps, as part of the Programme d’
Investissement d’Avenir, ANR-17-CONV-0003. Part of the equipment was
funded by iDEX Paris-Saclay, in the framework of the IRS project
ACE-ICSEN, and by the Communauté d’agglomération Versailles
Grand Parc – VGP - (www.versaillesgrandparc.fr). We are
thankful to VGP (Thomas Bonhoure) for facilitating the campaign. We
would like to thank all the members of the Polluscope consortia
who contributed in one way or another to this work: Salim Srairi
and Jean-Marc Naude (CEREMA) who conducted the campaign;
Boris Dessimond and Isabella Annesi-Maesano (Sorbonne
University) for their contribution to the campaign; Valerie Gros and
Nicolas Bonnaire (LSCE), and Anne Kaufman and Christophe
Debert (Airparif ) for their contribution in the periodic
qualification of the sensors and their active involvement in the project.
Finally, we would like to thank the participants for their great
efort in carrying the sensors, without whom this work would
not be possible.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>Samaneh</given-names>
            <surname>Aminikhanghahi</surname>
          </string-name>
          and
          <string-name>
            <given-names>Diane J.</given-names>
            <surname>Cook</surname>
          </string-name>
          .
          <year>2019</year>
          .
          <article-title>Enhancing activity recognition using CPD-based activity segmentation</article-title>
          .
          <source>Pervasive and Mobile Computing</source>
          <volume>53</volume>
          (
          <year>2019</year>
          ),
          <fpage>75</fpage>
          -
          <lpage>89</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <surname>Donald</surname>
            <given-names>J</given-names>
          </string-name>
          <string-name>
            <surname>Berndt and James Cliford</surname>
          </string-name>
          .
          <year>1994</year>
          .
          <article-title>Using Dynamic Time Warping to Find Patterns in Time Series</article-title>
          .
          <source>In KDD workshop</source>
          , Vol.
          <volume>10</volume>
          . Seattle, WA, USA:,
          <volume>359</volume>
          -
          <fpage>370</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>Kaixuan</given-names>
            <surname>Chen</surname>
          </string-name>
          , Dalin Zhang, Lina Yao, Bin Guo,
          <string-name>
            <given-names>Zhiwen</given-names>
            <surname>Yu</surname>
          </string-name>
          , and Yunhao Liu.
          <year>2020</year>
          .
          <article-title>Deep Learning for Sensor-based Human Activity Recognition: Overview, Challenges and Opportunities</article-title>
          . arXiv:
          <year>2001</year>
          .07416 [cs] (
          <year>Jan</year>
          .
          <year>2020</year>
          ). http://arxiv.org/abs/
          <year>2001</year>
          .07416 arXiv:
          <year>2001</year>
          .07416.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>Heeryon</given-names>
            <surname>Cho</surname>
          </string-name>
          and Sang Min Yoon.
          <year>2018</year>
          .
          <article-title>Divide and conquer-based 1D CNN human activity recognition using test data sharpening</article-title>
          .
          <source>Sensors</source>
          <volume>18</volume>
          ,
          <issue>4</issue>
          (
          <year>2018</year>
          ),
          <fpage>1055</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>Houtao</given-names>
            <surname>Deng</surname>
          </string-name>
          , George Runger, Eugene Tuv, and
          <string-name>
            <given-names>Martyanov</given-names>
            <surname>Vladimir</surname>
          </string-name>
          .
          <year>2013</year>
          .
          <article-title>A time series forest for classification and feature extraction</article-title>
          .
          <source>Information Sciences</source>
          <volume>239</volume>
          (
          <year>2013</year>
          ),
          <fpage>142</fpage>
          -
          <lpage>153</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>T. M. T.</given-names>
            <surname>Do</surname>
          </string-name>
          and
          <string-name>
            <given-names>D.</given-names>
            <surname>Gatica-Perez</surname>
          </string-name>
          .
          <year>2014</year>
          .
          <article-title>The Places of Our Lives: Visiting Patterns and Automatic Labeling from Longitudinal Smartphone Data</article-title>
          .
          <source>IEEE Transactions on Mobile Computing 13, 3 (March</source>
          <year>2014</year>
          ),
          <fpage>638</fpage>
          -
          <lpage>648</lpage>
          . https://doi. org/10.1109/TMC.
          <year>2013</year>
          .19
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>Hafsa</given-names>
            <surname>El Hafyani</surname>
          </string-name>
          .
          <year>2020</year>
          .
          <article-title>In 2020 21st IEEE International Conference on Mobile Data Management (MDM)</article-title>
          . IEEE,
          <fpage>246</fpage>
          -
          <lpage>247</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>Hafsa</given-names>
            <surname>El</surname>
          </string-name>
          <string-name>
            <surname>Hafyani</surname>
          </string-name>
          , Karine Zeitouni, Yehia Taher, and
          <string-name>
            <given-names>Mohammad</given-names>
            <surname>Abboud</surname>
          </string-name>
          .
          <year>2020</year>
          .
          <article-title>Leveraging Change Point Detection for Activity Transition Mining in the Context of Environmental Crowdsensing</article-title>
          .
          <source>The 9th SIGKDD International Workshop on Urban Computing</source>
          (
          <year>2020</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>Hassan</given-names>
            <surname>Ismail</surname>
          </string-name>
          <string-name>
            <surname>Fawaz</surname>
          </string-name>
          , Germain Forestier, Jonathan Weber,
          <string-name>
            <given-names>Lhassane</given-names>
            <surname>Idoumghar</surname>
          </string-name>
          , and
          <string-name>
            <surname>Pierre-Alain Muller</surname>
          </string-name>
          .
          <year>2019</year>
          .
          <article-title>Deep learning for time series classification: a review</article-title>
          .
          <source>Data Mining and Knowledge Discovery</source>
          <volume>33</volume>
          ,
          <issue>4</issue>
          (
          <year>2019</year>
          ),
          <fpage>917</fpage>
          -
          <lpage>963</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>Hassan</given-names>
            <surname>Ismail Fawaz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Lucas</surname>
          </string-name>
          , G. Forestier,
          <string-name>
            <given-names>Charlotte</given-names>
            <surname>Pelletier</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Schmidt</surname>
          </string-name>
          , Jonathan Weber,
          <string-name>
            <given-names>Geofrey I.</given-names>
            <surname>Webb</surname>
          </string-name>
          , L. Idoumghar,
          <string-name>
            <surname>Pierre-Alain Muller</surname>
            , and
            <given-names>Franccois</given-names>
          </string-name>
          <string-name>
            <surname>Petitjean</surname>
          </string-name>
          .
          <year>2020</year>
          .
          <article-title>InceptionTime: Finding AlexNet for Time Series Classification</article-title>
          . ArXiv abs/
          <year>1909</year>
          .04939 (
          <year>2020</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>Enrique</given-names>
            <surname>Garcia-Ceja</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Carlos E.</given-names>
            <surname>Galván-Tejada</surname>
          </string-name>
          , and
          <string-name>
            <given-names>Ramon</given-names>
            <surname>Brena</surname>
          </string-name>
          .
          <year>2018</year>
          .
          <article-title>Multiview stacking for activity recognition with sound and accelerometer data</article-title>
          .
          <source>Information Fusion</source>
          <volume>40</volume>
          (
          <year>March 2018</year>
          ),
          <fpage>45</fpage>
          -
          <lpage>56</lpage>
          . https://doi.org/10.1016/j.infus.
          <year>2017</year>
          .
          <volume>06</volume>
          .004
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <surname>Bin</surname>
            <given-names>Guo</given-names>
          </string-name>
          , Zhu Wang,
          <string-name>
            <surname>Zhiwen Yu</surname>
          </string-name>
          ,
          <string-name>
            <surname>Yu</surname>
            <given-names>Wang</given-names>
          </string-name>
          , Neil Y Yen,
          <string-name>
            <given-names>Runhe</given-names>
            <surname>Huang</surname>
          </string-name>
          , and
          <string-name>
            <given-names>Xingshe</given-names>
            <surname>Zhou</surname>
          </string-name>
          .
          <year>2015</year>
          .
          <article-title>Mobile crowd sensing and computing: The review of an emerging human-powered sensing paradigm</article-title>
          .
          <source>ACM computing surveys (CSUR) 48</source>
          ,
          <issue>1</issue>
          (
          <year>2015</year>
          ),
          <fpage>1</fpage>
          -
          <lpage>31</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <surname>Fazle</surname>
            <given-names>Karim</given-names>
          </string-name>
          , Somshubra Majumdar, Houshang Darabi, and
          <string-name>
            <given-names>Samuel</given-names>
            <surname>Harford</surname>
          </string-name>
          .
          <year>2019</year>
          .
          <article-title>Multivariate LSTM-FCNs for time series classification</article-title>
          .
          <source>Neural Networks</source>
          <volume>116</volume>
          (
          <year>2019</year>
          ),
          <fpage>237</fpage>
          -
          <lpage>245</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <surname>Baptiste</surname>
            <given-names>Languille</given-names>
          </string-name>
          , Valérie Gros, Nicolas Bonnaire, Clément Pommier, Cécile Honoré, Christophe Debert, Laurent Gauvin, Salim Srairi,
          <string-name>
            <surname>Isabella</surname>
            <given-names>AnnesiMaesano</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Basile Chaix</surname>
          </string-name>
          , et al.
          <year>2020</year>
          .
          <article-title>A methodology for the characterization of portable sensors for air quality measure with the goal of deployment in citizen science</article-title>
          .
          <source>Science of the Total Environment</source>
          <volume>708</volume>
          (
          <year>2020</year>
          ),
          <fpage>134698</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>Sheng</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>and Yun</given-names>
            <surname>Fu</surname>
          </string-name>
          .
          <year>2016</year>
          .
          <article-title>Multi-View Time Series Classification: A Discriminative Bilinear Projection Approach</article-title>
          .
          <source>Proceedings of the 25th ACM International on Conference on Information and Knowledge Management</source>
          (
          <year>2016</year>
          ),
          <fpage>989</fpage>
          -
          <lpage>998</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <surname>Jason</surname>
            <given-names>Lines</given-names>
          </string-name>
          , Sarah Taylor, and
          <string-name>
            <given-names>Anthony</given-names>
            <surname>Bagnall</surname>
          </string-name>
          .
          <year>2016</year>
          .
          <article-title>HIVE-COTE: The Hierarchical Vote Collective of Transformation-based Ensembles for Time Series Classification</article-title>
          .
          <source>In 2016 IEEE 16th international conference on data mining (ICDM)</source>
          .
          <volume>1041</volume>
          -
          <fpage>1046</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <surname>Li</surname>
            <given-names>Liu</given-names>
          </string-name>
          , Yuxin Peng, Shu Wang, Ming Liu, and
          <string-name>
            <given-names>Zigang</given-names>
            <surname>Huang</surname>
          </string-name>
          .
          <year>2016</year>
          .
          <article-title>Complex activity recognition using time series pattern dictionary learned from ubiquitous sensors</article-title>
          .
          <source>Information Sciences 340-341 (May</source>
          <year>2016</year>
          ),
          <fpage>41</fpage>
          -
          <lpage>57</lpage>
          . https: //doi.org/10.1016/j.ins.
          <year>2016</year>
          .
          <volume>01</volume>
          .020
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <surname>Guruprasad</surname>
            <given-names>Nayak</given-names>
          </string-name>
          , Varun Mithal, Xiaowei Jia, and
          <string-name>
            <given-names>Vipin</given-names>
            <surname>Kumar</surname>
          </string-name>
          .
          <year>2018</year>
          .
          <article-title>Classifying multivariate time series by learning sequence-level discriminative patterns</article-title>
          .
          <source>In Proceedings of the 2018 SIAM International Conference on Data Mining. SIAM</source>
          ,
          <fpage>252</fpage>
          -
          <lpage>260</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <surname>Juha</surname>
            <given-names>Pärkkä</given-names>
          </string-name>
          , Miikka Ermes, Panu Korpipää, Jani Mäntyjärvi, Johannes Peltola, and
          <string-name>
            <given-names>Ilkka</given-names>
            <surname>Korhonen</surname>
          </string-name>
          .
          <year>2006</year>
          .
          <article-title>Activity classification using realistic data from wearable sensors</article-title>
          .
          <source>IEEE transactions on information technology in biomedicine: a publication of the IEEE Engineering in Medicine and Biology Society</source>
          <volume>10</volume>
          ,
          <issue>1</issue>
          (Jan.
          <year>2006</year>
          ),
          <fpage>119</fpage>
          -
          <lpage>128</lpage>
          . https://doi.org/10.1109/titb.
          <year>2005</year>
          .856863
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <surname>Jake</surname>
            <given-names>Snell</given-names>
          </string-name>
          , Kevin Swersky, and Richard Zemel.
          <year>2017</year>
          .
          <article-title>Prototypical Networks for Few-shot Learning</article-title>
          .
          <source>In Advances in Neural Information Processing Systems</source>
          , Vol.
          <volume>30</volume>
          . Curran Associates, Inc.,
          <fpage>4077</fpage>
          -
          <lpage>4087</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <surname>Romain</surname>
            <given-names>Tavenard</given-names>
          </string-name>
          , Johann Faouzi, Gilles Vandewiele, Felix Divo, Guillaume Androz, Chester Holtz, Marie Payne, Roman Yurchak, Marc Rußwurm, Kushal Kolar, and
          <string-name>
            <given-names>Eli</given-names>
            <surname>Woods</surname>
          </string-name>
          .
          <year>2020</year>
          .
          <article-title>Tslearn, A Machine Learning Toolkit for Time Series Data</article-title>
          .
          <source>Journal of Machine Learning Research</source>
          <volume>21</volume>
          ,
          <issue>118</issue>
          (
          <year>2020</year>
          ),
          <fpage>1</fpage>
          -
          <lpage>6</lpage>
          . http: //jmlr.org/papers/v21/
          <fpage>20</fpage>
          -
          <lpage>091</lpage>
          .html
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <surname>Jesper</surname>
            <given-names>E. van Engelen and H.</given-names>
          </string-name>
          <string-name>
            <surname>Hoos</surname>
          </string-name>
          .
          <year>2019</year>
          .
          <article-title>A survey on semi-supervised learning</article-title>
          .
          <source>Machine Learning</source>
          <volume>109</volume>
          (
          <year>2019</year>
          ),
          <fpage>373</fpage>
          -
          <lpage>440</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <surname>Baoquan</surname>
            <given-names>Wang</given-names>
          </string-name>
          , Tonghai Jiang, Xi Zhou, Bo Ma, Fan
          <string-name>
            <surname>Zhao</surname>
            ,
            <given-names>and Yi</given-names>
          </string-name>
          <string-name>
            <surname>Wang</surname>
          </string-name>
          .
          <year>2020</year>
          .
          <article-title>Time-Series Classification Based on Fusion Features of Sequence and Visualization</article-title>
          .
          <source>Applied Sciences</source>
          <volume>10</volume>
          ,
          <issue>12</issue>
          (Jan.
          <year>2020</year>
          ),
          <volume>4124</volume>
          . https://doi.org/10. 3390/app10124124
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [24]
          <string-name>
            <surname>Jindong</surname>
            <given-names>Wang</given-names>
          </string-name>
          , Yiqiang Chen, Shuji Hao, Xiaohui Peng, and
          <string-name>
            <given-names>Lisha</given-names>
            <surname>Hu</surname>
          </string-name>
          .
          <year>2019</year>
          .
          <article-title>Deep Learning for Sensor-based Activity Recognition: A Survey</article-title>
          .
          <source>Pattern Recognition Letters</source>
          <volume>119</volume>
          (
          <year>March 2019</year>
          ),
          <fpage>3</fpage>
          -
          <lpage>11</lpage>
          . https://doi.org/10.1016/j.patrec.
          <year>2018</year>
          .
          <volume>02</volume>
          .010 arXiv:
          <fpage>1707</fpage>
          .
          <fpage>03502</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          [25]
          <string-name>
            <given-names>Li</given-names>
            <surname>Wei</surname>
          </string-name>
          and
          <string-name>
            <given-names>Eamonn</given-names>
            <surname>Keogh</surname>
          </string-name>
          .
          <year>2006</year>
          .
          <article-title>Semi-supervised time series classification</article-title>
          .
          <source>In Proceedings of the 12th ACM SIGKDD international conference on Knowledge discovery and data mining (KDD '06)</source>
          .
          <article-title>Association for Computing Machinery</article-title>
          , New York, NY, USA,
          <fpage>748</fpage>
          -
          <lpage>753</lpage>
          . https://doi.org/10.1145/1150402.1150498
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          [26]
          <string-name>
            <surname>David</surname>
            <given-names>H.</given-names>
          </string-name>
          <string-name>
            <surname>Wolpert</surname>
          </string-name>
          .
          <year>1992</year>
          .
          <article-title>Stacked generalization</article-title>
          .
          <source>Neural Networks</source>
          <volume>5</volume>
          ,
          <issue>2</issue>
          (
          <year>1992</year>
          ),
          <fpage>241</fpage>
          -
          <lpage>259</lpage>
          . https://doi.org/10.1016/S0893-
          <volume>6080</volume>
          (
          <issue>05</issue>
          )
          <fpage>80023</fpage>
          -
          <lpage>1</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          [27]
          <string-name>
            <given-names>Lexiang</given-names>
            <surname>Ye</surname>
          </string-name>
          and
          <string-name>
            <given-names>Eamonn</given-names>
            <surname>Keogh</surname>
          </string-name>
          .
          <year>2009</year>
          .
          <article-title>Time series shapelets: A New Primitive for Data Mining</article-title>
          .
          <source>Proceedings of the 15th ACM SIGKDD international conference on Knowledge discovery and data mining - KDD '09</source>
          (
          <year>2009</year>
          ),
          <fpage>947</fpage>
          -
          <lpage>956</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          [28]
          <string-name>
            <given-names>Mi</given-names>
            <surname>Zhang and Alexander</surname>
          </string-name>
          <string-name>
            <given-names>A.</given-names>
            <surname>Sawchuk</surname>
          </string-name>
          .
          <year>2012</year>
          .
          <article-title>Motion primitive-based human activity recognition using a bag-of-features approach</article-title>
          .
          <source>In Proceedings of the 2nd ACM SIGHIT International Health Informatics Symposium (IHI '12)</source>
          .
          <article-title>Association for Computing Machinery</article-title>
          , New York, NY, USA,
          <fpage>631</fpage>
          -
          <lpage>640</lpage>
          . https://doi.org/10. 1145/2110363.2110433
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          [29]
          <string-name>
            <surname>Xuchao</surname>
            <given-names>Zhang</given-names>
          </string-name>
          , Yifeng Gao,
          <string-name>
            <given-names>Jessica</given-names>
            <surname>Lin</surname>
          </string-name>
          , and
          <string-name>
            <surname>Chang-Tien Lu</surname>
          </string-name>
          .
          <year>2020</year>
          .
          <article-title>TapNet: Multivariate Time Series Classification with Attentional Prototypical Network</article-title>
          .
          <source>In Proceedings of the AAAI Conference on Artificial Intelligence</source>
          , Vol.
          <volume>34</volume>
          .
          <fpage>6845</fpage>
          -
          <lpage>6852</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          [30]
          <string-name>
            <surname>Yu</surname>
            <given-names>Zheng</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>Quannan</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Yukun</given-names>
            <surname>Chen</surname>
          </string-name>
          , Xing Xie, and
          <string-name>
            <surname>Wei-Ying Ma</surname>
          </string-name>
          .
          <year>2008</year>
          .
          <article-title>Understanding mobility based on GPS data</article-title>
          .
          <source>In Proceedings of the 10th international conference on Ubiquitous computing. Association for Computing Machinery</source>
          , New York, NY, USA,
          <fpage>312</fpage>
          -
          <lpage>321</lpage>
          . https://doi.org/10.1145/1409635.1409677
        </mixed-citation>
      </ref>
      <ref id="ref31">
        <mixed-citation>
          [31]
          <string-name>
            <surname>Zhi-Hua Zhou</surname>
          </string-name>
          .
          <year>2012</year>
          .
          <article-title>Ensemble Methods: Foundations and Algorithms</article-title>
          . CRC press.
        </mixed-citation>
      </ref>
      <ref id="ref32">
        <mixed-citation>
          [32]
          <string-name>
            <surname>Jingwei</surname>
            <given-names>Zuo</given-names>
          </string-name>
          , Karine Zeitouni, and
          <string-name>
            <given-names>Yehia</given-names>
            <surname>Taher</surname>
          </string-name>
          .
          <year>2019</year>
          .
          <article-title>Exploring Interpretable Features for Large Time Series with SE4TeC</article-title>
          .
          <source>In Proc. EDBT</source>
          .
          <volume>606</volume>
          -
          <fpage>609</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref33">
        <mixed-citation>
          [33]
          <string-name>
            <surname>Jingwei</surname>
            <given-names>Zuo</given-names>
          </string-name>
          , Karine Zeitouni, and
          <string-name>
            <given-names>Yehia</given-names>
            <surname>Taher</surname>
          </string-name>
          .
          <year>2019</year>
          .
          <article-title>Incremental and Adaptive Feature Exploration over Time Series Stream</article-title>
          .
          <source>In 2019 IEEE International Conference on Big Data (Big Data)</source>
          .
          <fpage>593</fpage>
          -
          <lpage>602</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>