=Paper= {{Paper |id=Vol-2626/paper9 |storemode=property |title=Lossy Compression Methods for Performance-Restricted Wearable Devices |pdfUrl=https://ceur-ws.org/Vol-2626/paper9.pdf |volume=Vol-2626 |authors=Lucie Klus,Elena Simona Lohan,Carlos Granell,Jari Nurmi |dblpUrl=https://dblp.org/rec/conf/icl-gnss/KlusLGN20 }} ==Lossy Compression Methods for Performance-Restricted Wearable Devices== https://ceur-ws.org/Vol-2626/paper9.pdf
Lossy Compression Methods for Performance-Restricted
Wearable Devices
Lucie Klusa , Elena Simona Lohana , Carlos Granellb and Jari Nurmia
a
    Tampere University, Electrical Engineering Unit, Tampere, Finland
b
    Universitat Jaume I, Institute of New Imaging Technologies, Castellon, Spain


                                         Abstract
                                         With the increasing popularity, diversity, and utilization of wearable devices, the data transfer-and-
                                         storage efficiency becomes increasingly important. This paper evaluates a set of compression tech-
                                         niques regarding their utilization in crowdsourced wearable data. Transform-based Discrete Cosine
                                         Transform (DCT), interpolation-based Lightweight Temporal Compression (LTC) and dimensional-
                                         ity reduction-focused Symbolic Aggregate Approximation (SAX) were chosen as traditional methods.
                                         Additionally, an altered SAX (ASAX) is proposed by the authors and implemented to overcome some
                                         of the shortcomings of the traditional methods. As one of the most commonly measured entities in
                                         wearable devices, heart rate data were chosen to compare the performance and complexity of the
                                         selected compression methods. Main results suggest that best compression results are obtained with
                                         LTC, which is also the most complex of the studied methods. The best performance-complexity
                                         trade-off is achieved with SAX. Our proposed ASAX has the best dynamic properties among the
                                         evaluated methods.

                                         Keywords
                                         Compression, Discrete Cosine Transform (DCT), Lightweight Temporal Compression (LTC),
                                         Heart Rate, Symbolic Aggregation Approximation (SAX), Wearables




1. Introduction
During the last decade, wearable devices have become increasingly popular, with smart watches
and wristbands dominating the market [1]. Wearables have found their utilization in a wide
variety of fields, including medical research and patient monitoring, person tracking for en-
hanced security, and various applications in industrial halls. As the size of the devices is one
determining factor in a wearable product, the battery requirements are typically very strict,
leading to limited computational capabilities in the sense of hardware- or software-restricted
processing power and data storage.
   The amount of data produced by wearable devices increases continuously, especially in health
monitoring applications, where physiological data such heart rate, ECG, breath rate, body tem-
perature etc. are often monitored to prevent certain diseases, to track patients with long-lasting
illnesses, or to monitor the changes in one’s health. With the threat of more infectious diseases
such as COVID-19 in the future, also location-related data and person tracking information
are likely to more and more harnessed from wearables for the purpose of disease control and
management. Long-term monitoring requires the storage of data for a long time in order to
ICL-GNSS 2020 WiP Proceedings, June 02–04, 2020, Tampere, Finland
email: lucie.klus@tuni.fi (L. Klus); elena-simona.lohan@tuni.fi (E.S. Lohan); carlos.granell@uji.es (C.
Granell); jari.nurmi@tuni.fi (J. Nurmi)
orcid: 0000-0003-0354-9421 (L. Klus); 0000-0003-1718-6924 (E.S. Lohan); 0000-0003-1004-9695 (C. Granell);
0000-0003-2169-4606 (J. Nurmi)
                                       ⃝
                                       c 2020 Copyright for this paper by its authors.
                                       Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0).
    CEUR
    Workshop
    Proceedings
                  http://ceur-ws.org
                  ISSN 1613-0073       CEUR Workshop Proceedings (CEUR-WS.org)
be able to detect the changes over the span of months or years, while the frequency of certain
measurements is required to be high (e.g. heart-connected measurements). In order to detect
heart attack, for example, heart measurements should be conducted within the intervals of a
few seconds, as in these cases high frequency is the key, because heart rate changes are rapid
and high. The demand for long-term monitoring comes not only from the medical side but also
from the patient’s side, as more and more patients acknowledge the advantages of prevention
and (cardiac) rehabilitation [2].
   Although the current wearable market for health applications offers a variety of heart track-
ers, most of them are not as precise as the heart monitors available in hospitals or are not
suitable for everyday usage (requiring a belt attached to the chest). The precision of the mea-
surements of wrist-worn devices is strongly dependent on the implemented sensors and it is
affected by many additional factors, such as hand movements and how tightly the device is
attached to the wrist.
   In [3], the authors conducted a comparison of 17 heart rate trackers regarding their accuracy
and precision. They found out that the accuracy of the trackers was between 79.8 % and 99.1 %,
concluding that wearable devices are able to collect measurements with reasonable accuracy.
   As the measurements themselves are reflecting the biometric data with a certain error,
reducing the size of gathered data using approximation, thinning, or lossy compression is
acceptable. Nevertheless, certain restrictions on the delay or error have to be applied, mostly
based on the purpose of the device (medical monitoring vs. activity tracking, etc.).
   The amount of data gathered increases with the number of sensors integrated in a wear-
able. Considering, for example, that a device conducts a measurement every 10 seconds, the
resulting total amount of measurements can add up to 8640 measurements per day and about
3153600 measurements per year and per user. Each measurement requires to store at least
the measured value and its timestamp. The most basic measurements may be stored in 4-byte
formats, with timestamp varying between 7 to 13 bytes. In case of 4 bytes per measurement
plus 7 bytes per timestamp, each entry for each sensor requires minimum of 11 bytes to be
stored, resulting up to 34.7 megabytes of data per sensor per user per year.
   The majority of the sensor data from wearables are time-series data, which are usually
highly correlated in time. Therefore, representing such data by sampling between correlated
samples or by other dimensionality-reduction methods may remove redundancy and reduce
noise within the data. Based on the above considerations, there are several reasons to apply
lossy compression on the data. In this paper, four most promising lossy compression techniques
are evaluated based on their compression ratio capabilities, performance requirements and
data distortion due to the compression. In order to select these four methods, an initial
literature review has been performed by the authors and three of the presented methods were
selected based on their good reported properties in the current literature. The fourth considered
compression technique was developed by the authors.


2. Chosen Methods - Theory and properties
This section presents and describes the lossy compression methods utilized in this paper. The
considered methods include Discrete Cosine Transform (DCT) with Run Length Encoding
(RLE), Symbolic Aggregate Approximation (SAX), Altered Symbolic Aggregate Approxima-
tion (ASAX) proposed by the authors for the purposes of this paper and Lightweight Temporal
Compression (LTC).
2.1. Discrete Cosine Transform with Run-Length Encoding (DCT-RLE)
Discrete Cosine Transform (DCT) is a transform-based compression widely utilized in media
compression, including JPEG, MPEG, and various other formats. DCT has strong filtering
properties, being able to reduce also the noise while reducing the size of the data. For these
properties, DCT is utilized in many applications, including wireless positioning [4], where the
authors used spectral analysis in the RSS-based indoor positioning scenario. The implemented
DCT was used to compress the database size of RSS measurements with up to 80 % compression
ratio, while achieving the performance of the conventional fingerprinting. DCT was also used
to retrieve medical images in [5] and finds its utilization in medical images watermarking
schemes [6][7].
  As the core component of the DCT compression, a DCT transformation is used on the input
data (metrics in the shape of N × M ). It results in the matrix of coefficients of the same shape
(N × M ) as the input, referring to the cosine-shaped trends of different frequencies present
within the data. Eq. (1) shows the formula for calculating the resulting coefficients of the
transformation, where u and v are the coordinates in the coefficient matrix, M and N are
the dimensions of input matrix (data block) and f (x, y) is the input matrix value at x and y
coordinates [8].
                 2
  F (u, v) = √      C(u)C(v)
                 NM
                      N −1 M −1              [︃          ]︃     [︃            ]︃
                      ∑︂   ∑︂                 (2x + 1)uπ           (2y + 1)uπ
                                  f (x, y)cos               cos
                                                  2N                   2M
                       x=0 y=0

                                                             for u = 0, ..., N − 1 and v = 0, ..., M − 1 (1)

                                                  {︄
                                                       √1      if k = 0
                                      C(k) =             2
                                                   1           if k > 0

   During the compression, the low-impact components, usually representing the higher fre-
quencies in the data having close-to-zero values, are set to zero, based on the utilized algo-
rithm. The compression ratio, as well as the resulting loss, can be varied based on the number
of components forced to be zero. The data can be reconstructed using an inverse Discrete
Cosine Transform (iDCT), which transforms the matrix of coefficients back to the original
data.
   From the computational point of view, the method is rather expensive [9]. In this paper we
refer to combination of DCT with lossless compression in order to reduce the amount of stored
data, as DCT itself does not decrease number of samples, but it rather transforms the data,
resulting in many close-to-zero values, which may be set to zero. Such a result may be stored
more efficiently by using lossless compression which reduces the redundancy in the data.
   Run-Length Encoding (RLE) is such a lossless compression method. It stores repetitive
values of input signal as a value and number of repetitions instead of storing every value
separately. There are several ways of implementing the method. For the purposes of this
paper, we apply the RLE per-sample. We stored the data as two same-length vectors, one
with values of coefficients, the second one with their counts, as the output of the compression.
     y
                a                  b               b                  c


     c




    b




     a




                                                                                           x
Figure 1: Illustration of SAX Implementation.


2.2. Symbolic Aggregate Approximation (SAX)
Symbolic Aggregate Approximation is a lossy compression method, composed of two basic
steps. The first step is a Piece-wise Aggregate Approximation (PAA) compression, and the
second one is a discretization of the y-axis and alphabet symbol assignment [10].
   PAA compression is a traditional and non-complex type of compression (see Fig. 1), which
takes a fixed range of samples (fixed amount of samples or fixed time interval), calculates their
mean and returns it as the result for the corresponding interval (red dashed lines in the Figure).
Next, SAX discretizes the y-axis into bins, e.g. (a, b, c) displayed as orange lines and assigns
the PAA output (red lines) to the closest alphabet value (orange line) as an output (a, b, b, c).
   The vertical distance between the bins is in most applications derived based on the distribu-
tion of the PAA data, as the normalized time series is assumed to have its distribution close to
the Gaussian [10]. The resulting levels are then set so that each alphabet letter has the same
probability of occurrence. In our implementation, the alphabet is based on the equidistant
values of y-axis. The final result is stored as the output values of y-axis bins, time stamp of
first sample and the length of equidistant sample range of x-axis.
   The main advantage of SAX is the ability to compress data in both x and y axis, resulting
in better compression capabilities compared to the transform-based methods such as DCT
methods. In the literature, SAX is shown to be a lossy compression with wide data-mining
utilization (e.g. clustering, classification or anomaly detection), while requiring less storage
due to the limited alphabet size.

2.3. Altered Symbolic Aggregate Approximation (ASAX)
Altered Symbolic Aggregate Approximation is a compression method proposed by us and
inspired by lossy SAX and lossless RLE. Based on SAX, it incorporates the discretized y-axis
and from RLE the redundancy reduction due to the limited number of symbols. ASAX is, in
     y      a      a     a      b     b     b      b     b    b   c     c     c



     c




     b




     a




                                                                                          x
Figure 2: Illustration of the proposed ASAX implementation.


our opinion, also similar to Adaptive Piecewise Constant Approximation (APCA), proposed
in [11], while limiting the alphabet size instead of averaging the sequences.
   In comparison to SAX, ASAX method does not apply PAA, and it directly transforms the
input values to the alphabet levels. Each sample is directly assigned to the discretized level
closest to its original y-axis value, while x-axis value remains the same (see Fig. 2). The only
parameter of this method is δ, representing the discretization step of the y-axis and determining
the compression ratio. In the next step, all samples but the first one sharing the same alphabet
symbol are excluded, until the change of symbol occurs. The output of the compression is the
sequence of all remaining samples and their time stamp (x-axis value).
   The authors proposed this method to maximize the ability to adapt to the dynamic changes
in the data. The ASAX method, compared to the original SAX, allows for more dynamic
system reactions in case a significant change occurs (e.g. heart attack), rather than waiting
for the entire block of samples to be measured, averaged, assigned and only then sent. On the
other hand, this method does not compress the x-axis equidistantly, causing the compression
ratio to be highly dependent on the data, as well as the chosen parameter delta. Therefore,
the compression ratio is greater for data with slower changes compared to SAX method.

2.4. Lightweight Temporal Compression (LTC)
Lightweight Temporal Compression is based on a piece-wise linear approximation. Its main
advantages are high compression rate capabilities and fast processing speed, i.e., low demands
on CPU and memory. Resembling RLE which compresses long strings of repeating symbols
by a single symbol, the LTC compresses the data based on the linear trend that occurs in
them [12].
  The LTC algorithm is simple to implement and the basic principle is captured in Fig. 3,
where the black dots represent the input samples in the time-series, Zi represents the current
     y                                               y
                                    U1


                                         U2                                      U2

                                         L1=L2                           Z2
                                                                                 L2

         Z1                                                                            U3


                                                                                      L3
           1         2        3          4       x       1    2         3        4          x
Figure 3: Illustration of LTC implementation.


output sample, Ui and Li represent the upper and lower bound of the compression, respectively.
It also has a single adjustable parameter called δ (red dashed lines), representing the tolerance
of the method to the changes within data. Sample 1 (marked as Z1 ) is the first symbol of input
data. Connecting Z1 with the extremes of the δ range of the following sample (coordinates
[x2 , y2 +δ] and [x2 , y2 -δ]) creates the upper and lower bounds for the compression, called U1
and L1 respectively.
   In case the δ-range (between [x3 , y3 +δ] and [x3 , y3 -δ]) of the following (third) sample in-
tersects with the area in between U1 and L1 (see Fig. 3, left), a new set of bounds is created
(U2 and L2 ). In case the upper point ([x3 , y3 +δ]) lies below U1 , U2 is created by connecting
Z1 to the upper point. In case the lower point ([x3 , y3 -δ]) lies above the L1 , L2 is created by
connecting Z1 to the lower point of δ-range.
   In case the δ-range of the following sample lies outside of the area between two bounds (see
Fig. 3, right), a new initial point (Z2 ) is set. It is placed in the middle of the current bounds
(U2 and L2 ) at the last valid δ-range line (see Eq. (2)). A new set of bounds (U3 and L3 ) are
created by connecting Z2 to the upper and lower point of the current sample’s δ-range.
                                           [︃                       ]︃
                                                U2 (x3 ) + L2 (x3 )
                                      Z2 = x3 ,                                                (2)
                                                         2
   The compression is completed by repeating the process. The output of the compression is
the sequence of all Z-point coordinates.
   The algorithmic expression of how the method is implemented can be found in [12]. LTC
was already evaluated to be feasible for wireless sensor networks. It allows to sample the input
data at a high rate, it requires low data storage and it assists with noise suppression [13]. The
method is extendable into multi-dimensional space [14]. One of possible alternatives to the
method is its altered version called Refined Lightweight Temporal Compression (RLTC). Its
compression ratio and energy requirements are improved in comparison to the traditional LTC,
while its memory usage and latency overhead are bigger [15], which is why we do not utilize it
in this paper.
                      105




                      104
          Count [-]




                      103




                      102




                      101

                            0   10      20              30          40   50       60
                                             Data granularity [s]

Figure 4: The granularity of heart rate measurements.


3. Dataset description for methods evaluation
The dataset titled An Open Dataset for Human Activity Analysis using Smart Devices, de-
scribed in [16] was used for the evaluation of compression methods. The dataset was pre-
viously utilized in [17] and [18]. The whole available collection of three datasets consists of
measurements by three wearable devices (phone, watch and glasses) of one subject, taken over
15 consecutive days (from morning to evening). Only the data measured by the smartwatch
are considered, including only its timestamps and heart rate values (beats per minute). The
LG Watch Urbane 2 type of smartwatch was used. The data consists of 91337 heat rate mea-
surements and is further explained in [16]. The granularity of these measurements may be seen
in Fig. 4. The dataset was chosen due to the high frequency of measurements, large number
of entries and availability of the dataset. We remark that an initial comparison of various
wearable datasets was done by the authors in [19].
  The selected dataset was pre-processed by removing all zero values, not-a-number values,
etc. (total of 9 values). Due to the high frequency of measurements, the heart rate values
of two neighboring samples are not changing rapidly, resulting into slow changes and high
correlation in the data as well as temporally linear segments in the course of the data.
4. Comparison of methods based on the selected wearable dataset
4.1. Performance metrics
To compare the compression performance, computational complexity and long-term storage
suitability of the compression methods, we utilize three basic metrics. The first one of them,
Root Mean Squared Error (RMSE), represents the degree of the distortion of the original data
compared to the reconstructed data (see Eq. (3), where N stands for number of samples in
each vector, sn is nth sample of original data vector and rn is nth sample of the reconstructed
vector). To represent the compression distortion more universally, the resulting error was
converted to % values by dividing it with y-axis range of the original data max(s) − min(s)
and multiplying by 100.
                                  ⌜
                                  ⃓
                                  ⃓ 1 ∑︂ N
                                                              100
                        RM SE =   ⎷        [sn − rn ]2 ·                                    (3)
                                     N                   max(s) − min(s)
                                         n=1

 Next, the Compression Ratio (CR) (see Eq. (4)) reflects the degree of compression of the
method, represented by number of samples in equivalent data format.

                                           size(original data)
                                 CR =                                                           (4)
                                         size(compressed data)
  Finally, Compression Time (CT) is employed to evaluate the computational complexity of
each method, represented by time required to compress the single data sample during the full
dataset compression. Big-O notation is in this case non-explanatory since the compression of
time-series data is preferably executed per-sample (or per-sample batch).

4.2. Performance comparison of the methods
In the following paragraphs, the methods are compared and evaluated based on the chosen
performance metrics. The comparison was generated under the assumption of transmitting
data approximately every 15 minutes, resulting in 100 samples per DCT-RLE block, while
measuring the data every 8-10 seconds. Regarding the compression efficiency, the LTC gives
the best results when comparing RMSE to the CR, as shown in Fig. 5, with 1.9 % RMSE at a
compression ratio of 10. The SAX performs only slightly worse than LTC with 2.2 % RMSE
at a CR equal to 10, followed by ASAX with 2.9 % RMSE at the same CR = 10. DCT-RLE
falls significantly behind the other methods with 21.6 % RMSE at CR = 10.
   The evaluation of computational complexity of the chosen methods is shown in Fig. 6, where
SAX performed the fastest for all CRs higher than 3, followed by ASAX. LTC and DCT-RLE
perform approximately 4 times slower than ASAX. Considering the above, LTC is the method
that trades slightly better compression results for higher computational complexity. ASAX
performs the fastest at the low compression values, overtaken by SAX at the higher compression
ratios. Both SAX and ASAX perform with moderate compression error compared to the two
remaining techniques, showing the promising trade-off between CT and RMSE. DCT-RLE
shows the poorest performance among the studied methods with regard to RMSE as well as
computational complexity.
   DCT, from its definition, is suitable for data with periodic changes as it is based on frequency-
domain transformation. In case the measurements are not taken in time spans of seconds,
                             RMSE Based on Compression Ratio
                5
                                                                            SAX
               4.5                                                          DCT-RLE
                                                                            LTC
                4                                                           ASAX

               3.5

                3
    RMSE [%]




               2.5

                2

               1.5

                1

               0.5

                0
                     0   2            4            6            8         10          12
                                       Compression Ratio [-]

Figure 5: Comparison of RMSE to Compression Ratio of Different Methods


but rather minutes, the changes between neighboring samples start to vary too unexpectedly
(e.g., the patient sleeps during one measurement and chases the bus during the next one),
which would result in high dynamic error. Therefore, the precision of the decompression
gets lowered, causing possible misinterpretation of the data. This case can be considered an
advantage in some use cases of DCT, where it is able to filter out high frequencies (e.g. noise
in measurements).
   From the definition of the methods, in case of sudden high-frequency changes, SAX, similarly
to the DCT, trims the extremes by averaging over the given set of samples (PAA part of the
algorithm). ASAX however reacts instantaneously to the peak, immediately reporting the
new alphabet value. LTC reacts to the sudden changes, but with certain delay as it always
reports the last Z-value instead of reacting to the current sample. In case the wearable health
application run on the sensor device is aiming to detect sudden changes (events such as cardiac
arrest), the most suitable method of the above-studied ones is ASAX, since it is the one able
to follow fast the rapid changes.

4.3. Transmission Frequency and Power Consumption
DCT-RLE requires the whole block of data to be compressed at once and therefore it is
limited in terms of block size (or step). As the high frequency components are usually weaker,
                                     Compression Time Based on Compression Ratio

                                                                               SAX
                                                                               DCT-RLE
                           0.5                                                 LTC
                                                                               ASAX
    Compression time [s]




                           0.4



                           0.3



                           0.2



                           0.1



                            0
                                 0        5          10           15          20
                                                 Compression Ratio [-]

Figure 6: Comparison of Computation Time to Compression Ratio of Different Methods


DCT-RLE is efficient if applied to larger blocks of data, as shown below. Furthermore, the
combination of DCT with RLE further improves the efficiency of large block size approach,
as it compresses more zero-values into only one symbol. Fig. 7 visualizes the performance of
DCT-RLE, based on the varying block size. It is clearly visible that the small block size limits
the method, while compressing blocks of 5000 samples result in 3 % RMSE while the CR is
40, making the method suitable for long-term data storage applications.
   Although the DCT-RLE utilization of bigger block sizes results in higher CR while having
smaller RMSE, DCT-RLE is not suitable for applications with high frequency of data trans-
mission. For the measurements taken every 8 to 10 seconds, approximately 100 samples are
produced every 15 minutes corresponding to the DCT block size equal to 100. This adds up
to 9600 measurements produced per each day. Therefore, DCT may be applied in wearable
devices that transmit the measurement once per day, but the utilization in low latency systems
is not suitable. In case of low latency systems, ASAX is the most suitable presented method
for wearables, as it reports the changes immediately, as explained above.
   To evaluate the dependency of various DCT block sizes on CT, first 20000 samples from the
evaluated dataset were utilized. Fig. 8 shows, that with the higher block size, the efficiency
also increases with regard to the computational complexity, when evaluating the CT based
on the CR. The CT of block size of 50 in comparison to the CT of block size 5000 is almost
                             RMSE of DCT based on different step size
               50
                                                                        Step = 50
               45                                                       Step = 100
                                                                        Step = 500
               40                                                       Step = 1000
                                                                        Step = 2000
               35                                                       Step = 5000
                                                                        Step = 20000
               30
    RMSE [%]




               25

               20

               15

               10

               5

               0
                    0   20    40    60     80    100    120      140   160   180       200
                                         Compression Ratio [-]

Figure 7: Comparison of RMSE to Compression Ratio of DCT


ten times higher. Lines referring to smaller step sizes are limited in CR, as high CRs are
impossible to obtain with limited block size. This result supports the claim, that DCT-RLE is
highly efficient in case the latency of the system is not an issue. Unlike DCT-RLE, the other
three studied algorithms, ASAX, SAX, and LTC are not highly dependent on the block size,
as they are compressing in more dynamic fashion.
   An additional point to consider when dealing with wearable devices and their power effi-
ciency is the power consumption of each connection and transmission of the data to a gateway
node. Frequent reporting leads to larger power consumption overhead due to communication
initiation cost.

4.4. Long-term Storage
On one hand, for the immediate diagnosis (e.g. in case of heart attack), single value measure-
ments taken at high frequency are important. These are also essential during activity tracking,
especially for high-performance exercise evaluation. On the other hand, the data stored for a
long time (e.g. years) may not have such information gain compared to the amount of data
stored in the short-term evaluation. For the purposes of long-term patient tracking, approxi-
mate heart rate measures may be sufficient.
  This way, the input alphabet size for the compression could be decreased and therefore the
                                      Compression time based on compression ratio for DCT
                           0.15
                                                                                      Step = 50
                                                                                      Step = 100
                                                                                      Step = 500
                                                                                      Step = 1000
                                                                                      Step = 2000
                                                                                      Step = 5000
    Compression Time [s]




                            0.1                                                       Step = 20000




                           0.05




                             0
                                  0    5     10    15     20     25    30       35   40    45        50
                                                        Compression Ratio [-]

Figure 8: Comparison of Computation Time to Compression Ratio of DCT


data size would be significantly reduced. For this purpose, a two-stage approach should be
utilized. The lossless compression, such as RLE, Huffman Coding (HC) or Lempel-Ziv-Welch
(LZW), should follow the lossy one. Therefore, another possible application of DCT-RLE with
big block size based on our studies is in the long-term storage. SAX can also be an efficient
technique for long-term storage, as the fixed time-step and limited alphabet size allow for
highly efficient bit-level compression in both dimensions. LTC and ASAX are rather lacking in
terms of data storage, as both methods and their compression efficiency are highly dependent
on the dynamics of the data they are applied to.

4.5. Results summary
The summary of the suitability of different methods for the chosen applications can be found
in Table 1. DCT-RLE is considered twice there, once with ”Small Step”, referring to block
sizes smaller than 100 for the applied dataset and ”Big Step”, referring to the block sizes of
2000 and more. The table gives a recommendation regarding the suitability of each method
for different use-cases considered in this paper. The suitability ranges from ”Suitable” to ”Not
Suitable”, with ”Neutral” referring to the cases, where the better option is available, yet the
performance is sufficient.
Table 1
The results summary

                                                  DCT-RLE
                                                  Small Big SAX      LTC   ASAX
                                                  Step Step
               Compression performance            ✗      ✓       -   ✓     -
               Computation complexity             ✗      ✓       ✓   ✗     -
               Long-term storage                  ✗      ✓       ✓   -     -
               Low-latency systems                -      ✗       -   -     ✓

                                  ✓Suitable ✗Not suitable -Neutral


5. Conclusions
The power consumption efficiency, as one of the main challenges of wearable-based devices,
can be improved by applying lossy compression techniques to the wearable sensor-based data.
This paper evaluates four different compression methods with regard to their applicability in
such devices on heart rate data, as most commonly measured biometric data from wearables.
The analysis shows that LTC performs the best with regard to the compression error, at
the cost of a high computational complexity. SAX performs the fastest, when compression
ratio is above 3, surpassed by ASAX at CR below 3. Both methods perform moderately
regarding compression distortion metric (RMSE per compression ratio of the method). The
proposed ASAX compression technique offers the best tradeoff between different performance
metrics if deployed in a system requiring short latency. This is because ASAX reacts to the
significant changes within the data without delay, unlike the other studied methods where the
reaction times are higher. The performance of DCT-RLE is strongly dependent on the block
size, determining the length of the vector fed to the algorithm. With shorter block size, e.g.,
below 1000 samples, DCT-RLE performs significantly worse than the remaining techniques.
On the other hand, with block size 10000 or higher, the DCT-RLE is able to achieve very
high compression ratios with low RMSE, making the DCT-RLE technique suitable for systems
tolerating very high latency and for long-term storage of data. The ability of SAX to compress
data in two dimensions makes it suitable for long-term compression as well.


Acknowledgments
The authors gratefully acknowledge funding from European Union’s Horizon 2020 Research
and Innovation programme under the Marie Sklodowska Curie grant agreement No. 813278
(A-WEAR, http://www.a-wear.eu/). Also, the authors would like to thank to Roman Klus
for his help and valuable impacts during the preparation of this paper.


References
 [1] E. Park, User acceptance of smart wearable devices: An expectation-confirmation model
     approach, Telematics and Informatics 47 (2020) 101318.
 [2] K. Lauber, G. Wort, F. Gillison, D. Thompson, D. Augustine, O. Peacock, Patient views
     of the role of wearable technology to enhance cardiac rehabilitation, in: United Kingdom
     Society for Behavioural Medicine annual conference, 2020.
 [3] F. El-Amrawy, M. I. Nounou, Are currently available wearable devices for activity track-
     ing and heart rate monitoring accurate, precise, and medically beneficial?, Healthcare
     informatics research 21 (2015) 315–320.
 [4] J. Talvitie, M. Renfors, E. S. Lohan, Novel indoor positioning mechanism via spectral
     compression, IEEE Communications Letters 20 (2015) 352–355.
 [5] C. Zhang, J. Li, S. Wang, Z. Wang, An encrypted medical image retrieval algorithm based
     on dwt-dct frequency domain, in: 2017 IEEE 15th International Conference on Software
     Engineering Research, Management and Applications, 2017.
 [6] R. Thanki, S. Borra, V. Dwivedi, K. Borisagar, An efficient medical image watermarking
     scheme based on fdcut–dct, Engineering science and technology, an international journal
     20 (2017) 1366–1379.
 [7] A. Mehto, N. Mehra, Adaptive lossless medical image watermarking algorithm based on
     dct & dwt, Procedia Computer Science 78 (2016) 88–94.
 [8] A. Raid, W. Khedr, M. A. El-Dosuky, W. Ahmed, Jpeg image compression using discrete
     cosine transform-a survey, arXiv preprint arXiv:1405.6147 (2014).
 [9] D. Del Testa, M. Rossi, Lightweight lossy compression of biometric patterns via denoising
     autoencoders, IEEE Signal Processing Letters 22 (2015) 2304–2308.
[10] J. Lin, E. Keogh, L. Wei, S. Lonardi, Experiencing sax: a novel symbolic representation
     of time series, Data Mining and knowledge discovery 15 (2007) 107–144.
[11] E. Keogh, K. Chakrabarti, M. Pazzani, S. Mehrotra, Locally adaptive dimensionality
     reduction for indexing large time series databases, in: Proc. of the 2001 ACM SIGMOD
     international conference on Management of data, 2001.
[12] T. Schoellhammer, B. Greenstein, E. Osterweil, M. Wimbrow, D. Estrin, Lightweight
     temporal compression of microclimate datasets (2004).
[13] R. Sharma, A data compression application for wireless sensor networks using ltc algo-
     rithm, in: IEEE International Conference on Electro/Information Technology, 2015.
[14] B. Li, O. Sarbishei, H. Nourani, T. Glatard, A multi-dimensional extension of the
     lightweight temporal compression method, in: 2018 IEEE International Conference on
     Big Data (Big Data), IEEE, 2018, pp. 2918–2923.
[15] O. Sarbishei, Refined lightweight temporal compression for energy-efficient sensor data
     streaming, in: 2019 IEEE 5th World Forum on Internet of Things, IEEE, 2019.
[16] S. Faye, N. Louveton, S. Jafarnejad, R. Kryvchenko, T. Engel, An open dataset for human
     activity analysis using smart devices (2017).
[17] A. Das Antar, M. Ahmed, M. A. R. Ahad, Challenges in sensor-based human activity
     recognition and a comparative analysis of benchmark datasets: A review, 2019. doi:10.
     1109/ICIEV.2019.8858508.
[18] H. Alrehamy, C. Walker, (personal data lake) semlinker: automating big data integration
     for casual users, Journal of Big Data 5 (2018). doi:10.1186/s40537-018-0123-x.
[19] L. Klus, E. S. Lohan, C. Granell, J. Nurmi, Crowdsourcing solutions for data gathering
     from wearables, in Proc. of XXXV Finnish URSI Convention on Radio Science, 2019.