=Paper=
{{Paper
|id=Vol-3665/short2
|storemode=property
|title=Methods for Predicting Failures in a Smart Home
|pdfUrl=https://ceur-ws.org/Vol-3665/short2.pdf
|volume=Vol-3665
|authors=Viktoriia Zhebka,Pavlo Skladannyi,Yurii Bazak,Andrii Bondarchuk,Kamila Storchak
|dblpUrl=https://dblp.org/rec/conf/decat/ZhebkaSBBS24
}}
==Methods for Predicting Failures in a Smart Home==
Methods for Predicting Failures in a Smart Home
Viktoriia Zhebka1, Pavlo Skladannyi2, Yurii Bazak1, Andrii Bondarchuk1,
and Kamila Storchak1
1 State University of Information and Communication Technologies, 7 Solomenskaya str., Kyiv, 03110, Ukraine
2 Borys Grinchenko Kyiv University, 18/2 Bulvarno-Kudriavska str., Kyiv, 04053, Ukraine
Abstract
Methods for predicting possible failures in smart home systems and analyzing the data
required for this have been considered in the study. A study of machine learning methods
has been carried out: their features, advantages, and disadvantages have been identified,
the metrics of each method have been studied, and the effectiveness of methods for
predicting failures in a smart home has been established. It has been found that the Long
Short-Term Memory (LSTM) model is distinguished by its ability to work with data
sequences and store information for a long time. The characteristics of the LSTM method
and its algorithm have been studied in detail. The study emphasizes the importance of
collecting and processing various data, such as sensor data, energy consumption, and
information about devices and users. The results of the study can be useful for the further
development of smart home control systems to improve their reliability and efficiency.
Keywords 1
Long short-term memory, LSTM, Machine learning, data processing, forecasting, smart
home, failure, information technology.
1. Introduction predicting possible problems and taking
measures to prevent them before they occur.
Smart houses are becoming increasingly Today, smart home failure prediction is
common thanks to the development of the mostly based on reactive data analysis. This
Internet of Things (IoT) and smart technologies means that systems detect anomalous
[1]. They provide automation and ease of situations or failures after they occur, which can
control of various systems such as lighting, make it difficult to avoid potential problems.
heating, security, energy efficiency, and many However, using machine learning methods,
others [2, 3]. such as classification, clustering, and prediction
However, as the complexity of these systems algorithms, it is possible to develop systems that
grows, the likelihood of failures or problems can predict failures in a smart home in advance.
increases. Network instability, software errors, Such systems use data analysis from sensors,
and faulty devices can all lead to unpredictable IoT devices, control systems, energy
situations that affect the usability and security consumption, and other data to identify
of a smart home [4]. patterns and anomalies that may precede
Therefore, predicting failures in a smart disruptions [5, 6]. Based on this information,
home is a relevant issue, and it is appropriate to machine learning systems can build predictive
conduct such a prediction using machine models that respond to certain signals or
learning methods. The use of machine learning changes in normal operation, warning of
algorithms allows for analyzing large amounts potential problems or taking steps to prevent
of data; and identifying deviations and patterns them [7, 8].
that precede failures. This approach allows This area of research is still evolving, but it
promises to improve smart home control
DECaT’2024: Digital Economy Concepts and Technologies, April 4, 2024, Kyiv, Ukraine
EMAIL: viktoria_zhebka@ukr.net (V. Zhebka); p.skladannyi@kubg.edu.ua (P. Skladannyi); jura.bazak@gmail.com (Y. Bazak);
dekan.it@ukr.net (A. Bondarchuk); kpstorchak@ukr.net (K. Storchak)
ORCID: 0000-0003-4051-1190 (V. Zhebka); 0000-0002-7775-6039 (P. Skladannyi); 0009-0000-6098-2809 (Y. Bazak); 0000-0001-5124-
5102 (A. Bondarchuk); 0000-0001-9295-4685 (K. Storchak)
©️ 2024 Copyright for this paper by its authors.
Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0).
CEUR Workshop Proceedings (CEUR-WS.org)
CEUR
ceur-ws.org
Workshop ISSN 1613-0073
Proceedings
70
systems by enabling them to predict and x − min
prevent possible failures in advance, providing xn = (1)
max − min .
greater reliability and security for users.
x−
xn = (2)
2. Research Results
2 x 2 x
xsin = sin , xcos = cos (3)
Machine learning techniques can help detect N N
and avoid some disruptions before they occur where х is the feature value, min is the
or restore system operations faster after a minimum value of the feature, max is the
failure. They can provide early detection of maximum value of the feature, µ is the average
anomalies in performance, which will prevent value of the feature, σ is the standard deviation
problems from occurring or respond quickly to of a feature, and N is the total number of
them, which in turn will help reduce the impact different values of the features.
of these failures on the smart home. Some of the algorithms may have problems
Machine learning algorithms are compared with dimensionality and run much slower than
on four different data representations: original, other algorithms. The problem can be solved
balanced, normalized, and standardized. The by reducing the number of dimensions using
original data is unchanged from the selected principal component analysis. The method is
data except for the removal of timestamp only applicable to specific algorithms and
values. Balancing is performed by under- specific data representations, depending on
sampling the faultless data in the training data. the speed of learning and prediction. In
Failure-free data inputs are randomly selected addition, some of the algorithms work only
and removed from the dataset, resulting in the with a certain input format.
same number of failures and no failures. Many machine learning algorithms can be
Insufficient sampling can erase important used to predict device failures. They can differ
information from the data, leading to poor in many properties and features [9]. They can
algorithm performance. The main advantage of be supervised or unsupervised, they can solve
data balancing is that it reduces the resources classification, regression, or clustering
and time required to train algorithms. Data problems or they can belong to different
normalization refers to the scaling of feature families such as deep learning, tree-based,
values in the range from zero to one. Scaling is probabilistic, or linear. The total number of
performed using (1) separately for each feature algorithms considered for comparison was
by finding its minimum and maximum values. limited due to their lengthy setup and training.
The data is also standardized by considering The algorithms were selected based on several
each feature separately. The feature values are criteria:
subtracted from the mean and then divided by 1. Supervised learning: based on the
the standard deviation, as shown in (2). These assumption that the data to be used have
results in values centered around zero with unit been chosen.
dispersion. An additional pre-processing step 2. Practical use: some of the algorithms are
that is performed before normalizing and used more than others for predictive
standardizing the data is the conversion of the maintenance.
time features of the day of the week and the 3. Diversity: algorithms were chosen to
hour. To indicate that the difference between represent different families, tasks, and
the hours 23 and 0 is the same as the difference functions, such as online learning or
between 22 and 23, the values are converted to prediction over time.
cyclic representations using the Fourier Based on these criteria, ten algorithms have
transform [6]. The transformation calculates been selected, nine of which are implemented
the sine and cosine values for each feature, as as classification algorithms and one as a time
shown in equation (3). Thus, each feature is series regression algorithm (Tables 1 and 2)
replaced by the corresponding sine and cosine [10, 11]. There are representatives of different
feature. The calculation depends on the total types. All algorithms support online learning
number of different feature values N, which is either implicitly or through certain variations.
24 for hours and 7 for days of the week.
71
Table 1
List of machine learning methods and their brief description
English name Description
k-Nearest Neighbor A classification method that determines the class of an object by analyzing its nearest neighbors in the
feature space [12].
Decision Tree A classification or regression algorithm that uses a tree structure to make decisions based on feature
separations [13].
Random Forest Decision tree ensemble, where multiple trees are combined to avoid overlearning and improve accuracy [14].
Extreme Gradient Boosting An ensemble of models that use gradient lift to improve the accuracy of each subsequent model [15].
Naive Bayes A probabilistic method that uses Bayes’ theorem for classification based on the probabilities of features
entering a class [16].
Support Vector Machine An algorithm that determines the optimal boundary of separation between classes using support vectors
[17].
Logistic Regression A classification method that uses a logistic function to determine the probability of an object belonging to a
certain class [12].
Stochastic Gradient Descent An optimization algorithm that uses a gradient to find the minimum of a loss function with a randomly
selected subset of data [16].
Multi-Layer Perceptron A neural network with one or more hidden layers is used for classification and regression based on
weighting coefficients [12].
LSTM A type of recurrent neural network designed to store and use information over a long period to predict
failures or events [16, 17].
Table 2
Comparative characteristics of machine learning methods
Method The principle of operation Application area Advantages Disadvantages
k-Nearest Determining the class of an Detecting anomalies, Easy to implement, no Sensitivity to emissions,
Neighbor object through its nearest predicting failures training required high computational
neighbors costs
Decision Tree Decision-making based on Anomaly detection, failure Ease of interpretation, Tendency to overlearn,
sequential divisions by features classification accommodating conditions instability
Random Forest Tree ensemble to avoid Detecting anomalies, High accuracy, and A large number of
overtraining predicting failures consistency of solutions hyperparameters,
training time
Extreme Gradient Using gradient lift to improve Failure prediction, High accuracy, less A large number of
Boosting accuracy anomaly detection tendency to overlearn hyperparameters,
complexity of
interpretation
Naive Bayes Using Bayes’ theorem for Filtering anomalies, Efficiency for small data, The predictions are not
probabilistic classification detecting failure patterns simplicity of the model flexible enough
Support Vector Determining the optimal Classification of Efficiency in large spaces, Requirements for data
Machine boundary of separation between anomalies, forecasting flexibility preparation, high
classes failures complexity of
customization
Logistic Determining the probability of an Failure classification, Interpretability, ease of Requires a linear
Regression object belonging to a certain class anomaly detection implementation separation surface
Stochastic Using a gradient to optimize the Model training, failure Fast learning, efficient for Requirements for
Gradient Descent loss function analysis big data hyperparameters,
tendency to stutter
Multi-Layer A neural network with one or Pattern recognition, time Ability to solve complex Requires a lot of data for
Perceptron more hidden layers series forecasting problems training, training time
LSTM Recurrent neural network for Time sequence analysis, Ability to recognize High computational costs,
long-term memorization failure prediction dependencies over time the complexity of setup
The main metrics for evaluating different Accuracy represents the percentage of
machine learning algorithms in prediction or correctly classified cases in the total number of
classification tasks allow for an objective cases, which gives a general idea of the
comparison of the effectiveness of these algorithm’s accuracy [11]:
algorithms.
72
Т = (TP + TN) / (FP + FN + TP + TN), (4) method are usually used, since the ROC-AUC
formula itself is an integral.
where TP is true positive (correctly categorized
The completeness (R) and frequency (S) are
positive ones), TN is true negative (correctly
determined using a confusion matrix for binary
categorized negative ones), FP is false positive
classification.
(incorrectly categorized positive ones), and FN is
Indicator S is calculated using formula (6),
false negative (incorrectly categorized negative
and the frequency is calculated using formula (9).
ones).
The precision determines the percentage of S=
FP
(9)
correctly identified positive classes among all FP + TN .
identified positive classes, which is useful for The Confusion Matrix provides detailed
working with uneven classes. information about the real and predicted classes,
P=
TP
(5) which helps to estimate the level of correctness
FP + TP . and errors for each class, which is important
Recall displays the percentage of correctly when analyzing the model.
identified positive classes among all actual The Confusion Matrix helps to evaluate the
positive classes, which is important for performance of a classification model by
identifying important cases that have been visualizing real and predicted values. It is the
missed. basis for calculating various metrics, such as
TP accuracy, sensitivity, specificity, F1 score, and
R= (6) others.
(TP + FN )
. These metrics are crucial for evaluating the
F1-average is a score that uses the harmonic effectiveness of algorithms and choosing the one
mean between accuracy and completeness to that is most suitable for a particular task,
understand how well the model solves the depending on the requirements and needs [18,
classification task. 19].
𝑃𝑅 Table 3 shows the performance of different
𝐹1 = 2 . (7) machine learning methods by the main metrics
𝑃+𝑅
considered.
ROC-AUC measures the area under the ROC
Table 3 and Fig. 1 show the performance of
curve and evaluates the model’s performance
different machine learning methods in terms of
depending on different classification thresholds,
the main metrics such as precision, accuracy,
helping to determine its ability to make correct
classification accuracy, completeness, F1-mean,
predictions.
1
and ROC-AUC. The score of “High,” “Medium”, or
ROC − AUC = R( S ) d(S) “Very High” in the “Performance” column is a
(8)
0 . generalized characterization of the methods’
performance based on these metrics.
To approximate this area, numerical methods
such as the trapezoidal method or Simpson’s
Table 3
Effectiveness of different machine learning methods by key metrics
Classification F1- ROC-
Method Accuracy Completeness Effectiveness
accuracy average AUC
k-Nearest Neighbor 0.85 0.81 0.89 0.85 0.92 High
Decision Tree 0.78 0.82 0.75 0.76 0.85 High
Random Forest 0.81 0.85 0.79 0.80 0.88 High
Extreme Gradient Boosting 0.87 0.88 0.86 0.87 0.94 High
Naive Bayes 0.75 0.79 0.72 0.73 0.82 Average
Support Vector Machine 0.82 0.84 0.80 0.81 0.89 High
Logistic Regression 0.79 0.83 0.77 0.78 0.86 Average
Stochastic Gradient Descent 0.80 0.82 0.79 0.80 0.87 Average
Multi-Layer Perceptron 0.84 0.86 0.82 0.83 0.91 High
LSTM 0.88 0.90 0.87 0.88 0.95 Very high
73
Figure 1: Comparison of machine learning methods
The study has found that the LSTM model is 5. New memory state Ct
distinguished by its ability to work with data 6. New exit ht.
sequences and store information for a long The model itself consists of the following
time. This makes LSTM effective for analyzing formulas:
time series, such as sensor data in a smart 1. Forgetting the previous memory state:
home, where information is usually sequential
in time. . (10)
The LSTM model is capable of storing 2. Defining what will be updated in memory:
information for a long time, allowing it to
effectively understand and analyze a sequence . (11)
of real-time sensor data. By using mechanisms
that ensure that some information is forgotten (12)
.
and others are retained, the LSTM can take into
account long-term dependencies and the 3. Update the memory status:
importance of individual events in time series. C = f *C + it * C t (13)
LSTM can adapt to and learn from different
t t t- 1
.
amounts of data, including large amounts of 4. Update the output value:
data from smart home sensors, which allows for
more accurate predictions of failures. The LSTM . (14)
model can adapt to changing conditions and
detect changes in time series, which allows for h = o * tg (C ).
t t t (15)
predicting failures and anomalies in real-time.
where ft is the forget gate, which decides that
LSTM can process a variety of data types (text,
the previous state should be forgotten, it is the
numbers, sequences, etc.), making it versatile
input gate, which determines how much of the
for use in various forecasting and analysis
new input will be added to the memory state,
scenarios.
𝐶̃𝑡 is a new candidate for the memory state, Ct
That is why it is not surprising that this
is memory status, ot is output gate, determines
algorithm showed the best results for
which output will be next, xt is input on a time
predicting failures in a smart home.
step t, ht−1 is preliminary output on the time
The LSTM model has the following elements
step t−1, W and а b is weights and
at each time step t:
displacements to be taught during the training
1. Input data xt
process.
2. Previous output ht−1
These equations allow the LSTM to
3. Previous memory state Ct−1
determine what information to forget, what to
4. Gates:
keep, and how to use it to generate output
• Forget gate ft
values.
• Input gate it
• Output gate ot.
74
The step-by-step algorithm for training an This is a general description of the LSTM
LSTM model includes the following steps: learning algorithm for predicting failures in a
1. Data preparation: smart home.
• Input: receive a set of data containing time The approach presented in this study takes
series or sequences. advantage of LSTM to predict time series. The
• Data preparation: normalization, and LSTM is implemented using Keras (a high-level
conversion of data format to meet model neural network interface that simplifies the
requirements. process of creating and training artificial neural
2. Building an LSTM architecture: networks; it is a machine learning library that
• Creating a model: using machine learning runs on the Tensorflow, Theano, and Microsoft
libraries to build an LSTM network. Cognitive Toolkit frameworks) as a sequential
• Defining parameters: number of layers, model with two LSTM layers and a dense output
number of neurons in each layer, layer. It receives a sequence of data inputs
activation functions, etc. (normalized feature values without rejections)
3. Data separation: and outputs a sequence of failure values.
• Training and testing sets: splitting data It was decided to use a single data input, as
into training and testing sets to evaluate this significantly reduces the training time and is
model performance. sufficient for the algorithm to recognize failure
4. Model training: patterns. The length of the source sequence
• Model training: fitting LSTM to training determines the duration of the runtime.
data using backpropagation. Prediction performance is tested on three
• Model evaluation: assessing whether the different input sequences: 1 (1 second), 300 (5
model’s predictions match the actual data minutes), and 1800 (30 minutes). As a result,
on the test set. three different LSTM models were built. For
5. Evaluation and improvement of the model: training, we used a dataset with 70% failures,
• Analyzing the results: reviewing the creating a split into training and test sets in the
forecast results and assessing their proportion of 25–75% without mixing. In
accuracy. addition, the data was prepared by creating
• Improving the model: use optimization output sequences for each data record, which
techniques, change hyperparameters, or were then used for training.
modify the architecture to improve Based on the conducted research, a data
results. prediction platform has been developed. Once
6. Testing and forecasting: the system has been successfully integrated
• Failure prediction: applying a trained into the smart home, the implementation
model to new data to predict failures in a process takes place, when the system becomes
smart home. an active part of the home environment.
7. Evaluation of the results: However, this is only the beginning: further
• Performance evaluation: comparing support, optimization, and continuous
forecasts with real data to determine the improvement of the system play a key role in
accuracy and efficiency of the model. ensuring its long-term and efficient operation
8. Maintaining and improving the model: in a smart home, adapting to changing needs
• Continuous learning: collecting new data and conditions (Fig. 2).
and improving the model based on it to
keep forecasts up-to-date.
75
100%
90%
80%
70%
60%
50%
40%
30%
20%
10%
0%
Energy efficiency Automation Security Comfort Resource Responding to
management forecasting and changes
optimization
System without machine learning Machine learning system
Figure 2: Smart home system performance with and without machine learning methods
3. Discussion
Machine learning helps to avoid certain
problems by analyzing previous data and
recognizing patterns, but it cannot predict
absolutely all possible scenarios, especially if
they arise from certain unpredictable factors
or third-party interventions.
A smart home system that uses machine
learning methods proves to be better than a
system without this technology (as the study
results show, the performance of a smart home
using failure prediction methods gives an
average of 22% better result compared to a
similar system without prediction—Fig. 3).
Machine learning allows the system to adapt to
changes in the environment and user
requirements, respond more quickly to new
conditions, and optimize resource use. This
helps to improve the system’s efficiency in
managing energy, comfort, safety, and user
satisfaction [20].
Machine learning allows the system to
predict and avoid failures, which ensures Figure 3: The process of integrating an
greater reliability and durability of the system. information system into a smart home
This approach also allows for increased
automation, helping the system perform 4. Conclusions
routine tasks without user intervention [21,
22]. Overall, a machine learning system The study results have shown a wide range of
remains the preferred choice due to its ability modern technologies, sensors, and control
to predict, optimize, and adapt to changes, systems used in smart homes. The overview
enabling it to provide more efficient and has shown that existing technologies have the
convenient smart home management [23, 24]. potential to improve convenience, security,
and energy efficiency.
The analysis of available machine learning
methods indicates their potential in predicting
76
and managing risks in smart homes. The communications, Science and
considered models have shown high accuracy Technology (2023) 522–526. doi:
in predicting failures. 10.1109/PICST57299.2022.10238518.
The following areas can be considered for [7] Smart Home Technology: How AI Creates
further development of this work: a Space that is Comfortable for life. URL:
• Improving machine learning methods to https://www.everest.ua/tehnologiya-
increase the accuracy of failure prediction. rozumnogo-budynku-yak-ai-stvoryuye-
• In-depth study of the impact of the prostirkomfortnyj-dlya-zhyttya/
introduction of machine learning systems [8] T. Arsan, Smart Systems: From Design to
on the functioning of a smart home. Implementation of Embedded SMART
Based on the obtained data, it is possible to Systems (2016). doi: 10.1109/HONET.
build a methodology for predicting failures in a 2016.7753420.
smart home, which will be the direction of the [9] V. Zhebka et al., Optimization of Machine
authors’ next research. Learning Method to Improve the
Management Efficiency of
References Heterogeneous Telecommunication
Network, Cybersecurity Providing in
Information and Telecommunication
[1] Z. Hu, et al., Bandwidth Research of
Systems Vol. 3288 (2022) 149–155.
Wireless IoT Switches, in: IEEE 15th
[10] A. Boiko, Modeling of the Automated
International Conference on Advanced
System of Operational Control of the
Trends in Radioelectronics,
Parameters of the “Smart Home” in the
Telecommunications and Computer
Proteus Environment, Technologies and
Engineering (2020). doi:
Design 2(35) (2020).
10.1109/tcset49122.2020.2354922
[11] V. Zhebka, Y. Bazak, K. Storchak,
[2] I. Kuzminykh, et al., Investigation of the
Features of Predicting Failures in a
IoT Device Lifetime with Secure Data
Smart Home based on Machine Learning
Transmission, Internet of Things, Smart
Methods, Telecommunications and
Spaces, and Next Generation Networks
Information Technologies 4(81) (2023)
and Systems, vol. 11660 (2019) 16–27.
4–12.
doi: 10.1007/978-3-030-30859-9_2
[12] I. Ruban, V. Martovytskyi, S. Partyka,
[3] B. Zhurakovskyi, et al., Secured Remote
Classification of Methods for Detecting
Update Protocol in IoT Data Exchange
Anomalies in Information Systems,
System, in: Workshop on Cybersecurity
Systems of Armament and Military
Providing in Information and
Equipment 3(47) (2016) 47.
Telecommunication Systems, vol. 3421
[13] V. Kharchenko, Fundamentals of
(2023) 67–76
Machine Learning, Sumy State
[4] O. Shevchenko, et al., Methods of the
University (2023).
Objects Identification and Recognition
[14] X. Sun, et al., System-Level Hardware
Research in the Networks with the IoT
Failure Prediction using Deep Learning,
Concept Support, in: Cybersecurity
56th Annual Design Automation
Providing in Information and
Conference 20 (2019) 1–6. doi:
Telecommunication Systems, vol. 2923
10.1145/3316781.3317918.
(2021) 277–282.
[15] V. Malinov, et al., Biomining as an
[5] M. Moshenchenko, et. al., Optimization
Effective Mechanism for Utilizing the
Algorithms of Smart City Wireless
Bioenergy Potential of Processing
Sensor Network Control, in:
Enterprises in the Agricultural Sector, in:
Cybersecurity Providing in Information
Cybersecurity Providing in Information
and Telecommunication Systems II, vol.
and Telecommunication Systems Vol.
3188 (2021) 32–42.
3421 (2023) 223–230.
[6] V. Sokolov, et al., Method for Increasing
[16] Y. Bazak, Comparison of Machine
the Various Sources Data Consistency for
Learning Methods for Predicting
IoT Sensors, in: IEEE 9th International
Failures in a Smart Home, Abstracts of
Conference on Problems of Info-
the Report at the Scientific and Practical
77
Conference “Problems of Computer
Engineering,” Collection of abstracts
(2023) 125–127.
[17] K. Kononova, Machine Learning:
Methods and Models: a Textbook for
Bachelors, Masters and Doctors of
Philosophy in Specialty 051
“Economics,” V. N. Karazin Kharkiv
National University (2020).
[18] Y. Bondarenko, Manual for the Study of
the Discipline “Statistical Analysis of
Data,” Lira (2018).
[19] K. Hureeva, O. Kudin, A. Lisnyak, A
Review of Machine Learning Methods in
the Task of forecasting Financial Time
Series, Computer Science and Applied
Mathematics (2) (2018) 18–28.
[20] I. Puleko, A. Yefimenko, Architecture and
Technologies of the Internet of Things: a
Textbookб State University “Zhytomyr
Polytechnic” (2022).
[21] B. Zhurakovskyi, et al., Coding for
Information Systems Security and
Viability, in: Information Technologies
and Security Vol. 2859 (2021) 71–84.
[22] M. Moshenchenko, et al., Optimization
Algorithms of Smart City Wireless
Sensor Network Control, in:
Cybersecurity Providing in Information
and Telecommunication Systems Vol.
3188 (2021) 32–42.
[23] B. Chen, et al., Smart Factory of Industry
4.0: Key Technologies, Application Case,
and Challenges, IEEE Access 6 (2018)
6505–6519. doi: 10.1109/ACCESS.2017.
2783682.
[24] J. Mineraud, et al., A Gap Analysis of
Internet-of-Things Platforms, Computer
Communications 89–90 (2016) 5–16.
doi: 10.1016/j.comcom.2016.03.015.
78