<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Methods for Predicting Failures in a Smart Home</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Viktoriia</forename><surname>Zhebka</surname></persName>
							<email>viktoria_zhebka@ukr.net</email>
							<affiliation key="aff0">
								<orgName type="institution">State University of Information and Communication Technologies</orgName>
								<address>
									<addrLine>7 Solomenskaya str</addrLine>
									<postCode>03110</postCode>
									<settlement>Kyiv</settlement>
									<country key="UA">Ukraine</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Pavlo</forename><surname>Skladannyi</surname></persName>
							<email>p.skladannyi@kubg.edu.ua</email>
							<affiliation key="aff1">
								<orgName type="institution">Borys Grinchenko Kyiv University</orgName>
								<address>
									<addrLine>18/2 Bulvarno-Kudriavska str</addrLine>
									<postCode>04053</postCode>
									<settlement>Kyiv</settlement>
									<country key="UA">Ukraine</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Yurii</forename><surname>Bazak</surname></persName>
							<email>jura.bazak@gmail.com</email>
							<affiliation key="aff0">
								<orgName type="institution">State University of Information and Communication Technologies</orgName>
								<address>
									<addrLine>7 Solomenskaya str</addrLine>
									<postCode>03110</postCode>
									<settlement>Kyiv</settlement>
									<country key="UA">Ukraine</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Andrii</forename><surname>Bondarchuk</surname></persName>
							<affiliation key="aff0">
								<orgName type="institution">State University of Information and Communication Technologies</orgName>
								<address>
									<addrLine>7 Solomenskaya str</addrLine>
									<postCode>03110</postCode>
									<settlement>Kyiv</settlement>
									<country key="UA">Ukraine</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Kamila</forename><surname>Storchak</surname></persName>
							<email>kpstorchak@ukr.net</email>
							<affiliation key="aff0">
								<orgName type="institution">State University of Information and Communication Technologies</orgName>
								<address>
									<addrLine>7 Solomenskaya str</addrLine>
									<postCode>03110</postCode>
									<settlement>Kyiv</settlement>
									<country key="UA">Ukraine</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Methods for Predicting Failures in a Smart Home</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">D2551776594D36449BAA413D2021925C</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T20:20+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Long short-term memory, LSTM, Machine learning, data processing, forecasting, smart home, failure, information technology (K. Storchak) ORCID: 0000-0003-4051-1190 (V. Zhebka)</term>
					<term>0000-0002-7775-6039 (P. Skladannyi)</term>
					<term>0009-0000-6098-2809 (Y. Bazak)</term>
					<term>0000-0001-5124-5102 (A. Bondarchuk)</term>
					<term>0000-0001-9295-4685 (K. Storchak)</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Methods for predicting possible failures in smart home systems and analyzing the data required for this have been considered in the study. A study of machine learning methods has been carried out: their features, advantages, and disadvantages have been identified, the metrics of each method have been studied, and the effectiveness of methods for predicting failures in a smart home has been established. It has been found that the Long Short-Term Memory (LSTM) model is distinguished by its ability to work with data sequences and store information for a long time. The characteristics of the LSTM method and its algorithm have been studied in detail. The study emphasizes the importance of collecting and processing various data, such as sensor data, energy consumption, and information about devices and users. The results of the study can be useful for the further development of smart home control systems to improve their reliability and efficiency.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>Smart houses are becoming increasingly common thanks to the development of the Internet of Things (IoT) and smart technologies <ref type="bibr">[1]</ref>. They provide automation and ease of control of various systems such as lighting, heating, security, energy efficiency, and many others <ref type="bibr">[2,</ref><ref type="bibr">3]</ref>.</p><p>However, as the complexity of these systems grows, the likelihood of failures or problems increases. Network instability, software errors, and faulty devices can all lead to unpredictable situations that affect the usability and security of a smart home <ref type="bibr">[4]</ref>.</p><p>Therefore, predicting failures in a smart home is a relevant issue, and it is appropriate to conduct such a prediction using machine learning methods. The use of machine learning algorithms allows for analyzing large amounts of data; and identifying deviations and patterns that precede failures. This approach allows predicting possible problems and taking measures to prevent them before they occur.</p><p>Today, smart home failure prediction is mostly based on reactive data analysis. This means that systems detect anomalous situations or failures after they occur, which can make it difficult to avoid potential problems.</p><p>However, using machine learning methods, such as classification, clustering, and prediction algorithms, it is possible to develop systems that can predict failures in a smart home in advance.</p><p>Such systems use data analysis from sensors, IoT devices, control systems, energy consumption, and other data to identify patterns and anomalies that may precede disruptions <ref type="bibr">[5,</ref><ref type="bibr">6]</ref>. Based on this information, machine learning systems can build predictive models that respond to certain signals or changes in normal operation, warning of potential problems or taking steps to prevent them <ref type="bibr">[7,</ref><ref type="bibr">8]</ref>.</p><p>This area of research is still evolving, but it promises to improve smart home control systems by enabling them to predict and prevent possible failures in advance, providing greater reliability and security for users.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Research Results</head><p>Machine learning techniques can help detect and avoid some disruptions before they occur or restore system operations faster after a failure. They can provide early detection of anomalies in performance, which will prevent problems from occurring or respond quickly to them, which in turn will help reduce the impact of these failures on the smart home.</p><p>Machine learning algorithms are compared on four different data representations: original, balanced, normalized, and standardized. The original data is unchanged from the selected data except for the removal of timestamp values. Balancing is performed by undersampling the faultless data in the training data. Failure-free data inputs are randomly selected and removed from the dataset, resulting in the same number of failures and no failures. Insufficient sampling can erase important information from the data, leading to poor algorithm performance. The main advantage of data balancing is that it reduces the resources and time required to train algorithms. Data normalization refers to the scaling of feature values in the range from zero to one. Scaling is performed using (1) separately for each feature by finding its minimum and maximum values. The data is also standardized by considering each feature separately. The feature values are subtracted from the mean and then divided by the standard deviation, as shown in (2). These results in values centered around zero with unit dispersion. An additional pre-processing step that is performed before normalizing and standardizing the data is the conversion of the time features of the day of the week and the hour. To indicate that the difference between the hours 23 and 0 is the same as the difference between 22 and 23, the values are converted to cyclic representations using the Fourier transform <ref type="bibr">[6]</ref>. The transformation calculates the sine and cosine values for each feature, as shown in equation (3). Thus, each feature is replaced by the corresponding sine and cosine feature. The calculation depends on the total number of different feature values N, which is 24 for hours and 7 for days of the week.</p><formula xml:id="formula_0">min max min n x x − = − . (1) n x x   − = (2) sin cos 22 sin , cos xx xx NN      ==         (3)</formula><p>where х is the feature value, min is the minimum value of the feature, max is the maximum value of the feature, µ is the average value of the feature, σ is the standard deviation of a feature, and N is the total number of different values of the features. Some of the algorithms may have problems with dimensionality and run much slower than other algorithms. The problem can be solved by reducing the number of dimensions using principal component analysis. The method is only applicable to specific algorithms and specific data representations, depending on the speed of learning and prediction. In addition, some of the algorithms work only with a certain input format.</p><p>Many machine learning algorithms can be used to predict device failures. They can differ in many properties and features <ref type="bibr">[9]</ref>. They can be supervised or unsupervised, they can solve classification, regression, or clustering problems or they can belong to different families such as deep learning, tree-based, probabilistic, or linear. The total number of algorithms considered for comparison was limited due to their lengthy setup and training. The algorithms were selected based on several criteria:</p><p>1. Supervised learning: based on the assumption that the data to be used have been chosen. 2. Practical use: some of the algorithms are used more than others for predictive maintenance. 3. Diversity: algorithms were chosen to represent different families, tasks, and functions, such as online learning or prediction over time. Based on these criteria, ten algorithms have been selected, nine of which are implemented as classification algorithms and one as a time series regression algorithm (Tables <ref type="table" target="#tab_1">1 and 2</ref>) <ref type="bibr">[10,</ref><ref type="bibr">11]</ref>. There are representatives of different types. All algorithms support online learning either implicitly or through certain variations.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Table 1</head><p>List of machine learning methods and their brief description </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Support Vector Machine</head><p>An algorithm that determines the optimal boundary of separation between classes using support vectors <ref type="bibr" target="#b1">[17]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Logistic Regression</head><p>A classification method that uses a logistic function to determine the probability of an object belonging to a certain class <ref type="bibr">[12]</ref>.</p><p>Stochastic Gradient Descent An optimization algorithm that uses a gradient to find the minimum of a loss function with a randomly selected subset of data <ref type="bibr">[16]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Multi-Layer Perceptron</head><p>A neural network with one or more hidden layers is used for classification and regression based on weighting coefficients <ref type="bibr">[12]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>LSTM</head><p>A type of recurrent neural network designed to store and use information over a long period to predict failures or events <ref type="bibr">[16,</ref><ref type="bibr" target="#b1">17]</ref>.</p><p>Т = (TP + TN) / (FP + FN + TP + TN), (4) where TP is true positive (correctly categorized positive ones), TN is true negative (correctly categorized negative ones), FP is false positive (incorrectly categorized positive ones), and FN is false negative (incorrectly categorized negative ones).</p><p>The precision determines the percentage of correctly identified positive classes among all identified positive classes, which is useful for working with uneven classes.</p><formula xml:id="formula_1">TP P FP TP = + .</formula><p>(5)</p><p>Recall displays the percentage of correctly identified positive classes among all actual positive classes, which is important for identifying important cases that have been missed.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>(</head><p>)</p><formula xml:id="formula_2">TP R TP FN = + . (<label>6</label></formula><formula xml:id="formula_3">)</formula><p>F1-average is a score that uses the harmonic mean between accuracy and completeness to understand how well the model solves the classification task.</p><formula xml:id="formula_4">𝐹1 = 2 𝑃𝑅 𝑃 + 𝑅 .<label>(7)</label></formula><p>ROC-AUC measures the area under the ROC curve and evaluates the model's performance depending on different classification thresholds, helping to determine its ability to make correct predictions.</p><formula xml:id="formula_5">1 0 ( ) d(S) ROC AUC R S − =  . (<label>8</label></formula><formula xml:id="formula_6">)</formula><p>To approximate this area, numerical methods such as the trapezoidal method or Simpson's method are usually since the ROC-AUC formula is an integral.</p><p>The completeness (R) and frequency (S) are determined using a confusion matrix for binary classification.</p><p>Indicator is calculated using formula (6), and the frequency is calculated using formula (9).</p><formula xml:id="formula_7">FP S FP TN = + . (<label>9</label></formula><formula xml:id="formula_8">)</formula><p>The Confusion Matrix provides detailed information about the real and predicted classes, which helps to estimate the level of correctness and errors for each class, which is important when analyzing the model.</p><p>The Confusion Matrix helps to evaluate the performance of a classification model by visualizing real and predicted values. It is the basis for calculating various metrics, such as accuracy, sensitivity, specificity, F1 score, and others.</p><p>These metrics are crucial for evaluating the effectiveness of algorithms and choosing the one that is most suitable for a particular task, depending on the requirements and needs <ref type="bibr" target="#b2">[18,</ref><ref type="bibr" target="#b3">19]</ref>.</p><p>Table <ref type="table" target="#tab_2">3</ref> shows the performance of different machine learning methods by the main metrics considered.</p><p>Table <ref type="table" target="#tab_2">3</ref> and Fig. <ref type="figure" target="#fig_0">1</ref> show the performance of different machine learning methods in terms of the main metrics such as precision, accuracy, classification accuracy, completeness, F1-mean, and ROC-AUC. The score of "High," "Medium", or "Very High" in the "Performance" column is a generalized characterization of the methods' performance based on these metrics.  The study has found that the LSTM model is distinguished by its ability to work with data sequences and store information for a long time. This makes LSTM effective for analyzing time series, such as sensor data in a smart home, where information is usually sequential in time.</p><p>The LSTM model is capable of storing information for a long time, allowing it to effectively understand and analyze a sequence of real-time sensor data. By using mechanisms that ensure that some information is forgotten and others are retained, the LSTM can take into account long-term dependencies and the importance of individual events in time series. LSTM can adapt to and learn from different amounts of data, including large amounts of data from smart home sensors, which allows for more accurate predictions of failures. The LSTM model can adapt to changing conditions and detect changes in time series, which allows for predicting failures and anomalies in real-time. LSTM can process a variety of data types (text, numbers, sequences, etc.), making it versatile for use in various forecasting and analysis scenarios.</p><p>That is why it is not surprising that this algorithm showed the best results for predicting failures in a smart home.</p><p>The LSTM model has the following elements at each time step t:</p><p>1  </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Discussion</head><p>Machine learning helps to avoid certain problems by analyzing previous data and recognizing patterns, but it cannot predict absolutely all possible scenarios, especially if they arise from certain unpredictable factors or third-party interventions.</p><p>A smart home system that uses machine learning methods proves to be better than a system without this technology (as the study results show, the performance of a smart home using failure prediction methods gives an average of 22% better result compared to a similar system without prediction-Fig. <ref type="figure" target="#fig_2">3</ref>). Machine learning allows the system to adapt to changes in the environment and user requirements, respond more quickly to new conditions, and optimize resource use. This helps to improve the system's efficiency in managing energy, comfort, safety, and user satisfaction <ref type="bibr" target="#b4">[20]</ref>.</p><p>Machine learning allows the system to predict and avoid failures, which ensures greater reliability and durability of the system. This approach also allows for increased automation, helping the system perform routine tasks without user intervention <ref type="bibr" target="#b5">[21,</ref><ref type="bibr" target="#b6">22]</ref>. Overall, a machine learning system remains the preferred choice due to its ability to predict, optimize, and adapt to changes, enabling it to provide more efficient and convenient smart home management <ref type="bibr" target="#b7">[23,</ref><ref type="bibr" target="#b8">24]</ref>. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Conclusions</head><p>The study results have shown a wide range of modern technologies, sensors, and control systems used in smart homes. The overview has shown that existing technologies have the potential to improve convenience, security, and energy efficiency.</p><p>The analysis of available machine learning methods indicates their potential in predicting </p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: Comparison of machine learning methods</figDesc><graphic coords="5,72.00,72.00,451.95,187.15" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figure 2 :</head><label>2</label><figDesc>Figure 2: Smart home system performance with and without machine learning methods</figDesc><graphic coords="7,304.80,287.40,155.25,307.93" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>Figure 3 :</head><label>3</label><figDesc>Figure 3: The process of integrating an information system into a smart home</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0"><head></head><label></label><figDesc></figDesc><graphic coords="1,0.00,191.15,594.96,459.74" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 2</head><label>2</label><figDesc>Comparative characteristics of machine learning methods</figDesc><table><row><cell>Method</cell><cell>The principle of operation</cell><cell cols="2">Application area</cell><cell>Advantages</cell><cell>Disadvantages</cell></row><row><cell>k-Nearest</cell><cell>Determining the class of an</cell><cell cols="2">Detecting anomalies,</cell><cell>Easy to implement, no</cell><cell>Sensitivity to emissions,</cell></row><row><cell>Neighbor</cell><cell>object through its nearest</cell><cell cols="2">predicting failures</cell><cell>training required</cell><cell>high computational</cell></row><row><cell></cell><cell>neighbors</cell><cell></cell><cell></cell><cell></cell><cell>costs</cell></row><row><cell>Decision Tree</cell><cell>Decision-making based on</cell><cell cols="2">Anomaly detection, failure</cell><cell>Ease of interpretation,</cell><cell>Tendency to overlearn,</cell></row><row><cell></cell><cell>sequential divisions by features</cell><cell>classification</cell><cell></cell><cell>accommodating conditions</cell><cell>instability</cell></row><row><cell>Random Forest</cell><cell>Tree ensemble to avoid</cell><cell cols="2">Detecting anomalies,</cell><cell>High accuracy, and</cell><cell>A large number of</cell></row><row><cell></cell><cell>overtraining</cell><cell cols="2">predicting failures</cell><cell>consistency of solutions</cell><cell>hyperparameters,</cell></row><row><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell>training time</cell></row><row><cell>Extreme Gradient</cell><cell>Using gradient lift to improve</cell><cell cols="2">Failure prediction,</cell><cell>High accuracy, less</cell><cell>A large number of</cell></row><row><cell>Boosting</cell><cell>accuracy</cell><cell cols="2">anomaly detection</cell><cell>tendency to overlearn</cell><cell>hyperparameters,</cell></row><row><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell>complexity of</cell></row><row><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell>interpretation</cell></row><row><cell>Naive Bayes</cell><cell>Using Bayes' theorem for</cell><cell cols="2">Filtering anomalies,</cell><cell>Efficiency for small data,</cell><cell>The predictions are not</cell></row><row><cell></cell><cell>probabilistic classification</cell><cell cols="2">detecting failure patterns</cell><cell>simplicity of the model</cell><cell>flexible enough</cell></row><row><cell>Support Vector</cell><cell>Determining the optimal</cell><cell>Classification of</cell><cell></cell><cell>Efficiency in large spaces,</cell><cell>Requirements for data</cell></row><row><cell>Machine</cell><cell>boundary of separation between</cell><cell cols="2">anomalies, forecasting</cell><cell>flexibility</cell><cell>preparation, high</cell></row><row><cell></cell><cell>classes</cell><cell>failures</cell><cell></cell><cell></cell><cell>complexity of</cell></row><row><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell>customization</cell></row><row><cell>Logistic</cell><cell>Determining the probability of an</cell><cell cols="2">Failure classification,</cell><cell>Interpretability, ease of</cell><cell>Requires a linear</cell></row><row><cell>Regression</cell><cell>object belonging to a certain class</cell><cell cols="2">anomaly detection</cell><cell>implementation</cell><cell>separation surface</cell></row><row><cell>Stochastic</cell><cell>Using a gradient to optimize the</cell><cell cols="2">Model training, failure</cell><cell>Fast learning, efficient for</cell><cell>Requirements for</cell></row><row><cell>Gradient Descent</cell><cell>loss function</cell><cell>analysis</cell><cell></cell><cell>big data</cell><cell>hyperparameters,</cell></row><row><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell>tendency to stutter</cell></row><row><cell>Multi-Layer</cell><cell>A neural network with one or</cell><cell cols="2">Pattern recognition, time</cell><cell>Ability to solve complex</cell><cell>Requires a lot of data for</cell></row><row><cell>Perceptron</cell><cell>more hidden layers</cell><cell cols="2">series forecasting</cell><cell>problems</cell><cell>training, training time</cell></row><row><cell>LSTM</cell><cell>Recurrent neural network for</cell><cell cols="2">Time sequence analysis,</cell><cell>Ability to recognize</cell><cell>High computational costs,</cell></row><row><cell></cell><cell>long-term memorization</cell><cell cols="2">failure prediction</cell><cell>dependencies over time</cell><cell>the complexity of setup</cell></row><row><cell cols="3">The main metrics for evaluating different</cell><cell cols="3">Accuracy represents the percentage of</cell></row><row><cell cols="3">machine learning algorithms in prediction or</cell><cell cols="3">correctly classified cases in the total number of</cell></row><row><cell cols="3">classification tasks allow for an objective</cell><cell cols="3">cases, which gives a general idea of the</cell></row><row><cell cols="3">comparison of the effectiveness of these</cell><cell cols="2">algorithm's accuracy [11]:</cell></row><row><cell>algorithms.</cell><cell></cell><cell></cell><cell></cell><cell></cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>Table 3</head><label>3</label><figDesc>Effectiveness of different machine learning methods by key metrics</figDesc><table><row><cell>Method</cell><cell>Accuracy</cell><cell>Classification accuracy</cell><cell>Completeness</cell><cell>F1-average</cell><cell>ROC-AUC</cell><cell>Effectiveness</cell></row><row><cell>k-Nearest Neighbor</cell><cell>0.85</cell><cell>0.81</cell><cell>0.89</cell><cell>0.85</cell><cell>0.92</cell><cell>High</cell></row><row><cell>Decision Tree</cell><cell>0.78</cell><cell>0.82</cell><cell>0.75</cell><cell>0.76</cell><cell>0.85</cell><cell>High</cell></row><row><cell>Random Forest</cell><cell>0.81</cell><cell>0.85</cell><cell>0.79</cell><cell>0.80</cell><cell>0.88</cell><cell>High</cell></row><row><cell>Extreme Gradient Boosting</cell><cell>0.87</cell><cell>0.88</cell><cell>0.86</cell><cell>0.87</cell><cell>0.94</cell><cell>High</cell></row><row><cell>Naive Bayes</cell><cell>0.75</cell><cell>0.79</cell><cell>0.72</cell><cell>0.73</cell><cell>0.82</cell><cell>Average</cell></row><row><cell>Support Vector Machine</cell><cell>0.82</cell><cell>0.84</cell><cell>0.80</cell><cell>0.81</cell><cell>0.89</cell><cell>High</cell></row><row><cell>Logistic Regression</cell><cell>0.79</cell><cell>0.83</cell><cell>0.77</cell><cell>0.78</cell><cell>0.86</cell><cell>Average</cell></row><row><cell>Stochastic Gradient Descent</cell><cell>0.80</cell><cell>0.82</cell><cell>0.79</cell><cell>0.80</cell><cell>0.87</cell><cell>Average</cell></row><row><cell>Multi-Layer Perceptron</cell><cell>0.84</cell><cell>0.86</cell><cell>0.82</cell><cell>0.83</cell><cell>0.91</cell><cell>High</cell></row><row><cell>LSTM</cell><cell>0.88</cell><cell>0.90</cell><cell>0.87</cell><cell>0.88</cell><cell>0.95</cell><cell>Very high</cell></row></table></figure>
		</body>
		<back>
			<div type="annex">
<div xmlns="http://www.tei-c.org/ns/1.0"><p>The step-by-step algorithm for training an LSTM model includes the following steps:</p><p>1. Data preparation:</p><p>• Input: receive a set of data containing time series or sequences. The approach presented in this study takes advantage of LSTM to predict time series. The LSTM is implemented using Keras (a high-level neural network interface that simplifies the process of creating and training artificial neural networks; it is a machine learning library that runs on the Tensorflow, Theano, and Microsoft Cognitive Toolkit frameworks) as a sequential model with two LSTM layers and a dense output layer. It receives a sequence of data inputs (normalized feature values without rejections) and outputs a sequence of failure values.</p><p>It was decided to use a single data input, as this significantly reduces the training time and is sufficient for the algorithm to recognize failure patterns. The length of the source sequence determines the duration of the runtime. Prediction performance is tested on three different input sequences: 1 (1 second), 300 (5 minutes), and 1800 (30 minutes). As a result, three different LSTM models were built. For training, we used a dataset with 70% failures, creating a split into training and test sets in the proportion of 25-75% without mixing. In addition, the data was prepared by creating output sequences for each data record, which were then used for training.</p><p>Based on the conducted research, a data prediction platform has been developed. Once the system has been successfully integrated into the smart home, the implementation process takes place, when the system becomes an active part of the home environment. However, this is only the beginning: further support, optimization, and continuous improvement of the system play a key role in ensuring its long-term and efficient operation in a smart home, adapting to changing needs and conditions (Fig. <ref type="figure">2</ref>). and managing risks in smart homes. The considered models have shown high accuracy in predicting failures.</p><p>The following areas can be considered for further development of this work:</p><p>• Improving machine learning methods to increase the accuracy of failure prediction. • In-depth study of the impact of the introduction of machine learning systems on the functioning of a smart home. Based on the obtained data, it is possible to build a methodology for predicting failures in a smart home, which will be the direction of the authors' next research. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>References</head></div>			</div>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Problems of Computer Engineering</title>
	</analytic>
	<monogr>
		<title level="m">Conference</title>
		<title level="s">Collection of abstracts</title>
		<imprint>
			<date type="published" when="2023">2023</date>
			<biblScope unit="page" from="125" to="127" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Machine Learning: Methods and Models: a Textbook for Bachelors</title>
		<author>
			<persName><forename type="first">K</forename><surname>Kononova</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Masters and Doctors of Philosophy in Specialty 051</title>
				<imprint>
			<date type="published" when="2020">2020</date>
		</imprint>
		<respStmt>
			<orgName>V. N. Karazin Kharkiv National University</orgName>
		</respStmt>
	</monogr>
	<note type="report_type">Economics</note>
</biblStruct>

<biblStruct xml:id="b2">
	<monogr>
		<author>
			<persName><forename type="first">Y</forename><surname>Bondarenko</surname></persName>
		</author>
		<title level="m">Manual for the Study of the Discipline</title>
				<imprint>
			<publisher>Lira</publisher>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
	<note>Statistical Analysis of Data</note>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">A Review of Machine Learning Methods in the Task of forecasting Financial Time Series</title>
		<author>
			<persName><forename type="first">K</forename><surname>Hureeva</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Kudin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Lisnyak</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Computer Science and Applied Mathematics</title>
		<imprint>
			<biblScope unit="volume">2</biblScope>
			<biblScope unit="page" from="18" to="28" />
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<monogr>
		<author>
			<persName><forename type="first">I</forename><surname>Puleko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Yefimenko</surname></persName>
		</author>
		<title level="m">Architecture and Technologies of the Internet of Things: a Textbookб State University</title>
				<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
		<respStmt>
			<orgName>Zhytomyr Polytechnic</orgName>
		</respStmt>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Coding for Information Systems Security and Viability</title>
		<author>
			<persName><forename type="first">B</forename><surname>Zhurakovskyi</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Information Technologies and Security</title>
		<imprint>
			<biblScope unit="volume">2859</biblScope>
			<biblScope unit="page" from="71" to="84" />
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Optimization Algorithms of Smart City Wireless Sensor Network Control</title>
		<author>
			<persName><forename type="first">M</forename><surname>Moshenchenko</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Cybersecurity Providing in Information and Telecommunication Systems</title>
		<imprint>
			<biblScope unit="volume">3188</biblScope>
			<biblScope unit="page" from="32" to="42" />
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Smart Factory of Industry 4.0: Key Technologies, Application Case, and Challenges</title>
		<author>
			<persName><forename type="first">B</forename><surname>Chen</surname></persName>
		</author>
		<idno type="DOI">10.1109/ACCESS.2017.2783682</idno>
	</analytic>
	<monogr>
		<title level="j">IEEE Access</title>
		<imprint>
			<biblScope unit="volume">6</biblScope>
			<biblScope unit="page" from="6505" to="6519" />
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">A Gap Analysis of Internet-of-Things Platforms</title>
		<author>
			<persName><forename type="first">J</forename><surname>Mineraud</surname></persName>
		</author>
		<idno type="DOI">10.1016/j.comcom.2016.03.015</idno>
	</analytic>
	<monogr>
		<title level="j">Computer Communications</title>
		<imprint>
			<biblScope unit="volume">89</biblScope>
			<biblScope unit="issue">90</biblScope>
			<biblScope unit="page" from="5" to="16" />
			<date type="published" when="2016">2016</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
