<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Machine learning techniques for fault-tolerance analysis and forecasting</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Nataliia Kuznietsova</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Petro Bidyuk</string-name>
          <email>pbidyuke_00@ukr.net</email>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Maryna Kuznietsova</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Jules Lepretre</string-name>
          <email>jules.lepretre@ecl17.ec-lyon.fr</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Ecole Centrale de Lyon</institution>
          ,
          <addr-line>36 Avenue Guy de Collongue, Écully, 69134</addr-line>
          ,
          <country country="FR">France</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Institute for Information Recording of the National Academy of Sciences of Ukraine</institution>
          ,
          <addr-line>Kyiv,03113</addr-line>
          ,
          <country country="UA">Ukraine</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>National Technical University of Ukraine "Igor Sikorsky Kyiv Polytechnic Institute"</institution>
          ,
          <addr-line>Kyiv, 03056</addr-line>
          ,
          <country country="UA">Ukraine</country>
        </aff>
      </contrib-group>
      <fpage>56</fpage>
      <lpage>68</lpage>
      <abstract>
        <p>This article is dedicated to the research of data storage reliability, fault-tolerance, and survivability. The paper performs modeling on the basis of accumulated historical data on the disks state which provided storage of information for cloud services customers. The requirements for uninterrupted access and high quality (without loss) of stored information for them are important. The main idea of our approach is as follows: if we have the SMART (SelfMonitoring, Analysis and Reporting Technology) characteristics of Hard Disk Drives (HDD) we can predict the time of disk failures and, so, restore the information on the new disks in time. To predict these moments of failure, and the performance of disks, we can use some different models such as ARMA, ARIMA, and linear regression, and forecast the daily failure rate. We took the BackBlaze datasets and showed the possibility to forecast risk exposure levels using SMART characteristics. We used a high-level analysis of risk quantification that allowed us to forecast both the failure rate and the consequent loss. Also, we used the classification models to predict the fact of failure and survival analysis to predict the moment when it could happen.</p>
      </abstract>
      <kwd-group>
        <kwd>1 Survivability</kwd>
        <kwd>fault-tolerance</kwd>
        <kwd>data mining modelling</kwd>
        <kwd>risk analysis</kwd>
        <kwd>neural networks</kwd>
        <kwd>LSTM</kwd>
        <kwd>RNN</kwd>
        <kwd>ARMA</kwd>
        <kwd>ARIMA</kwd>
        <kwd>survival models</kwd>
        <kwd>dynamic approach</kwd>
        <kwd>loss prediction</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>Ensuring the survivability and fault-tolerance of complex technical systems is a quite important task
and now it is even more urgent due to the enormous quantity of the IT-systems, programs, and support
networks of suppliers and producers. Even, for example, a logistic chain includes not only delivering
goods but also such processes as support, connection, databases of goods and tracks, and roads
networking. All these processes include a lot of data that should be gathered, processed and further used
for solving optimization tasks.</p>
      <p>The requirements for the operation of various online web services are associated with a heavy load
on the equipment and technical systems that ensure their smooth operation. The simultaneous failures
of the largest social networking services that occurred in October-November 2021 in Europe and around
the world were explained by the developers as a simultaneous technical load that the technical
equipment failed to cope with. Similar challenges arose in Ukraine before the application "Dija", when
a large number of users were forced to use the service for Covid-certificates download, thus creating an
excessive load. It led to system failure and in some cases even partial data loss. Now the requirements
to data especially to their confidentiality, storage reliability and also access due to the requirements of
their owners are currently the most relevant. Today, there is a growing digitalization of various areas of
activity and the transition to digital data which are regularly collected and stored in huge databases and
data warehouses, updated and supplemented daily, and therefore require a lot of space for their storage.</p>
      <p>Survivability of such systems means that they can achieve their functioning goals even in the
presence of faults and troubles, ensuring required efficiency at required time. So it’s necessary to have
all data and information, with all their characteristics of volume, correctness, security, links and means
of processing, etc. That’s why a good storage data system is required. The well-known company in this
area is BackBlaze, a cloud storage and data back-up company [1]. Its business is based on the data
protection and storage of the customers’ information. BackBlaze is involved in cloud storage and data
backup, providing extensive cloud storage services, helping customers develop and create their own
cloud solutions, backing up existing systems and files, and providing reliable interoperability with
workflows and tools. Backblaze has hundreds of pre-built integrations and partners, including
Facebook. Given the above-mentioned technical failures in the services of Facebook, the purpose of
our study was to test the reliability of data storage on disks from different manufacturers and determine
the longest reliable models to recommend Backblaze.</p>
      <p>Among other topics such as data privacy or data encryption, the company must provide its ability to
minimize the risk of data loss. For this reason, it is needed to assess and predict its Hard Disk Drives
(HDD) risk of failure, and to recommend or use some means to restore the information on the disks.</p>
      <p>The first disk monitoring technologies of the HDD have been introduced in the 1990’s. They aimed
to forecast the HDD failure. These reporting technologies were checking critical performance
parameters and providing binary information {Disk OK; Disk will fail soon}. A standardization of the
HDD reporting technology was introduced in 1995 and named SMART (Self-Monitoring, Analysis and
Reporting Technology). Despite the importance and criticality significance of this subject only very
few studies on failure characteristics of disk drives in this area were made and published. Most of the
available information comes from the disk manufacturers themselves [2].</p>
      <p>In work [3], the authors are working on BackBlaze 2011-2015 HDD reporting. They perform a
Kmean clustering down sampling to manually correct the huge class imbalance and are focusing in only
4 different HDD models (2 Seagate’s and 2 Hitachi’s).</p>
      <p>In [4] instead of forecasting the failure events, Ni, Jun, and Xiaoning Jin are providing five different
decision support tools for planning maintenance operations which aim at considering both production
and maintenance decisions jointly. Though providing interesting insights, the approach will be just
shortly discussed for the BackBlaze application.</p>
      <p>While utilizing and assessing the performance of different machine learning classifiers for predicting
failure in industrial equipment for anode production, the work of [5] really aims at emphasizing the
importance of the time window considered and its impact on the accurate measurement of a failure.</p>
      <p>In [6] the authors considered the risk of a failed HDD over the next month based on the last 3 months’
SMART characteristics. All disks were divided into two types of classes, defined as “0” (if the disk
would not fail over the next month) and “1” (if it would fail during the next month). To forecast this
fact all disk information was analyzed and the most important variables were defined. Based on the five
SMART characteristics it was calculated their last 3 months’ statistics. The obtained results were really
accurate.</p>
      <p>In our research, we will extend the application of this approach on new obtained data for the next
years and will study in time the efficiency of SMART-characteristics time series analysis and how we
can use their change for predicting disk failure.</p>
    </sec>
    <sec id="sec-2">
      <title>Problem statement</title>
      <p>Despite already existing and implemented tools for disk monitoring and replacement (mostly based
on the average lifespan of an HDD), the preliminary study of the BackBlaze dataset highlights that the
problem of disk failure still exists, and thus our research in this area will be relevant. It is still not
sufficient to prevent all failures and consequent data losses. Therefore the present work aims at
evaluating and forecasting the hard drive faults using the Self-Monitoring, Analysis, and Reporting
Technology characteristics of each hard drive.</p>
      <p>In this research, we focus on two different and complementary approaches. First, we will predict the
level of risk exposure, faced by BackBlaze company over time based on the failure rate and the total
storage capacity modeling. In addition to this high-level approach, we will use the survival models for
individually forecast the failure probability of each HDD independently. This implementation is
considering SMART characteristics as time-varying features. The approach is risk-oriented and
therefore should be used as a part of a decision support system for predictive maintenance or early
replacement.
3.</p>
    </sec>
    <sec id="sec-3">
      <title>Mathematical problem definition</title>
      <p>The aim of this part is to globally assess and forecast the level of risk exposure, faced by BackBlaze
company over time, and to describe the dataset at disposal. To properly assess the BackBlaze’s risk
exposure we propose to denote the risk of failure at the mathematical composition of the two
components: the probability of failure (the failure rate) and the estimated loss which is related to the
fact that because of the failure the total amount of available information on the disk is damaged or
missed. Equation 1 summarizes the problem definition we will model:</p>
      <p>
        R = P ⊗U ,
(
        <xref ref-type="bibr" rid="ref1">1</xref>
        )
where P is the probability of risk occurrence in general; U are related losses. In our task P is the
probability of HDD failure in the Datacenter and is dependent in time variable and could be determined
also as follows: P(t) = f (t, P(t − 1), P(t − 2),..., P(t − n)) . Losses in these cases are the losses of total
size storage and are increasing in time, and thus U (t) = f (t,U (t − 1),U (t − 2),...,U (t − n)) .
3.1.
      </p>
    </sec>
    <sec id="sec-4">
      <title>Preliminary Data Analysis</title>
      <p>Normalized SMART characteristics were used for time series analysis: each SMART characteristic
will be considered as a time series (like a sensor).</p>
      <p>In Figure 1 it is presented the number of reported HDD drive over time. While this number keeps
increasing as BackBlaze warehouses are expanding, the figure also underlines that there are 4 days with
missing a report for HDD. These missing reports shall not be considered as failed HDD, but rather as
reporting error. In the following work, the missing values for these dates were replaced by a linear
function between the two nearest neighbors. In Figure 2 it is shown the daily number of HDD reported
as failed. This value varies over time and seems close to a stationary signal. In addition, the maximum
is 24 failures in one day in 2017. This value will not be considered as an outlier, but rather as a probable
important technical incident in the data warehouse.</p>
      <p>As the number of HDD varies over time, we will consider the HDD daily failure rate as shown in
Figure 3.</p>
      <p>
        Overall, the daily failure rate is rarely exceeding 0.015%. The graph of the density function,
generally, has the shape of a right-skewed Gamma distribution. But in fact, and according to the results
provided by the best model, i.e. the model minimizing the chi-square value, it seems to be the beta
distribution (Equation 2) with parameters (2.7, 67.4, -6.6e-06, 1.4e-3). Tested distributions and their
fitting results are shown in Table 1. The fitted model on the distribution is shown in Figure 4.
(
        <xref ref-type="bibr" rid="ref2">2</xref>
        )
4.
      </p>
    </sec>
    <sec id="sec-5">
      <title>Modelling and forecasting the disks failure using time series analysis</title>
      <p>In this part, we will try to model and predict the daily failure rate (signal shown in Figure 3) using
different time series models. Based on Figure 3 we can make an assumption that the quantity of HDD
failures is non-stationary and failures can have the seasonal effects. That’s why we will try to remove
these effects and then return to the initial time series.
4.1.</p>
    </sec>
    <sec id="sec-6">
      <title>ARMA preprocessing</title>
      <p>To perform the ARMA modelling, we will implement the following signal transformations:</p>
      <p>Apply a natural logarithm to stabilize the variance: the Equation 3 will be applied to the signal.
Logarithm transformation applied to the signal is presented as such Equation 3:</p>
      <p>
        Failure_Ratelog = ln(2 + 100 * HDDrate% )
(
        <xref ref-type="bibr" rid="ref3">3</xref>
        )
b. Perform a differentiation of the values to remove the eventual trend and seasonality.
      </p>
      <p>The transformed signal is represented in Figure 5. The signal seems now stationary with a stabilized
variance, an ARMA model can therefore be considered for modelling the failure rate.</p>
      <p>
        Based on the computed Autocorrelation and Partial Autocorrelation Functions (PACF) we can do
some assumptions regarding the order of the autocorrelation or the moving average (with the ACF plot)
parts of our model. ARMA (
        <xref ref-type="bibr" rid="ref2">5,2</xref>
        ) seems to be a good candidate for sufficiently capturing our signal. In
addition, it is worth mentioning that the Dickey-Fuller test for stationary performed rejected the null
hypothesis, suggesting the time series does not have a unit root and is stationary.
4.2.
      </p>
    </sec>
    <sec id="sec-7">
      <title>ARMA modelling and results</title>
      <p>An ARMA model has the form depicted by the Equation 4.</p>
      <p>p q
X t = c + εt + ∑ ϕi X t−i + ∑ θ j εt− j .</p>
      <p>i=1 j=1</p>
      <p>AR( p) MA(q)</p>
      <p>The model implemented is a rolling-window ARMA model, where 85% of the dataset is used to fit
the model, and the remaining 15% are used for testing. As ARMA models are performing better in
short-term forecasting, we will forecast the next day’s failure rate value based on all past values (rolling
window).</p>
      <p>In Table 2, the performance of the model on the training set is provided with the AIC (Akaike
Information Criterion), the BIC (Bayesian Information Criterion), ranked from lowest AIC to highest
[7] gives the best model for the testing set and provides the R2 score as well as the MAPE (Mean
Average Percentage Error).
(4)</p>
      <p>
        As shown in the Table 2 the most accurate model seems to be the ARMA (
        <xref ref-type="bibr" rid="ref3 ref3">3,3</xref>
        ). In the ARMA
modelling it is also important to check the residuals and ensure they are close to a white noise.
5.
      </p>
    </sec>
    <sec id="sec-8">
      <title>Forecast of failure rate using linear regression</title>
      <p>It is also possible to convert our time series analysis problem to a linear regression. Indeed, a matrix
containing the n lagged values of the failure rate is built. The results considering different number of
failure rate’s lagged values are shown in Table 3. It appears that the consideration of the 16 past values
to forecast next day’s failure rate holds the best results. The feature importance of the past values is
shown in Figure 6.</p>
    </sec>
    <sec id="sec-9">
      <title>Forecast of total storage size</title>
      <p>As described before, the probability of failure is varying over time. To meet customer’s demand
the total storage amount of BackBlaze datacentre is also increasing over time (Figure 8).</p>
      <p>While different modelling techniques (including exponential and polynomial) have been tested, the
most accurate model for predicting the total storage size is the ARIMA. The best model results are
summarized in Table 4.</p>
      <p>
        The ARIMA (
        <xref ref-type="bibr" rid="ref1 ref1 ref1">1,1,1</xref>
        ) provides a really accurate model for storage prediction with a R2 score of more
than 99%. The model and forecasted time series are provided in Figure 9.
      </p>
      <p>Thanks to the modelling of both the failure rate and the total storage capacity, BackBlaze is now
able to forecast its risk exposure level. This risk level quantification enables it to properly set-up the
back-up solutions (quantity of HDD that will probably need to be replaced) and to quantify the required
maintenance/replacement actions. So, the disks are removed before they fail, and customers even don’t
know about it.</p>
    </sec>
    <sec id="sec-10">
      <title>7. Classification analysis – next-day failure forecasting based on the</title>
    </sec>
    <sec id="sec-11">
      <title>SMART characteristics and machine-learning classification algorithms</title>
      <p>In our research [6] we used different data mining methods for forecasting the failure of the disk. The
best model on received data was the random forest. It seems for us that we need also to develop approach
for finding appropriate compromise between monitoring and preventing disk failure in time. The most
used statistical indicators for assessing the classification quality are general accuracy, error matrix
(Confusion Matrix), first and second kind of errors, index GINI [7]. Usually, the error matrix is
determined for binary classification, size 2 × 2 . Let’s give some definitions.</p>
      <p>FN (False Negative) – first kind of errors, false negative value. For our task such error means that
the disk was evaluated as damaged in next day but actually it worked.</p>
      <p>FP (False Positive) – second kind of errors, false positive value. Here, this error means that the disk
was considered as a good one but actually it failed [8].</p>
      <p>The values of the confusion matrix are used to calculate the main metrics for estimating the
classification model.</p>
      <p>Precision shows that some of the objects which were called positive by the classifier and in fact they
were indeed positive:</p>
      <p>TP
precision = . (5)</p>
      <p>TP + FP</p>
      <p>Recall shows which part of the positive class objects out of all the positive class objects were found
by the algorithm:</p>
      <p>TP
recall = . (6)</p>
      <p>TP + FN</p>
      <p>The recall characteristic demonstrates the ability of the algorithm to find this class in general, and
precision – the ability to distinguish this class from other classes [9].</p>
      <p>F-measure (f1-score) is the average harmonic between precision and recall:</p>
      <p>presicion ⋅ recall
fβ = (1+ β2 ) ⋅
(β2 ⋅ precision) + recall
,
where β determines the accuracy measure in this metric, and for β = 1 it is the harmonic mean (with
the factor (1 + β2 ) equal to 2, so that in the case when precision = 1 and recall = 1), the f1-score
reaches its maximum at completeness and accuracy equal to one. The measure of f1-score approaches
zero if one of the indicators approaches zero [9].</p>
      <p>For our task we propose to build the confusion matrix for evaluating the faults’ risks as the tool for
evaluating the costs of losses in a case when the disk was not changed timely and the information was
lost, and if the disk was detected as required to change and in fact, it could still operate. In Table 5 the
results of the classification models tested in our investigation and confusion matrixes are presented.</p>
      <p>At the next step let us evaluate the value of wrong management decisions. Here we assume that the
cost of disks is defined as some value D , but the cost of information in case of loss is guaranteed by
the sum of the contract between BackBlaze and the client of storage. Let’s denote it as L . Usually, the
penalty in case of a fault is more than the cost of the disk. We can make also an assumption that
L = 10 × D . Now let us evaluate for each case the cost of the wrong decision not in time to change the
disks. There are four possible cases based on the different data mining methods for the classification of
the disks and prediction of their failure:
Losses1 = 329 × L + 82 × D = 329 ×10 × D + 82 × D = 3372 × D – obtained for the random forest
classifier with deep = 1;
Losses2 = 257 × L + 71× D = 257 ×10 × D + 71× D = 2641× D – obtained for the random forest
classifier with deep=10;
Losses3 = 359 × L + 79 × D = 359 ×10 × D + 79 × D = 3669 × D – obtained for the logistic regression;
Losses4 = 307 × L + 82 × D = 307 ×10 × D + 82 × D = 3152 × D – obtained for the support vector
machine.
(7)</p>
      <p>As we can see from the presented results, the lowest cost required for this time period is for the
model of Random Forest for deep = 10 and thus this method is more effective from the costs point of
view. Certainly, we can see that the most effective management decisions are really dependable on the
costs of the disks and penalty for the company BackBlaze in the case of information losses. Of course,
the reputation for the company is more expensive and thus the real time monitoring and prevention of
risks is really significant. That’s why our next proposal was to develop such dynamic approach based
on the time changing of the disks statuses and their statistics which gives us the opportunity to catch
the increasing risks of the failure.
8.
8.1.</p>
    </sec>
    <sec id="sec-12">
      <title>The task of predicting the time of failure risk</title>
    </sec>
    <sec id="sec-13">
      <title>Approach development</title>
      <p>For each hard disk we have five main characteristics time series – measurements of
SMARTcharacteristics (as well as in the previous point, the zero instances were rejected identically and the data
were logarithmically transformed). Different approaches can be used to predict the time of risk. If it
could be formed a sample similar to the first part, and built a regression (i.e., the vector of features is
formed according to the data for three months, and as a label – the time that the disk worked after this
point). If we consider the tools of survival analysis, then, since we have time-varying covariates, it is
advisable to use the Cox Proportional Hazards Model with Time-Dependent Covariates [7, 10]. To do
this, the data for each disk was divided into intervals of 1 day, and the table will have columns: ID –
disk number; tStart and tStop – numbers of the day of measurement and the next day; Censoring –
whether the disk broke that day: Smart_5, Smart_187, Smart_188, Smart_197, Smart_198.</p>
      <p>Using the coxphfit function (from the survival library of the R language, also similar functions are
available for Matlab and SAS) from the given interval measurements it is possible to obtain coefficients
for the proportional Cox model. However, for its use in calculating the probability of failure after the
certain number of days, the predicted values of the covariant are required (and the task of predicting
SMART-characteristic itself implies additional error), so this model is generally not used for
forecasting, but still allows to obtain survival function.</p>
      <p>In order to move to a time-independent covariant and use Cox's Proportional Hazard model let's
assume that the basic statistics sufficiently describe the state of the hard disk, and do not depend much
from the length of the observed series. To form a training sample for each hard drive, a feature vector
was compiled: maximum, minimum, average, and nonzero values of each of the five selected SMART
characteristics over the entire drive observation period. The label was the lifetime of the disc and the
censorship indicator. The following coefficient was obtained for the covariance: Concordance index =
0.87. To predict the time to failure, we form a vector of features from all available observations for a
given disk, using the built model we can determine the length of the disk lifetime for a given probability
threshold, and subtract the time that the disk has already worked.
8.2.</p>
    </sec>
    <sec id="sec-14">
      <title>Example of calculation</title>
      <p>For example, for the disk 'Z3029FSB' we have the vector formed from data for 100 days:
[2.1972245773362196, 0.0, 0.06462425227459469, 0.029411764705882353, 2.1972245773362196,
0.0, 0.06462425227459469, 0.029411764705882353, 0.0, 0.0, 0.0, 0.0, 1.0986122886681098, 0.0,
0.8778107128746405, 0.8294117647058824, 1.6094379124341003, 0.0, 0.18858392848979214,
0.14411764705882352].</p>
      <p>In figure 10 obtained probability of surviving, i.e. the probability that the disk will continue working
without failing in time, is presented. We build the survival model for one year ahead. Next, the company
defines the probability trigger which is the less acceptable probability that the disk will work next
months. So from figure 10 we can define that for example, with a probability of 0.95, the disk will work
for up to 210 days. So, confirmed by surviving models, the estimated operating time is 110 days. In
fact, the disk had been working for another 240 days.</p>
      <p>It means that if the company determines the triggers as the cut-off for the probability of disk survival,
it means the moment when the disk should be removed. In our previous section, we showed that
spending money on disks 30 days ahead could lead to extra losses. But on the practical and reputational
side it is better to change the disk later and for this reason, the cut-off should be decreased to 0.94. So,
survival models are good instruments to define the correct disk cut-off and are more flexible for risk
triggers varying.</p>
    </sec>
    <sec id="sec-15">
      <title>9. Conclusions</title>
      <p>The research performed in the area of fault-tolerance and survivability analysis of hard disk drives
confirms that the problem of fault-tolerance in the technical sphere is still relevant [11]. The methods
of data mining applied for solving the forecasting problem of the fault probability give quite a high
accuracy but still, there are some disks that are not forecasted as suspicious and required change.
Actually, the task of searching for the compromise between changing the disks as a prevention measure
is required, and defining a more precise forecasting of fault is required. That’s why the approach
proposed in this paper and based on the survival models, and time SMART-characteristics monitoring
is a good tool for solving this dilemma. Survival modelling gives the possibility to forecast for each
disk its fault tolerance [12] and based on the experience of the technical system and production needs
to make the triggers, which would allow forecasting the critical moment for disks and prevent the
failure. In our next study, we are going to apply this approach in combination with other data mining
methods and evaluate the accuracy of forecasting. Time-monitoring and dynamic risk assessment
confirmed their effectiveness in analyzing financial systems [7], so we expect that a combination of
integrated and dynamic approaches will be effective also for technical systems.
Knowledge Discovery and Data Mining (KDD '16). Association for Computing Machinery, New
York, NY, USA, 2016, pp. 39–48, doi: 10.1145/2939672.2939699
[4] N. Jun, X. Jin, Decision Support Systems for Effective Maintenance Operations, CIRP Annals,
vol. 61, no 1, 2012, pp. 411‑14. DOI.org, doi:10.1016/j.cirp.2012.03.065.
[5] N. Kolokas, T. Vafeiadis, D. Ioannidis and D. Tzovaras, Forecasting faults of industrial equipment
using machine learning classifiers, 2018 Innovations in Intelligent Systems and Applications
(INISTA), 2018, pp. 1-6, doi: 10.1109/INISTA.2018.8466309.
[6] N. Kuznietsova, M. Kuznietsova, Data mining methods application for increasing the data storage
systems fault-tolerance, IEEE 2nd International Conference on System Analysis &amp; Intelligent
Computing (SAIC), 2020, pp. 315–318. https://doi.org/10.1109/SAIC51296.2020.9239222
[7] N. V. Kuznietsova, P. I. Bidyuk, Theory and practice of financial risk analysis: systemic approach,</p>
      <p>Lira-K, Kyiv, 2020.
[8] F. Herrera, F. Charte, A.J. Rivera, M.J. del Jesus, Multilabel Classification Problem Analysis,
Metrics and Techniques, Springer International Publishing, Switzerland, 2016. doi:
10.1007/9783-319-41111-8.
[9] M. Hossin, M.N, Sulaiman, A Review on Evaluation Metrics for Data Classification Evaluations,
International Journal of Data Mining &amp; Knowledge Management Process 5 (2015) 1-11. doi:
10.5121/ijdkp.2015.5201.
[10] D. R. Cox, Regression Models and Life-Tables. Journal of the Royal Statistical Society, Series B,
1972, Vol. 34, No2. P. 187–220.
[11] A. Dodonov, O. Gorbachyk, M. Kuznietsova, Increasing the survivability of automated systems of
organizational management as a way to security of critical infrastructures, CEUR Workshop
Proceedings, 2018, Vol. 2318, pp. 261‒270, http://ceur-ws.org/Vol-2318/.
[12] P. Zheng, S. Yuan, X. Wu, SAFE: A Neural Survival Analysis Model for Fraud Early Detection,
2018. URL: http://arxiv.org/abs/1812.07142.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          <article-title>[1] Backblase homepage</article-title>
          . URL: https://www.backblaze.com/b2/hard
          <article-title>-drive-test-data</article-title>
          .
          <source>html.</source>
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>D.</given-names>
            <surname>Anderson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Dykes</surname>
          </string-name>
          , E. Riedel,
          <article-title>More than an interface - scsi vs</article-title>
          . ata,
          <source>in Proceedings of the 2nd USENIX Conference on File and Storage Technologies (FAST'03)</source>
          ,
          <year>2003</year>
          , pp.
          <fpage>245</fpage>
          -
          <lpage>257</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <surname>M. M. Botezatu</surname>
            ,
            <given-names>I. Giurgiu</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Bogojeska</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Wiesmann</surname>
          </string-name>
          ,
          <article-title>Predicting Disk Replacement towards Reliable Data Centers</article-title>
          ,
          <source>in: Proceedings of the 22nd ACM SIGKDD International Conference on</source>
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>