<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>ORCID:</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>On Parallel Processing of Machine Learning Based On Big Data and Voronoi Tessellation</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Vasyl Martsenyuk</string-name>
          <email>vmartsenyuk@ath.bielsko.pl</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Marcin Bernas</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Aleksandra Klos-Witkowska</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Tomasz Gancarczy</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>University of Bielsko-Biala</institution>
          ,
          <addr-line>2 Willowa, Bielsko-Biala, 43-309</addr-line>
          ,
          <country country="PL">Poland</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Voronoi tessellation. Computational</institution>
        </aff>
      </contrib-group>
      <volume>000</volume>
      <fpage>0</fpage>
      <lpage>0001</lpage>
      <abstract>
        <p>The paper is devoted to the development of an approach to machine learning for Big Data under epistemic and aleatoric uncertainties which are taken into account with the help of corresponding minimax criteria. The keystone of the method is processing subsets of training data in parallel, using the partitioning based on complexity is analyzed and compared with sequential data processing. An example from a medical application is considered, where the method is investigated for different learners and resampling strategies.</p>
      </abstract>
      <kwd-group>
        <kwd>Parallel machine learning</kwd>
        <kwd>Big Data</kwd>
        <kwd>learner</kwd>
        <kwd>uncertainty</kwd>
        <kwd>minimax</kwd>
        <kwd>Voronoi diagram</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <sec id="sec-1-1">
        <title>The active usage of Big Data technology in various branches [1-8] requires the development of high</title>
        <p>performance algorithm for solving the Machine Learning (ML) problems. To cope with Big Data
parallel computing is one of the most effective solution in the case of ML. It leads to the necessity of
the partitioning on Big Data sets. Voronoi diagrams are traditionally used for such type of problems.</p>
      </sec>
      <sec id="sec-1-2">
        <title>The minimax approach (together with maximin and maximax) is traditionally used for regression</title>
        <p>
          problems [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ], [
          <xref ref-type="bibr" rid="ref10">10</xref>
          ]. In the case of ML, one of the generalized minimax approaches is known as the
        </p>
      </sec>
      <sec id="sec-1-3">
        <title>Minimax Probability Machine (MPM) [11].</title>
      </sec>
      <sec id="sec-1-4">
        <title>It can be argued that MPM is a classic result of studying the reliability of intelligent models [12], [9], which can be considered as a typical method of classifying the reliability of learning. The task of MPM optimization is to minimize the upper limit of the probability of incorrect classification of the study of model parameters.</title>
        <p>
          The upper limit of the probability of incorrect classification can be used as an explicit indicator to
assess the reliability of classification models. A version of MPM with parametric reduction was
proposed in [
          <xref ref-type="bibr" rid="ref11">11</xref>
          ], [
          <xref ref-type="bibr" rid="ref13">13</xref>
          ] for nonlinear classification problems. Several advanced MPM algorithms have
been presented from different points of view [14], [15], [16], [17]. In [15], [16] it was pointed out that
in some cases it is necessary to distinguish the probability of incorrect classification of two classes, as
one class may be more important than another. In [18], MPM was extended for regression. In [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ], MPM
was introduced to prepare the fuzzy classifier for a more transparent and understandable classification
model. In addition to MPM, the study of the reliability of intelligent models has been considered from
other points of view.
classification models in [19], [20].
        </p>
      </sec>
      <sec id="sec-1-5">
        <title>For example, the concepts of "conflict" and "ignorance" were introduced to denote the reliability of</title>
      </sec>
      <sec id="sec-1-6">
        <title>To make the method of adopting the minimum probabilistic method available for learning additional intelligent models and to implement the study of the reliability of these models, a generalized hidden minimum probability machine (GHM-MPM) is proposed. MPM classification was used as an explicit indicator to characterize the reliability of the classification model.</title>
        <p>2022 Copyright for this paper by its authors.</p>
      </sec>
    </sec>
    <sec id="sec-2">
      <title>2. Description of the method</title>
      <p>The problem of supervised ML, which means the prediction of  with the help of  , loss function
 , and the set of probability distributions  on ( ,  ) can be formulated as minimax problem with the
respect to  , provided that the maximization is due to all possible distributions  and minimization is
with the respect to decision rules  ∈  .</p>
      <p>∈  ∈
 [ ( ,  ( ))]
where  [•] is expectancy.</p>
      <sec id="sec-2-1">
        <title>The problem (2.1) can be solved with the help of introducing the generalization of the entropy</title>
        <p>maximum principle. Mathematical description of the problem of supervised ML in systemic medical
research was presented in [21], [22]. Here we formulate it in the case of minimax criterion.
Mathematically ML problem for the systemic medical research is based on the following data. We have
dataset  , which includes  tuples</p>
        <p>D  X i | i  1, N</p>
        <p>In order to model aleatory uncertainties, consider supervised ML regarding the distribution of
learning tuples. For the class of all subsets of  we introduce  , including the distributions of classes
of training and testing datasets
  (Dtrain, j , Dtest, j)  D  D | Dtrain, j  Dtest, j  , Dtrain, j  Dtest, j  D, j  1,2N ,,

 
where Dtrain, j and Dtest, j are all possible datasets for training and testing correspondingly. In practice,
resampling strategies are distributions of the classes of tuples which are characterizing aleatoric
uncertainties the best. We introduce the resampling strategies  ⊂ 
  (Dtrain,k , Dtest,k)  D  D | Dtrain, j  Dtest, j  ,k Dtrain,k  k Dtest,k  D,
(2.4)
(2.1)
(2.2)
(2.3)
(2.5)
2.1.</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>The problem of the dimension reduction</title>
      <sec id="sec-3-1">
        <title>Because real sets of systemic medical studies include dozens of vital signs, morphological, biochemical, and clinical assessments, it is natural to want to reduce the number of symptoms, leaving</title>
      </sec>
      <sec id="sec-3-2">
        <title>As examples of resampling strategies we can consider  3,  5,  10, which correspond to fold cross-validation for different  .</title>
        <p>Each  th tuple   = ( 1,  2, . . . ,   ,   ) consists of input data ( 1,  2, . . . ,   ) (called by also
attributes) and output data   .</p>
        <p>Let raw   = (  1,   2, . . . ,    ) present the value of  th attribute of all  tuples. Output attribute  =
( 1,  2, . . . ,   ) includes all output data. The attributes  1, . . . ,   and  (depending on the tasks of
classification or regression) can accept both numerical and categorial values.</p>
        <p>In the simplest case, the supervised ML problem is to predict, using the certain predictor, the value
of output attribute   +1 based on the values of attributes  1 +1, . . . ,    +1. The predictor should
maximize the accuracy of prediction of output attribute, namely the probability
 {     1 +1, . . . ,    +1} for arbitrary  ∈ {1, . . . ,  }1. Further, applying minimax
approach, we introduce ℎ ∈  for the considered class of ML models ℎ( ,  ), which can be trained and
tuned for the data  ⊂  and assessed taking into account certain resampling strategies  . Comparing
different ML models, the goal is to minimize expected losses. But you also need to consider resampling
strategies, which should also assess the loss function. This formulation of the ML problem considers
two types of uncertainty. Namely, uncertainty in oversampling is aleatory because it is related to data.
At the same time, the uncertainty in the choice of models is epistemic. Mathematically, the minimum
problem of MN is described as a search for a model ℎ due to


ℎ∈

 ∈
 [ ( , ℎ( ,  ))]



strategy 
the ones with the greatest differences. The principal components analysis method (PCA) is one of the
widely used methods of dimensional reduction. Although it is used for unsupervised ML problems, it
helps us refine the results when used for supervised ML, such as a classification or regression problem.
The task of reducing the number of attributes is extremely important for medical use in interpreting the
results. Below we propose a method of its application in conditions of aleatory uncertainty.</p>
        <p>When regarding the Voronoi tessellation, the dimension reduction algorithm has to be applied for
each Voronoi cell. Moreover, since in the minimax ML (machine learning) problem the loss function
is calculated for all resampling strategies, the dimension reduction algorithm must be applied separately
for each strategy  . We can present arbitrary resampling strategy  as    =



( ), where
( ),  ∈ 1,  is lth sample of indices from 1 to N, which corresponds to training tuples of the

Let   ( ) be input data coming from  , if sample of indices  
( ) were applied. Namely,
  ( ) = {(</p>
        <p>1,  2, . . . ,   ,   ) }
the lth training sample.</p>
        <p>∈{1... }∩ 
( )</p>
        <p>, where Nl &lt; N is the number of training tuples in</p>
      </sec>
      <sec id="sec-3-3">
        <title>Before training and tuning the model ℎ( ,  ), we reduce the dimension  ∈  of D with the respect to  . For this purpose we offer the modification of PCA method with the respect to Voronoi cell and resampling strategy  (see Algorithm 1).</title>
      </sec>
      <sec id="sec-3-4">
        <title>Algorithm 1: PCA for the resampling strategy</title>
      </sec>
      <sec id="sec-3-5">
        <title>Input data:</title>
        <p>= {( 
1,  2, . . . ,   ,   ) }
 =1
, 
Output: principle components together with the attributes.
1 transform data D into the matrix A including all numerical entries;
2 apply resampling strategy  for А:   = {( 1, , 
2, , . . . ,  
 , ,   , ) } 
 =1
∈   1+1×  ,  = 1,  ;
3
4
5
6
7
8
9
  , := 1</p>
        <p>calculate 

  : =
1
 1
for each   ( ),  = 1,  do
∑ 
 =1

 , ,  = 1,  1;

(  ): =
∑ 1 
 =1</p>
        <p>(  , );
(  , ),  = 1,  1;
 ′ (</p>
        <p>′ ) ∈   1.,
 ′ : = { 
 , −   , } =1, 1, =1,</p>
        <p>∈   1×  .;
calculate eigenvalues:   ,1 ≤   ,2 ≤. . . ≤   , 1
10 calculate</p>
        <p>eigenvectors
11 
  , 1respectively
Var(PC1 ), ExplainedVar(PC2 ): =
Var(PC2 );
Var(  )
ExplainedVar(PC12): =</p>
        <p>∑ ExplainedVar(PC12)
1


 =1
17 return names of attributes з  (  , 1 ),  = 1,  і  (  , 1−1),  = 1,</p>
      </sec>
      <sec id="sec-3-6">
        <title>Next, we describe the basic steps of Algorithm 1. In step 1, we convert all categorical attributes, encoding them as a set of boolean inputs, each of which represents one category 0 or 1. We can generate columns with category flags automatically. 106</title>
        <p>values of raws (Step 4), variance</p>
        <p>(  , ),  = 1,  1 (Step 5), general variance (sum of sample
variances)</p>
        <p>(  ) (Step 6), deviation matrix  ′ ∈   1×  (Step 7), covariance matrix   ∈   1 (Step
8), eigenvalues of matrix   due to increasing order (Step 9), eigenvectors   (Step 10). Here we
consider eigenvectors   , 1</p>
        <p>and   , 1−1 ∈   1, which correspond to   , 1and   , 1−1 respectively. At
the step 11 we get two principle components 
(PC2l) respectively (Step 13). Next, we organize the values of the eigenvectors   , 1і   , 1−1 in
descending order of their absolute values. For this purpose, we use permutations  (  , 1) and
 (  , 1−1). We use the denotion  ( ) for a permutation that organizes the vector x in descending order
of the absolute values of its elements.</p>
        <p>Next we return the names of the first ExplainedVar (PC1l) 100% attributes in permutation  (  , 1 )
and the first ExplainedVar (PC2l) 100% attributes in permutation  (  , 1−1) (Step 14).</p>
      </sec>
      <sec id="sec-3-7">
        <title>After completing the main cycle, we calculate the variance of the main components for the</title>
        <p>resampling strategy \ gamma (Step 16). Finally, we return the names of the first ExplainedVar (PC1l)
100% attributes, which are most common in permutations  (  , 1 ),  = 1,  and the first ExplainedVar
(PC2l) 100% attributes that are most common in permutations  (  , 1−1),  = 1,  (Step 17).</p>
        <p>As a result of reduction
of dimension
we receive some
numerical
matrix  
=
∈   2+1× ,  2 ≤  1. These data can then be used as training to solve ML
{( 1;  2, . . . ,   2,   ) }</p>
        <p>=1
problems based on the minimax approach.</p>
      </sec>
      <sec id="sec-3-8">
        <title>Note 1. Stages 2 and 10-14 are modifications of the traditional PCA algorithm. First, in step 1, we</title>
        <p>convert all categorical attributes that are widely used in systemic medical research into boolean data.
Second, when considering the two main components traditionally used for planar presentation of
training kits, we propose an approach to selecting some reduced number of attributes for further research
(e.g., developing a ML model). This number is related to the number of variations explained. The latter
assumption allows us to truly reduce the size of ML problems in systemic medical research under
uncertainty.</p>
      </sec>
      <sec id="sec-3-9">
        <title>Note 2. Of course, we must take into account the case if the variance due to the first two components</title>
        <p>is low. In such cases, we need to take into account the components PC3, PC4 and so on to obtain the
appropriate dispersion. Steps 10-14 and other algorithms should be changed accordingly.</p>
        <p>Note 3. It should be noted that the PCA should be calculated depending on the resampling strategy,
as the PCA is applied to training tuples.   ( ) (not for the whole data set D). Therefore, in Step 14,

different features may be selected depending on the sample of indices  
the selection of attributes in the last step of the algorithm for the entire resampling strategy.
( ). In turn, this affects
2.2.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>General flowchart of parallel machine learning with the help of Voronoi diagrams</title>
      <sec id="sec-4-1">
        <title>The general block diagram (Figure 2.1) allows us to obtain a learner of the ML problem based on a</title>
        <p>minimax approach with the possibility of accurate, acceptable and stable results. The MN model,
formulated under conditions of certainty, is presented in [22]. Here, we summarize a flowchart for
solving the problem under uncertainty in both the model and the resampling strategy.</p>
      </sec>
      <sec id="sec-4-2">
        <title>We start with the import and preparation of data (feature generation, gap filling, normalization)</title>
        <p>collected in EMR systems. Methods of importing data sets from EMR systems are presented. Note that
the choice of open source EMR systems over commercial ones is extremely important because it allows
open access to clinical data that can be processed and selected for subsequent stages of ML [22].</p>
        <p>Then we should define the task from the point of view of MN. This can be regression, classification,
grouping, and so on. Resampling strategies  are also defined. For example, resampling strategies
supported by the mlr package include: cross-validation (cv), cross-validation (LOO),
re-crossvalidation (RepCV), color subsampling, also called Monte Carlo cross-validation (Subsample), Holdout
method (training / testing) (Holdout) [23]. In a real application, we are dealing with a large number of
attributes. Only some of them can be important for the tasks of MN. Therefore, it is natural to try to
reduce the dimension by discarding the attributes with the largest deviations.</p>
      </sec>
      <sec id="sec-4-3">
        <title>Next we specify the set Ѱ of appropriate methods (learner) of the solution. The most important is</title>
        <p>the choice of parameters for the methods, which affects the accuracy of the model. In the next cycle,
configure the parameters for each model with Ѱ based on all resampling strategies  , which are used.</p>
      </sec>
      <sec id="sec-4-4">
        <title>The original model will satisfy the criterion of the minimax approach (2).</title>
        <p>Prepar
ation
of
data
Determ
ining
task
Regression,
classification,
grouping, ...</p>
        <p>Voronoi
tesselat
ion
Minimax
ML model
Result voting for
Voronoi cells
Resam
pling
strateg</p>
        <p>ies
 ∈ Γ
Choos</p>
        <p>ing
model
ℎ ∈
Ψ</p>
        <p>Minimization of ℎ ∈ Ψ
Dimension
reduction
Tuning
model
ℎ( , γ)
Assessment
of loss
function
 ( , ℎ( , γ))
Maximization of  ∈ Γ</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>Computational complexity</title>
      <sec id="sec-5-1">
        <title>In order to analyze the computational complexity of the proposed approach, consider an example of</title>
        <p>a set Ѱ, which includes the method of a 4-layer neural network with the number of neurons  ,  ,  ,  on
layers based on inverse error propagation and method C5.0 induction of decision tree height ℎ.</p>
      </sec>
      <sec id="sec-5-2">
        <title>Assume that the training data sample  includes #( ) tuples based on  attributes.</title>
      </sec>
      <sec id="sec-5-3">
        <title>Let  be the number of seeds for Voronoi tessellation. Corresponding computational complexity is</title>
        <p>≔  (  
 +  ⌈2⌉)</p>
      </sec>
      <sec id="sec-5-4">
        <title>The computational complexity of the specified neural network method based on  iterations is</title>
        <p>Error! Reference source not found.:</p>
      </sec>
      <sec id="sec-5-5">
        <title>Computational complexity of decision tree induction [28]:</title>
        <p>Computational complexity of resampling based on  -fold cross validation is Error! Reference
source not found.:
 
: =  ( #( )( +</p>
        <p>+  ))
  5.0 ≔  (ℎ#( )</p>
        <p>#(  ))
 
=  ( #( ))
(2.6)
(2.7)
(2.8)</p>
      </sec>
      <sec id="sec-5-6">
        <title>Thus, the computational complexity of constructing the ML model based on the scheme in Figure 2.1 is</title>
        <p>=      ( 
+   5.0)
(2.9)</p>
      </sec>
      <sec id="sec-5-7">
        <title>Since k is constant, then from (2.9) shows that the computational complexity increases by one order of magnitude.</title>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>3. Example of the medical data</title>
      <sec id="sec-6-1">
        <title>Modern systemic medical research (evidence-based medicine) is the integration of the best scientific</title>
        <p>evidence with clinical experience and patient expectations [30]. They are aimed at improving health
care in the future. Systematic medical research helps doctors and researchers gain knowledge about
human health and disease. They also allow you to find more effective ways to prevent and treat disease.
Assessment of health is based on a comprehensive and systematic examination of the patient, which
includes history, objective examination of the body, analysis of laboratory blood tests and various
secretions, instrumental and interventional studies, including X-ray, CT, MRI, endoscopy, biopsy and
others methods.</p>
        <p>Nowadays, cardiovascular diseases attract attention because they are "the number one cause of death
in the world" [31]. In the study of cardiac diseases, there are quite a number of nuances and indicators
that experts pay attention to during diagnosis. Diagnostic criteria include both physical tests and history,
as well as laboratory, instrumental research methods. During the survey, the doctor may ask questions
about the patient's family members (genetic predisposition), lifestyle and habits. Physical inactivity
(sedentary lifestyle), unhealthy diet, alcohol consumption and smoking significantly increase the risk
of cardiovascular disease. During laboratory studies, much attention is paid to the assessment of the
level of lipids and their fractions (lipid profile). It includes indicators of total cholesterol, triglycerides,
high, low, very high and very low lipoprotein density, as well as the level of atherogenicity. Lipid
imbalance increases the risk of atherosclerosis. Among other things, the patient's overweight is one of
the dangerous risk factors for heart disease. Blood glucose and glycated hemoglobin are among the
most important indicators of carbohydrate metabolism in the body and markers of diabetes. Diabetes is
a separate disease, but its presence significantly increases the risk of cardiovascular disease. In addition
to the risk assessment, the necessary extended hematological, biochemical and instrumental studies are
performed. In addition to the general blood test, the patient's blood pressure is measured, the following
instrumental methods are used: electrocardiogram (ECG), Holter monitoring, echocardiography,
coronary angiography, MR angiography.</p>
        <p>This experimental study includes data from 1651 patients diagnosed with myocardial infarction. The
target attribute of forecasting is life expectancy. Each patient's data includes 97 attributes that contain
both numerical and categorical values. Such information includes data on the type of heart attack (focal
or transmural), the location of the heart attack (anterior or posterior). Mortality information (hospital,
short-term and long-term) is also used. The presence of concomitant pathologies is described. And here
we use a detailed analysis, because such pathologies can be combined. Risk factors typical of
cardiovascular diseases are investigated, namely, clinical evaluation includes data on such risk factors
as gastritis, gallstone disease, lung disease, nephrological disorders, rheumatic thyroid disease,
angiopathology, gastrointestinal diseases, oncology, chronic obstructive pulmonary disease,
hypertension, diabetes, smoking.</p>
        <p>The considered detailed clinical course includes indicators of vital functions, namely heart rate,
systolic blood pressure and diastolic blood pressure, analyzes of heart attack complications in the form
of arrhythmias, in particular, detailed heart attack complications developed in the hospital. The data
lists all indicators of the general analysis of blood. Special attention is paid to leukocytes (WBC),
biochemical analysis of blood is presented, information on medicines which the patient received in
hospital is included. After the dimension reduction algorithm, the following features remained: sex,
age, re-myocardial infarction (RMI), life expectancy after MI (death_days), body mass index (BMI),
leukocyte density (White_blood_cells_count), left ejection fraction ventricle (LVEF).</p>
      </sec>
      <sec id="sec-6-2">
        <title>We consider set Ѱ, which includes the models of linear regression (regr.lm), SVM model with radial</title>
        <p>base kernel (regr.ksvm), and random forest (regr.ranger).</p>
        <p>Resampling strategies include cross-validation of cv3, cv5, cv7, cv9, cv10. The loss function L was
calculated as the rmsse rmse and the training time. In the case of RMSE as an indicator of efficiency
(Table 2.1), the regr.ksvm model is a solution of the ML problem based on the minimax. Namely, we
first compare the error values for all the models considered. In the second step, we see that the RMSE
value for the ksvm model will be minimal among the maximum. In Figure 2.2 we can see the analysis
of the effectiveness of ML models with different resampling strategies for standard deviation as an
indicator of efficiency.</p>
      </sec>
    </sec>
    <sec id="sec-7">
      <title>4. Conclusions</title>
      <sec id="sec-7-1">
        <title>For example, as can be seen from Table 2.2, there is a resampling strategy (namely cv10), in which the random forest model shows the lowest value of the RMSE loss function on the set of all models considered.</title>
      </sec>
      <sec id="sec-7-2">
        <title>At the same time, there are resampling strategies (cv3), in which this model shows greater errors</title>
        <p>compared to the model of SVM. In this situation, the choice of a random forest model would lead to
unexpected losses arising from aleatory uncertainty.</p>
        <p>Therefore, the minimax approach proposes to establish a resampling strategy with the maximum
("worst") value of the loss function, on which the desired model should behave best (get the minimum
value of the loss function).
5. References
[14] Z. Deng , L. Cao , Y. Jiang , S. Wang , Minimax probability tsk fuzzy system classifier: a more
transparent and highly interpretable classification model, IEEE Trans. Fuzzy Syst. 23 (4) (2015)
813–826 .
[15] K. Huang. , H. Yang , I. King , et al. , The minimum error minimax probability machine, J. Mach.</p>
        <p>Learn. Res, 5 (4) (2004) 1253–1286 .
[16] K. Huang , H. Yang , I. King , M.R. Lyu , Imbalanced learning with a biased minimax probability
machine, IEEE Trans. Syst. Man Cybern. Part B 36 (4) (2006) 913–923 .
[17] T. Strohmann , G.Z. Grudic , A formulation for minimax probability machine regression, in: 2002</p>
      </sec>
      <sec id="sec-7-3">
        <title>Neural Information Processing Systems (NIPS), (2002) 769–776 .</title>
        <p>[18] T. Strohmann , G.Z. Grudic , A formulation for minimax probability machine regression, in: 2002</p>
      </sec>
      <sec id="sec-7-4">
        <title>Neural Information Processing Systems (NIPS)(2002) 769–776 .</title>
        <p>[19] E. Lughofer , Single-pass active learning with conflict and ignorance, Evolving Syst 3 (4) (2012)
251–271 .
[20] E. Lughofer , O. Buchtala , Reliable all-pairs evolving fuzzy classifiers, IEEE Trans. Fuzzy Syst
21 (4) (2013) 625–641.
[21] V. Martsenyuk, L. Babinets, Y. Dronyak, O. Paslay, O. Veselska, K. Warwas, I. Andrushchak, and</p>
      </sec>
      <sec id="sec-7-5">
        <title>A. Klos-Witkowska, On development of machine learning models with aim of medical differential</title>
        <p>diagnostics of the comorbid states," in 2019 10th IEEE International Conference on Intelligent</p>
      </sec>
      <sec id="sec-7-6">
        <title>Data Acquisition and Advanced Computing Systems: Technology and Applications (IDAACS),</title>
        <p>IEEE, Sep. 2019,pp313-318 doi:10.1109/idaacs.2019.8924345.
[22] V. Martsenyuk, V. Povoroznyuk, A. Semenets, and L. Martynyuk, On an approach of the solution
of machine learning problems integrated with data from the open-source system of electronic
medical records: Application for fractures prediction, Artificial Intelligence and Soft Computing,
Springer International Publishing, 2019, pp. 228-239.doi:10.1007/978-3-030-20915-5_21.
[23] Resampling, https://mlr.mlr-org.com/articles/tutorial/resample.html, (Accessed on 10/08/2020)
[24] Jun Ma, Liming Yang, Yakun Wen, and Qun Sun, Twin minimax probability extreme learning
machine for pattern recognition, Knowledge-Based Systems 187, (2020) 104806.
doi:10.1016/j.knosys.2019.06.014.
[25] Zhaohong Deng, Junyong Chen, Te Zhang, Longbing Cao, Shitong Wang, Generalized
Hidden</p>
      </sec>
      <sec id="sec-7-7">
        <title>Mapping Minimax Probability Machine for the training and reliability learning of several classical</title>
        <p>intelligent models, Information Sciences 436–437 (2018) 302-319. doi:10.1016/j.ins.2018.01.034.
[26] Jun Ma ,Jumei Shen, A novel twin minimax probability machine for classification and regression,</p>
        <p>Knowledge-Based Systems 196(2020) 105703. doi.org/10.1016/j.knosys.2020.105703.
[27] Khadiev K. The Quantum Version Of Classification Decision Tree Constructing Algorithm C5.0
/ K. Khadiev, I. Mannapov, L. Safina. – 2019.
[28] Z. Pawlak. Rough sets. International Journal of Information and Computer Sciences 11 (1982)
341–356.
[29] W. Xizhao . Learning with Uncertainty / W. Xizhao, Z. Junhai., 2016. – (Taylor &amp; Francis Groups).
[30] D.L. Sackett , W.M. Rosenberg, J.A Gray , R.B. Haynes RB, W.S Richardson , Evidence based
medicine: what it is and what it isn’t, BMJ, 312(7023),(1996) 71–72.
[31] Cardiovascular diseases, https://www.who.int/health-topics/cardiovascular-diseases/#tab=tab_1,
366 (Accessed on 11/24/2020).</p>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>I.G.</given-names>
            <surname>Kryvonos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>I.V.</given-names>
            <surname>Krak</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.V.</given-names>
            <surname>Barmak</surname>
          </string-name>
          ,
          <string-name>
            <surname>A.I. Kulias</surname>
          </string-name>
          ,
          <article-title>Methods to create systems for the analysis and synthesis of communicative information</article-title>
          ,
          <source>Cybern. Syst. Anal</source>
          <volume>53</volume>
          (
          <issue>6</issue>
          ), (
          <year>2017</year>
          )
          <fpage>847</fpage>
          -
          <lpage>856</lpage>
          . doi:
          <volume>10</volume>
          .1007/s10559-017-9986-7
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>I.</given-names>
            <surname>Krak</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Barmak</surname>
          </string-name>
          , E. Manziuk,
          <article-title>Using visual analytics to develop human and machine-centric models: A review of approaches and proposed information technology</article-title>
          ,
          <source>Computitional Intelligence</source>
          (
          <year>2020</year>
          )
          <fpage>1</fpage>
          -
          <lpage>26</lpage>
          . doi./10.1111/coin.12289
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>I.G.</given-names>
            <surname>Kryvonos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>I.V.</given-names>
            <surname>Krak</surname>
          </string-name>
          ,
          <article-title>Modeling human hand movements, facial expressions, and articulation to synthesize and visualize gesture information</article-title>
          ,
          <source>Cybernetics and Systems Analysis</source>
          <volume>47</volume>
          (
          <issue>4</issue>
          ) (
          <year>2011</year>
          )
          <fpage>501</fpage>
          -
          <lpage>505</lpage>
          . doi:
          <volume>10</volume>
          .1007/s10559-011-9332-4
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>I.G.</given-names>
            <surname>Kryvonos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>I.V.</given-names>
            <surname>Krak</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.V.</given-names>
            <surname>Barmak</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.S.</given-names>
            <surname>Ternov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.V.</given-names>
            <surname>Kuznetsov</surname>
          </string-name>
          ,
          <article-title>Information technology for the analysis of mimic expressions of human emotional states</article-title>
          ,
          <source>Cybernetics and Systems Analysis</source>
          ,
          <volume>51</volume>
          (
          <issue>1</issue>
          ) (
          <year>2015</year>
          )
          <fpage>25</fpage>
          -
          <lpage>33</lpage>
          . doi:
          <volume>10</volume>
          .1007/s10559-015-9693-1
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>I.V.</given-names>
            <surname>Krak</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.I.</given-names>
            <surname>Kudin</surname>
          </string-name>
          ,
          <string-name>
            <surname>A.I. Kulyas</surname>
          </string-name>
          ,
          <article-title>Multidimensional scaling by means of pseudoinverse operations</article-title>
          .
          <source>Cybernetics and Systems Analysis</source>
          ,
          <volume>55</volume>
          (
          <issue>1</issue>
          )(
          <year>2019</year>
          )
          <fpage>22</fpage>
          -
          <lpage>29</lpage>
          . (
          <year>2019</year>
          ).
          <source>doi:10.1007/s10559- 019-00108-9</source>
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <fpage>1</fpage>
          .
          <string-name>
            <given-names>O.</given-names>
            <surname>Bychkov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Merkulova</surname>
          </string-name>
          and
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhabska</surname>
          </string-name>
          ,
          <source>Software Application for Biometrical Person's Identification by Portrait Photograph Based on Wavelet Transform</source>
          ,
          <source>2019 IEEE International Conference on Advanced Trends in Information Theory (ATIT)</source>
          ,
          <year>2019</year>
          , pp.
          <fpage>253</fpage>
          -
          <lpage>256</lpage>
          , doi: 10.1109/ATIT49449.
          <year>2019</year>
          .9030462
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <fpage>2</fpage>
          .
          <string-name>
            <given-names>O.</given-names>
            <surname>Bychkov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Merkulova</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhabska</surname>
          </string-name>
          ,
          <source>Information Technology of Person's Identification by Photo Portrait</source>
          ,
          <source>2020 IEEE 15th International Conference on Advanced Trends in Radioelectronics</source>
          , Telecommunications and Computer Engineering (TCSET),
          <year>2020</year>
          , pp.
          <fpage>786</fpage>
          -
          <lpage>790</lpage>
          , doi: 10.1109/TCSET49122.
          <year>2020</year>
          .235542
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>O.</given-names>
            <surname>Bychkov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Ivanchenko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Merkulova</surname>
          </string-name>
          , Y. Zhabska,
          <article-title>Mathematical methods for information technology of biometric identification in conditions of incomplete data</article-title>
          ,
          <source>CEUR Workshop Proceedings of the 7th International Conference "Information Technology and Interactions" (IT&amp;I-2020)</source>
          ,
          <volume>2845</volume>
          ,
          <year>2021</year>
          , pp.
          <fpage>336</fpage>
          -
          <lpage>349</lpage>
          . URL: http://ceur-ws.
          <source>org/</source>
          Vol-
          <volume>2845</volume>
          /Paper_31.pdf
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>A. G.</given-names>
            <surname>Nakonechnyi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. B.</given-names>
            <surname>Kachinskiy</surname>
          </string-name>
          ,
          <article-title>Minimax parameter estimators of a linear regression with multiplicative noises</article-title>
          ,
          <source>Journal of Automation and Information Sciences</source>
          <volume>29</volume>
          (
          <year>1997</year>
          )
          <fpage>98</fpage>
          -
          <lpage>104</lpage>
          . doi:
          <volume>10</volume>
          .1615/jautomatinfscien.
          <source>v29.i2-3</source>
          .
          <fpage>130</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>J.</given-names>
            <surname>Michálek</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Nakonechny</surname>
          </string-name>
          ,
          <article-title>Minimax estimates of a linear parameter function in a regression model under restrictions on the parameters and variance-covariance matrix</article-title>
          ,
          <source>Journal of Mathematical Sciences</source>
          ,
          <volume>102</volume>
          (
          <issue>1</issue>
          ) (
          <year>2000</year>
          )
          <fpage>3790</fpage>
          -
          <lpage>3802</lpage>
          . doi:
          <volume>10</volume>
          .1007/bf02680236.
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>G.R.G.</given-names>
            <surname>Lanckriet</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.E.</given-names>
            <surname>Ghaoui</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Bhattacharyya</surname>
          </string-name>
          , and
          <string-name>
            <given-names>M.I.</given-names>
            <surname>Jordan</surname>
          </string-name>
          ,
          <article-title>A robust minimax approach to classification</article-title>
          ,
          <source>J. Mach. Learn. Res</source>
          ,
          <volume>3</volume>
          (
          <issue>3</issue>
          ) (
          <year>2003</year>
          ),
          <fpage>555</fpage>
          -
          <lpage>582</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>A. G.</given-names>
            <surname>Nakonechny</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V. P.</given-names>
            <surname>Marzeniuk</surname>
          </string-name>
          ,
          <source>Uncertainties in Medical Processes Control, Lecture Notes in Economics and Mathematical Systems</source>
          <volume>581</volume>
          (
          <year>2006</year>
          )
          <fpage>185</fpage>
          -
          <lpage>192</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <surname>G.R.G Lanckriet.</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.E.</given-names>
            <surname>Ghaoui</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Bhattacharyya</surname>
          </string-name>
          , et al. ,
          <source>Minimax probability machine, Neural Information Processing Systems (NIPS)</source>
          (
          <year>2001</year>
          )
          <fpage>801</fpage>
          -
          <lpage>807</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>