<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Application of Machine Algorithms for Classification and Formation of the Optimal Plan</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Nataliya</forename><surname>Boyko</surname></persName>
							<email>nataliya.i.boyko@lpnu.ua</email>
							<affiliation key="aff0">
								<orgName type="institution">Lviv Polytechnic National University</orgName>
								<address>
									<addrLine>Profesorska Street 1</addrLine>
									<postCode>79013</postCode>
									<settlement>Lviv</settlement>
									<country key="UA">Ukraine</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Rostyslav</forename><surname>Hlynka</surname></persName>
							<email>hlynka1608@gmail.com</email>
							<affiliation key="aff0">
								<orgName type="institution">Lviv Polytechnic National University</orgName>
								<address>
									<addrLine>Profesorska Street 1</addrLine>
									<postCode>79013</postCode>
									<settlement>Lviv</settlement>
									<country key="UA">Ukraine</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Application of Machine Algorithms for Classification and Formation of the Optimal Plan</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">2EE591CA88B0590141524C289FBF401E</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T11:53+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Data Mining</term>
					<term>Mathematical Programming</term>
					<term>Linear Programming</term>
					<term>Nonlinear Programming</term>
					<term>Quadratic Programming</term>
					<term>Problem of the Quadratic Programming</term>
					<term>Support Vector Machine</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>The paper presents three methods for data classification and finding the optimal plan: the study of the quadratic programming problem, the double problem and the Support Vector Machine method. It is known that linear programming is used to solve resource allocation problems. Also, its purpose is widely used to determine the highest profit or lowest cost, inventory management, the formation of an optimal transportation plan or to determine research, and so on. An important approach to the application of linear programming problems is the use of the duality principle, which is methodologically related to the theory of systems of dependent inequalities. This aspect better explains the concept of duality in linear programming problems with general mathematical rigor.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>Known methods of transition from a primal problem to a dual one are based on qualitative transformations and are meaningful. Formalization and proof of the correctness of the algorithm for constructing a dual problem for an arbitrary form of representation of a primal problem will make it easy to obtain correct pairs of known dual problems. The relevance of research is due to the requirements for simplification of solutions of linear programming problems based on the development of a formal algorithm for the transformation of a primal problem to a dual linear optimization problem <ref type="bibr" target="#b0">[1,</ref><ref type="bibr" target="#b5">6]</ref>.</p><p>Quadratic programming is a area of mathematical programming devoted to the theory of solving problems characterized by a quadratic dependency between variables <ref type="bibr" target="#b1">[2]</ref>. The usage of this method is relevant today, as the use of mathematical models is an important factor in improving the planning of the company. Mathematical representation of data allowed to create and model different options for choosing the optimal solution <ref type="bibr" target="#b2">[3,</ref><ref type="bibr" target="#b8">9,</ref><ref type="bibr" target="#b10">11]</ref>.</p><p>The paper considers the Support Vector Machine (SVM) method that is taught by examples and used to classify objects. It is established that SVM can be successfully used to control complex electromechanical systems, it can ensure the adaptability of control algorithms, perform the functions of an observer, an identifier of unknown parameters, a reference model, with its help you can control complex nonlinear objects, as well as objects with stochastic parameters <ref type="bibr" target="#b3">[4,</ref><ref type="bibr" target="#b9">10,</ref><ref type="bibr" target="#b16">17]</ref>.</p><p>The aim of the work is to solve a dual problem by SVM, the comparison with the primal problem and classification of the dataset.</p><p>Achieving this goal involves solving specific tasks:  determine the problem of the method of SVM for the dual problem;  compare dual SVM and primary;</p><p> analyze this method;  apply it in practice.</p><p>The object of research is to solve a dual problem by the method of SVM. The purpose of the study is to apply the problems of linear and nonlinear programming to study the properties of the studied problems, to determine their advantages and disadvantages.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Methods</head><p>Mathematical programming is an applied mathematical discipline that investigates the extremum of a function (maximum or minimum search problems) and develops methods for solving them. Such problems are also called optimization <ref type="bibr" target="#b4">[5,</ref><ref type="bibr" target="#b6">7]</ref>.</p><p>The area of mathematical programming can be applied to the type of objective function and to the system of constraints. As a result, we obtain a division into:  Linear programming -objective function and constraint functions included in the constraint system are linear (first order equation).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head></head><p>Nonlinear programming -the objective function or one of the constraint functions included in the constraint system is nonlinear (higher order equations).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head></head><p>Integer (discrete) programming -if at least one variable has an integer constaint. Dynamic programming -if the parameters of the objective function and / or system of constraints change over time or the objective function has an additive / multiplicative form or the decisionmaking process itself is multi-step <ref type="bibr" target="#b7">[8,</ref><ref type="bibr" target="#b11">12]</ref>.</p><p>Depending on that all the information about the process is known in advance, the field of mathematical programming is divided into:  Stochastic programming -not all information about the process is known in advance: the parameters included in the objective function or in the constraint function are random or have to make decisions in conditions of risk.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head></head><p>Deterministic programming -all information about the process is known in advance. Depending on the number of objective functions, the tasks are divided into:  Single-criteria;  Multicriteria. The optimization problem can be classified as follows: those problems that describe the properties of the constraint system and, accordingly, others that are determined by the objective function:</p><p> Unconditional optimization problems or problems without restrictions -they do not impose restrictions on quantitative variables.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head></head><p>Conditional optimization problems or constrained problems -in these problems, quantitative variables are constrained.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head></head><p>Optimization problems for incomplete data -they have a goal function or a system of constraints depend on some parameter p (numerical, vector), the value of which is completely undefined at the time of solving the problem. The first type includes optimization problems, the task of which is to minimize or maximize the quadratic function of several variables with linear constraints on these variables.</p><p>Quadratic programming problems include a special class of NP problems in which the objective function ) (x f is quadratic and concave (or convex), and all constraints are linear <ref type="bibr" target="#b12">[13,</ref><ref type="bibr" target="#b14">15]</ref>. Each linear programming problem can be matched to another that relates in some way to the original task. Such problems are called dual, or conjugate. Joint consideration of dual pairs of problems is very important in the economic analysis of the optimal plan. The correspondence between the original and dual problems is to build a dual problem on the basis of the first problem (as the source can be considered any of the conjugate pair of problems). Dual problems are symmetric and asymmetric.</p><p>The quadratic programming includes the SVM method. He constructs a model in the form of points in space using a binary linear classifier. This model goes through a series of iterations in which new patterns are displayed in a given space and determine the side of the gap. On the basis of these data the forecast of belonging of samples to a certain category is made.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Observation and analysis of the existing methods and means</head><p>Each linear programming problem corresponds to a dual, formed by certain rules directly from the condition of the primal problem. Comparing these two formulated problems, we conclude that the dual linear programming problem is formed from a primal problem by the following rules:</p><p>1. Each constraint of a primal problem corresponds to a variable of a dual problem. The number of unknowns of a dual problem is equal to the number of constraints of the primal problem. 2. For the primary problem, a certain variable corresponds to the specified constraint of the double problem and, accordingly, their number determines the number of unknowns in the primary problem.</p><p>3. If the objective function of the primary problem goes to max, then the objective function of the double problem goes to min, and vice versa. 4. In the objective function of a double problem, the coefficients of the variables are free values of the system of constraints for the primary problem. 5. The column of free members of the double problem is the coefficients for the variables in the objective function of the primary problem. 6. The coefficients of variables in the system of constraints of the primary problem are written in the matrix and accordingly it is transposed to determine the coefficients of constraints for the double problem.</p><p>As a result of intensive research in the field of machine learning, aimed at improving the quality of classifiers, a new generation of methods appeared, in particular -SVM. This method refers to machine learning methods based on vector spatial models, the purpose of which is to find dividing surfaces between classes as far as possible from all points of the study population (perhaps ignoring some points such as emissions or noise) <ref type="bibr" target="#b13">[14,</ref><ref type="bibr" target="#b15">16]</ref>.</p><p>If the training set contains two classes of data that allow linear division, then there are a large number of linear classifiers with which you can divide this data. It is intuitively clear that a dividing surface passing through the middle of a strip separating two classes. For example, the perceptron allows you to find at least one linear, other methods, such as the naive Bayesian method, find the best linear separator using a certain criterion. In particular, SVM necessarily assumes that the decisive function is completely determined by a subset of data that affect the position of the delimiter. In vector space, a point can be considered as a vector passing through the origin. Consider a dataset of n points in the form ) , ( ),..., , (</p><formula xml:id="formula_0">1 1 n n y x y x</formula><p>, where</p><formula xml:id="formula_1">  1 , 1 1   y</formula><p>identifies to which class the point x belongs.</p><p>Each point is a p-dimensional real vector. SVM means to find the maximum hyperplane that separates groups of points belonging to 1  y from 1   y . In Equation <ref type="formula" target="#formula_2">1</ref>we write a hyperplane through many points that satisfy the condition:</p><formula xml:id="formula_2">0 *   b x w , (<label>1</label></formula><formula xml:id="formula_3">)</formula><p>where w is a optional vector to the hyperplane. Parameter w b determines the displacement of the hyperplane from the origin on the normal vector (Figure <ref type="figure">1</ref>).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Figure 1: Visual representation of SVM</head><p>In Equation <ref type="formula" target="#formula_4">2</ref>we define the data that are not linearly separated:</p><formula xml:id="formula_4">)) ( 1 , 0 max( b wx y i i   . (<label>2</label></formula><formula xml:id="formula_5">)</formula><p>In Equation <ref type="formula" target="#formula_6">3</ref>, we minimize the function:</p><formula xml:id="formula_6">2 1 )) ( 1 , 0 max( 1 w b wx y n n i i i             ,<label>(3)</label></formula><p>where  determines the trade-off between the size of the margin and the guarantee that the point lies on the correct side of the margin. Hence, if  is very small, the second operand becomes insignificant, and the function will behave as with a hard margin. The calculation of the soft margin classifier is to minimize the expression of the form (Equation <ref type="formula" target="#formula_7">4</ref>):</p><formula xml:id="formula_7">2 1 ) ) ( 1 , 0 max( 1 w b wx y n n i i i             . (<label>4</label></formula><formula xml:id="formula_8">)</formula><p>Therefore, in a further study, we will consider a classifier with a bounded boundary.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1.">Primal problem</head><p>Equation 4 presents the minimization of a bounded optimization problem with a differentiated objective function. For each i we enter a variable i e -the least positive number that satisfies</p><formula xml:id="formula_9">i i i e b x w y   ) * ( and 0  i e .</formula><p>Equation 5 presents the problem of optimization taking into account additions. (</p><p>Solving the primary problem for the dual Lagrange, we obtain a simplified problem (Equation <ref type="formula" target="#formula_11">6</ref>):</p><formula xml:id="formula_11">        n i n j j j j i i i n i i n c y x x c y c c 1 1 1 1 ) * ( 2 1 ) ,..., maxmizef(c , (<label>6</label></formula><formula xml:id="formula_12">) if:    n j i i c y 1 0 and  n c i 2 1 0   .</formula><p>The quadratic function solves double maximization problems. its results satisfy linear constraints. Equations 7-8 determine the variables i c to determine the second problem.  <ref type="formula" target="#formula_15">9</ref>shows a linear combination of reference vectors, which determines the offset through a point on the field boundary.</p><formula xml:id="formula_13">        n i n j j j j i i i n i i n c y x x c y c c 1 1 1 1 ) * ( 2 1 ) ,..., maxmizef(c . (<label>7</label></formula><formula xml:id="formula_14">) i n i i i x c y w    1 . (<label>8</label></formula><formula xml:id="formula_15">i i i i y wx b b wx y      1 ) ( . (<label>9</label></formula><formula xml:id="formula_16">)</formula></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.">Comparison of problems</head><p>Equation (Formula 8) gives the optimal value of w in terms of с . Suppose we have adjusted the parameters of our model to the training set, and now we want to make a prediction for the new point</p><p>x . Then we will calculate b x w T  and forecast 1  y , if and only if this value is greater than zero. But, using (Formula 3), this value can also be written as (Equation <ref type="formula" target="#formula_17">10</ref>):</p><formula xml:id="formula_17">         n i i i i n i T i i i T b x x c y b x x c y b x w 1 1 ) , ( )) ( ( . (<label>10</label></formula><formula xml:id="formula_18">)</formula><p>So, if we find i c to make a prediction, we have to calculate a value that depends only on the internal product between</p><p>x . Moreover, we have previously seen that i c will be equal to all but zero support vectors. Thus, many terms in the above sum will be zero, and we really only need to find the internal products between</p><p>x and the reference vectors (which are often only a small number) to calculate (Formula 9) and make our prediction.</p><p>Considering the dual form of the optimization problem, we got a good idea of the structure of the problem, and we can write the whole algorithm in terms of only the internal products between the vectors of the input functions. This property is important to apply kernels to our classification problem. The obtained algorithm, supporting vector machines, will be able to effectively learn in spaces with high dimensions.</p><p>Also dual SVM requires fewer kernel estimates than the primary. Therefore, it gives a more stable result in less computational time (Table <ref type="table" target="#tab_0">1</ref>). In terms of stable convergence and learning speed, a dual SVM is better than a basic SVM.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Experiments</head><p>The study requires solving a double problem by the method of reference vectors, comparison with the primary problem and classification of the data set.</p><p>Achieving this goal involves solving specific tasks:  determine the problem of the method of support vectors for the dual problem;  compare dual SVM and primal;  analyze the algorithm of the method;  apply it in practice.</p><p>The Iris dataset was chosen for the implementation of the support vector method. This is a wellknown set of data used in the area of machine learning.</p><p>Dataset attribute information: The main purpose of visualizationinterpretation of a large data set into visual graphics to easily understand complex data relationships and quickly get an imagination of the dataset.</p><p>The histogram shown in Figure <ref type="figure">2</ref> shows the frequencies of the data set, which can be used to understand the trend of length / width of the petals for each type of plant.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Figure 2: Histogram of frequencies</head><p>From Figure <ref type="figure">2</ref> it is seen that the type of petal virginica is much larger than others and sepals are mostly too, setosa has the smallest size and small range of values, versicolor -medium in size.</p><p>The box plot (Figure <ref type="figure" target="#fig_2">3</ref>) shows the distribution of data by quartiles, the average values and statistical emissions are highlighted. The vertical lines (whiskers) drawn to the rectangles reflect the variability of the values outside the upper and lower quartiles, any point on these lines is considered a statistical ejection. The other two species are larger and have larger tendrils, which characterize a large scope.</p><p>The heatmap (Figure <ref type="figure" target="#fig_3">4</ref>) shows the levels of correlation between attributes and their linear dependencies:</p><p> (-0.09; 0.0)(0.0; 0.09) -linear independence;  (-0.3; -0.1)(0.1; 0.3) -low linear dependence;  (-0.5; -0.3) (0.3; 0.5)medium linear dependence;  (-1.0; -0.5) (0.5; 1.0) -high linear dependence. The values of the length of the sepals depend entirely on the values of its width, or vice versa. In the case of a petal, the length and width are independent of each other. It is also interesting that the length of the petal depends on the length and width of the sepals.</p><p>In Figure <ref type="figure" target="#fig_4">5</ref> you can see the linear separation between classes. Setosa is linearly separated from Versicolor and Virginica, and Versicolor and Virginica are not linearly separated. It means that in the first case it is necessary to apply SVM with a hard margin, and in the second -with soft.    From this statistical in the Table <ref type="table" target="#tab_3">2</ref> it is seen: that the size of the sepals mostly is larger than the size of the petal. Also, since the variance and standard deviation characterize the scattering of values around the distribution center, it can be concluded that the variation of petal size values between plant species is much larger than for the sepals.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Results</head><p>Two arbitrary points from the iris dataset are selected for analysis, and the optimal hyperplane is calculated.</p><p>Let first arbitrary point - , where xindex in the dataset, y - length of the sepal (Figure <ref type="figure" target="#fig_6">6</ref>).</p><p>The point A will be considered negative, You need to find the optimal separate hyperplane.</p><formula xml:id="formula_19">B -positive, thus 1  yA , 1   yB .</formula><p>It is known that any hyperplane can be described as (equation 11</p><formula xml:id="formula_20">): 0   b wx ,<label>(11)</label></formula><p>where wnormal vector to the hyperplane and w b -perpendicular distance from the hyperplane to the origin.</p><p>To find w and b dual form should be introduced. It contains a quadratic objective function with constraints (equation 12). </p><formula xml:id="formula_21">j i j i j i j i L i i D a x x y y a a a L , , 1 2 1 max      , (<label>12</label></formula><formula xml:id="formula_22">) if i a y a i L i i i      0 0 1 ,</formula><formula xml:id="formula_23">1 max 2 2 2 1 2 1 2 1 2 2 2 1 1 1 2 1 , 1 , a a a a a a a a a a a a a a a a i a L j i j i j i j i L i D a x x y y                                                                    </formula><p>Using the Lagrange equation we can solve this problem:</p><formula xml:id="formula_24">) 0 ( * 69 . 10039 48 . 2460 04 . 167 ( 2 1 ) , ( 2 1 2 2 2 1 2 1 2 1         x x x x x x x x X L   .</formula><p>Under the condition of the extremum of the Lagrange function, we equate the partial derivatives to zero.</p><p>Built system:  This system can now be written as:   Figure <ref type="figure" target="#fig_10">7</ref> shows the hyperplane that best classifies our data, and theoretically all positive points will be on the left and negative points on the right. Dark blue color indicates the hyperplane itself, dotted lines form a margin. Figure <ref type="figure" target="#fig_11">8</ref> shows the calculated results using the program for the entire dataset:  )), so the support vectors and their number for each class are found correctly. Plants with an index less than 75 will belong to setosa, and more -to versicolor. But it is necessary to remember about possible cases where the sepal_length value will be considered.</p><formula xml:id="formula_25">                             0 0 1</formula><formula xml:id="formula_26">            </formula><formula xml:id="formula_27">                                                        <label>96837</label></formula><formula xml:id="formula_28">,                                                     L S m</formula></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.">Conclusion</head><p>The dual method of support vectors is to solve the Lagrange problem, with found Lagrange multipliers it is easy to calculate the normal vector w and draw a separate hyperplane.</p><p>Solving the primary problem, it is obtained the optimal w, but nothing about the Lagrange multipliers. To classify the point x, it is needed to clearly calculate the scalar product x w T , which can be very expensive. Solving the dual problem: obtained Lagrange factors (where 0  i a for all but a few points -support vectors). This problem is very efficiently calculated if there are few support vectors. Also, with a scalar product that includes only data vectors, it is possible to use kernel trick for nonlinear problems.</p><p>Dual SVM in nonlinear problems is more stable and faster than the primary because it performs fewer kernel estimates.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head></head><label></label><figDesc>point lies on the right side of the field and edge of the field. Equation</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>Figure 3 :</head><label>3</label><figDesc>Figure 3: Box plot</figDesc><graphic coords="7,79.20,97.30,436.55,267.85" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_3"><head>Figure 4 :</head><label>4</label><figDesc>Figure 4: Heatmap</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_4"><head>Figure 5 :</head><label>5</label><figDesc>Figure 5: Scatter plot</figDesc><graphic coords="7,184.10,555.95,240.95,186.15" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_6"><head>Figure 6 :</head><label>6</label><figDesc>Figure 6: Visualization of selected points</figDesc><graphic coords="8,208.10,472.20,192.95,175.70" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_7"><head></head><label></label><figDesc>multiply the first row by 1230;  multiply the second row by 167;  add the second row to the first. multiply the third row by (-1230);  add the third row to the second. multiply the first row by (-0.0538);  add the second row to the first.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_8"><head></head><label></label><figDesc>Calculating w and b :</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_9"><head></head><label></label><figDesc>Returning to the representation of the hyperplane and substitute the numbers:</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_10"><head>Figure 7 :</head><label>7</label><figDesc>Figure 7: Visualization of the hyperplane</figDesc><graphic coords="11,199.80,72.00,209.50,191.05" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_11"><head>Figure 8 :</head><label>8</label><figDesc>Figure 8: Visualization of the dataset</figDesc><graphic coords="11,199.90,492.95,209.20,184.45" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_12"><head>Figure 9 :</head><label>9</label><figDesc>Figure 9: Visualization of the hyperplane</figDesc><graphic coords="12,202.10,275.15,204.80,186.75" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0"><head></head><label></label><figDesc></figDesc><graphic coords="6,138.95,302.05,331.10,189.95" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1</head><label>1</label><figDesc>Comparison between primal and dual SVMs on the different datasets</figDesc><table><row><cell>SVM</cell><cell>Data</cell><cell>SVs</cell><cell>Iterations</cell><cell>Kernels</cell><cell>Rec.</cell><cell>Time</cell></row><row><cell></cell><cell>Thyroid</cell><cell>98</cell><cell>1439</cell><cell>2,660,742,540</cell><cell>97.81 (100)</cell><cell></cell></row><row><cell></cell><cell>Blood cell</cell><cell>188</cell><cell>13</cell><cell>86,324,885</cell><cell>93.58 (97.19)</cell><cell>10</cell></row><row><cell></cell><cell>H-50</cell><cell>77</cell><cell>20</cell><cell>386,391,968</cell><cell>99.28 (100)</cell><cell>50</cell></row><row><cell>Dual</cell><cell>H-13</cell><cell>39</cell><cell>83</cell><cell>2,343,986,783</cell><cell>99.55 (100)</cell><cell></cell></row><row><cell></cell><cell>H-105</cell><cell>91</cell><cell>22</cell><cell>812,259,183</cell><cell>100 (100)</cell><cell></cell></row><row><cell></cell><cell>Satimage</cell><cell>1001</cell><cell>28</cell><cell>600,106,924</cell><cell>91.70 (100)</cell><cell></cell></row><row><cell></cell><cell>USPS</cell><cell>597</cell><cell>19</cell><cell>593,529,638</cell><cell>95.47 (100)</cell><cell></cell></row><row><cell></cell><cell>Thyroid</cell><cell></cell><cell></cell><cell>No convergence</cell><cell></cell><cell></cell></row><row><cell></cell><cell>Blood cell</cell><cell>203</cell><cell>10</cell><cell>445,319,582</cell><cell>93.61 (97.19)</cell><cell>21</cell></row><row><cell></cell><cell>H-50</cell><cell>70 (78)</cell><cell>15</cell><cell>605,637,116</cell><cell>99.28 (100)</cell><cell>52</cell></row><row><cell>Primal</cell><cell>H-13</cell><cell>99</cell><cell>12</cell><cell>749,712,635</cell><cell>99.70 (99.96)</cell><cell>93</cell></row><row><cell></cell><cell>H-105</cell><cell>111</cell><cell>13</cell><cell>907,360,383</cell><cell>100 (100)</cell><cell></cell></row><row><cell></cell><cell>Satimage</cell><cell>1006</cell><cell>25</cell><cell>26,125,955,619</cell><cell>91.70 (99.71)</cell><cell>1258</cell></row><row><cell></cell><cell>USPS</cell><cell>604</cell><cell>16</cell><cell>8,116,273,966</cell><cell>95.47 (99.99)</cell><cell></cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 1</head><label>1</label><figDesc></figDesc><table /><note>shows the results of the primal and dual SVM, using mark datasets. Contains 7 columns:  SVMthe type of problem of the used method of support vectors;  Datathe name of the used dataset;  SVssolution function to determine the number of vectors;  Iterationsnumber of steps;  Kernelsthe number of calls to the kernel;  Rec.recognition accuracy;  Timetraining time.</note></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head></head><label></label><figDesc>The best way to analyze a large data set is to visualize it. Data visualization refers to the approaches used to understand data through visual representation.</figDesc><table><row><cell>1. Sepallength.</cell></row><row><cell>2. Sepalwidth.</cell></row><row><cell>3. Petallength.</cell></row><row><cell>4. Petalwidth.</cell></row><row><cell>5. Classes:</cell></row><row><cell> Iris Setosa;</cell></row><row><cell> Iris Versicolour;</cell></row><row><cell> Iris Virginica.</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_3"><head>Table 2</head><label>2</label><figDesc>Statistic of the iris dataset</figDesc><table><row><cell></cell><cell></cell><cell></cell><cell cols="10">Formula</cell><cell></cell><cell></cell><cell>sepallength</cell><cell>sepalwidth</cell><cell>petallength</cell><cell>petalwidth</cell></row><row><cell>Amount</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell cols="2">n</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell>150</cell><cell>150</cell><cell>150</cell><cell>150</cell></row><row><cell>Mean value</cell><cell></cell><cell></cell><cell></cell><cell cols="6">n i  1</cell><cell>x</cell><cell>i</cell><cell></cell><cell></cell><cell></cell><cell>5.84</cell><cell>3.057</cell><cell>3.758</cell><cell>1.19</cell></row><row><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell cols="2">n</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell>Dispersion</cell><cell></cell><cell cols="2">n i 1</cell><cell cols="2">(</cell><cell cols="2">x</cell><cell>i</cell><cell cols="2"></cell><cell cols="2">x</cell><cell cols="2">)</cell><cell>2</cell><cell>0.6724</cell><cell>0.19</cell><cell>3.115</cell><cell>0.578</cell></row><row><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell>j</cell><cell></cell><cell></cell></row><row><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell cols="2">n</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell>Standard deviation</cell><cell cols="2"></cell><cell cols="2">n i 1</cell><cell cols="2">(</cell><cell cols="2">x</cell><cell>i</cell><cell cols="2"></cell><cell>x</cell><cell>j</cell><cell cols="2">)</cell><cell>2</cell><cell>0.828</cell><cell>0.436</cell><cell>1.765</cell><cell>0.762</cell></row><row><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell cols="2">n</cell><cell></cell><cell></cell><cell></cell><cell></cell></row></table></figure>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Using the Lagrange function helps to distribute the data linearly. The kernel trick is used to separate data in different ways, but not line.</p><p>SVM can be successfully used to control complex electromechanical systems, it can ensure the adaptability of control algorithms, perform the functions of an observer, an identifier of unknown parameters, a reference model, it can be used to control complex nonlinear objects.</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Amoeba: Hierarchical clustering based on spatial proximity using Delaunay diagram</title>
		<author>
			<persName><forename type="first">V</forename><surname>Estivill-Castro</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Lee</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">9th Intern. Symp. on spatial data handling</title>
				<meeting><address><addrLine>Beijing, China</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2000">2000</date>
			<biblScope unit="page" from="26" to="41" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Density connected clustering with local subspace preferences</title>
		<author>
			<persName><forename type="first">C</forename><surname>Boehm</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Kailing</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Kriegel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Kroeger</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. of the 4th IEEE Intern. conf. on data mining</title>
				<meeting>of the 4th IEEE Intern. conf. on data mining<address><addrLine>Los Alamitos</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2004">2004</date>
			<biblScope unit="page" from="27" to="34" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Information system of catering selection by using clustering analysis</title>
		<author>
			<persName><forename type="first">N</forename><surname>Boyko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Shakhovska</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">IEEE Ukraine Student, Young Professional and Women in Engineering Congress (UKRSYW)</title>
				<meeting><address><addrLine>Kyiv, Ukraine</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2018">2018. 2018</date>
			<biblScope unit="page" from="7" to="13" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Clustering spatial data using random walks</title>
		<author>
			<persName><forename type="first">D</forename><surname>Harel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Koren</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. of the 7th ACM SIGKDD Intern. conf. on knowledge discovery and data mining</title>
				<meeting>of the 7th ACM SIGKDD Intern. conf. on knowledge discovery and data mining<address><addrLine>San Francisco, California</addrLine></address></meeting>
		<imprint>
			<biblScope unit="volume">200</biblScope>
			<biblScope unit="page" from="281" to="286" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Spatial clustering in the presence of obstacles</title>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">K</forename><surname>Tung</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Hou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Han</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">The 17th Intern. conf. on data engineering (ICDE&apos;01)</title>
				<meeting><address><addrLine>Heidelberg</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2001">2001</date>
			<biblScope unit="page" from="359" to="367" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Application of Artificial Intelligence Algorithms for Image Processing</title>
		<author>
			<persName><forename type="first">N</forename><surname>Boyko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Bronetskyi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Shakhovska</surname></persName>
		</author>
		<idno>nbn: de: 0074-2386-1</idno>
	</analytic>
	<monogr>
		<title level="m">CEUR. Workshop Proceedings of the 8th International Conference on &quot;Mathematics. Information Technologies. Education</title>
				<meeting><address><addrLine>MoMLeT&amp;DS; Shatsk, Ukraine</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2019-06-02">2019. June 2-4, 2019</date>
			<biblScope unit="volume">2386</biblScope>
			<biblScope unit="page" from="194" to="211" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Automatic sub-space clustering of high dimensional data</title>
		<author>
			<persName><forename type="first">R</forename><surname>Agrawal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Gehrke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Gunopulos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Raghava</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Data mining knowledge discovery</title>
		<imprint>
			<biblScope unit="volume">11</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page" from="5" to="33" />
			<date type="published" when="2005">2005</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Towards an effective cooperation of the user and the computer for classification</title>
		<author>
			<persName><forename type="first">M</forename><surname>Ankerst</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Ester</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H.-P</forename><surname>Kriegel</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. of the 6th ACM SIGKDD Intern. conf. on knowledge discovery and data mining</title>
				<meeting>of the 6th ACM SIGKDD Intern. conf. on knowledge discovery and data mining<address><addrLine>Boston, Massachusetts, USA</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2000">2000</date>
			<biblScope unit="page" from="179" to="188" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Testing local spatial autocorrelation using</title>
		<author>
			<persName><forename type="first">С</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Murayama</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Intern. J. of Geogr. Inform. Science</title>
		<imprint>
			<biblScope unit="volume">14</biblScope>
			<biblScope unit="page" from="681" to="692" />
			<date type="published" when="2000">2000</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Amoeba: Hierarchical clustering based on spatial proximity using Delaunay diagram</title>
		<author>
			<persName><forename type="first">V</forename><surname>Estivill-Castro</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Lee</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">9th Intern. Symp. on spatial data handling</title>
				<meeting><address><addrLine>Beijing, China</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2000">2000</date>
			<biblScope unit="page" from="26" to="41" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">ICEAGE: Interactive clustering and exploration of large and high-dimensional geodata</title>
		<author>
			<persName><forename type="first">D</forename><surname>Guo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">J</forename><surname>Peuquet</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Gahegan</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Geoinformatica</title>
		<imprint>
			<biblScope unit="volume">3</biblScope>
			<biblScope unit="issue">7</biblScope>
			<biblScope unit="page" from="229" to="253" />
			<date type="published" when="2003">2003</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Comparison Of Machine Learning Libraries Performance Used For Machine Translation Based On Recurrent Neural Networks</title>
		<author>
			<persName><forename type="first">N</forename><surname>Boyko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Basystiuk</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">2018 IEEE Ukraine Student, Young Professional and Women in Engineering Congress (UKRSYW)</title>
				<meeting><address><addrLine>Kyiv, Ukraine</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2018">2018</date>
			<biblScope unit="page" from="78" to="82" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Finding generalized projected clusters in high dimensional spaces</title>
		<author>
			<persName><forename type="first">C</forename><surname>Aggarwal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Yu</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">ACM SIGMOD Intern. conf. on management of data</title>
				<imprint>
			<date type="published" when="2000">2000</date>
			<biblScope unit="page" from="70" to="81" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">Application of Machine Learning Algorithms for Classification and Security of Diagnostic Images</title>
		<author>
			<persName><forename type="first">R</forename><surname>Thanki</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Borra</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Machine Learning in Bio-Signal Analysis and Diagnostic Imaging</title>
		<imprint>
			<biblScope unit="page" from="273" to="292" />
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<monogr>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">J</forename><surname>Peuquet</surname></persName>
		</author>
		<title level="m">Representations of space and time</title>
				<meeting><address><addrLine>N. Y.</addrLine></address></meeting>
		<imprint>
			<publisher>Guilford Press</publisher>
			<date type="published" when="2000">2000</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">A Monte Carlo algorithm for fast projective clustering</title>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">M</forename><surname>Procopiuc</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Jones</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><forename type="middle">K</forename><surname>Agarwal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">M</forename><surname>Murali</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Intern. conf. on management of data</title>
				<meeting><address><addrLine>Madison, Wisconsin, USA</addrLine></address></meeting>
		<imprint>
			<publisher>ACM SIGMOD</publisher>
			<date type="published" when="2002">2002</date>
			<biblScope unit="page" from="418" to="427" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">A Comparative Study of Various Clustering Algorithms in Data Mining</title>
		<author>
			<persName><forename type="first">K</forename><surname>Chitra</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Dr</forename><forename type="middle">D</forename><surname>Maheswari</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">International Journal of Computer Science and Mobile Computing</title>
		<imprint>
			<biblScope unit="volume">6</biblScope>
			<biblScope unit="issue">8</biblScope>
			<biblScope unit="page" from="109" to="115" />
			<date type="published" when="2017-08">August 2017</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
