<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">The Regression Analysis of the Data to Determine the Buffer Size When Serving a Self-Similar Packets Flow</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Gennadiy</forename><surname>Linets</surname></persName>
							<affiliation key="aff0">
								<orgName type="institution">North Caucasus Federal University</orgName>
								<address>
									<addrLine>2 Kulakova str</addrLine>
									<postCode>355029</postCode>
									<settlement>Stavropol</settlement>
									<country key="RU">Russia</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Roman</forename><surname>Voronkin</surname></persName>
							<email>roman.voronkin@gmail.com</email>
							<affiliation key="aff0">
								<orgName type="institution">North Caucasus Federal University</orgName>
								<address>
									<addrLine>2 Kulakova str</addrLine>
									<postCode>355029</postCode>
									<settlement>Stavropol</settlement>
									<country key="RU">Russia</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Svetlana</forename><surname>Govorova</surname></persName>
							<affiliation key="aff0">
								<orgName type="institution">North Caucasus Federal University</orgName>
								<address>
									<addrLine>2 Kulakova str</addrLine>
									<postCode>355029</postCode>
									<settlement>Stavropol</settlement>
									<country key="RU">Russia</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Ilya</forename><surname>Palkanov</surname></persName>
							<affiliation key="aff0">
								<orgName type="institution">North Caucasus Federal University</orgName>
								<address>
									<addrLine>2 Kulakova str</addrLine>
									<postCode>355029</postCode>
									<settlement>Stavropol</settlement>
									<country key="RU">Russia</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Carlos</forename><surname>Grilo</surname></persName>
							<email>carlos.grilo@ipleiria.pt</email>
							<affiliation key="aff1">
								<orgName type="institution">Instituto Polit´écnico de Leiria</orgName>
								<address>
									<addrLine>Rua General Norton de Matos</addrLine>
									<postBox>Apartado 4133</postBox>
									<postCode>2411-901</postCode>
									<settlement>Leiria</settlement>
									<country key="PT">Portugal</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">The Regression Analysis of the Data to Determine the Buffer Size When Serving a Self-Similar Packets Flow</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">0E0CAE53EB45A1BDAE9F47EC9F41FEB3</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T05:32+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Telecommunication network, self-similar traffic, Hurst exponent, Pareto distribution, packet loss, regression analysis, quality metrics, penalty score, machine learning 0000-0002-2279-3887 (Gennadiy Linets)</term>
					<term>0000-0002-7345-579X (Roman Voronkin)</term>
					<term>0000-0002-3225-1088 (Svetlana Govorova)</term>
					<term>0000-0003-0751-3928 (Ilya Palkanov)</term>
					<term>0000-0001-9727-905X (Carlos Grilo)</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Using the methods of regression analysis on the basis of simulation data, a model for predicting the queue size of the input self-similar packet flow, distributed according to the Pareto law when it is transformed into a flow having an exponential distribution, is constructed. Since the amount of losses in the general case does not give any information about the efficiency of using the buffer memory space in the process of transforming a self-similar packet flow, a quality metric (penalty) was introduced to get the quality of the models after training, which is a complex score. This criterion considers both packet loss during functional transformations and ineffective use of the buffer space in switching nodes. The choice of the best model for predicting the queue size when servicing a self-similar packet flow was carried out using the following characteristics: the coefficient of determination; root-mean-square regression error; mean absolute error; the penalty score. The best in terms of the investigated characteristics are the models using the isotonic regression and the support vector regression.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>The main reason leading to a buffer overflow is the presence of a long-term dependence in network traffic due to its self-similarity, as a result of which the total cumulative effect in a wide range of delays can significantly differ from that observed in a short-term dependent process <ref type="bibr" target="#b0">[1]</ref>. To eliminate self-similarity of network traffic, various models and traffic transformation devices are used, one of which is the asynchronous simulation model described in <ref type="bibr" target="#b2">[2]</ref><ref type="bibr" target="#b3">[3]</ref><ref type="bibr" target="#b4">[4]</ref>, for which there is a software implementation <ref type="bibr" target="#b5">[5]</ref>.</p><p>An important indicator of the operation of this model is the queue size used in the traffic transformation process. Since, due to limited computer resources, the queue cannot have an infinite size, the problem arises of predicting the queue size depending on the measure of self-similarity of the input traffic, which is the Hurst exponent.</p><p>The solution to the problem of finding the optimal buffer size for a given value of the Hurst exponent H can be found using the methods of regression analysis, based on simulation data obtained using the developed software <ref type="bibr" target="#b5">[5]</ref>.</p><p>Since machine learning includes many methods, at the initial stage, for further comparison with more complex models built, in particular, using deep learning methods, it is advisable to consider only methods of pairwise regression analysis, isotonic regression and support vector machines.</p><p>Let us involve a quality metric (penalty), which is a complex score and considers both packet loss during traffic transformation and inefficient use of buffer space.</p><p>Next, we choose the best model for predicting the queue size, depending on the Hurst exponent of the input flow, using the following quality metrics:  coefficient of determination;  root mean square error of regression;  mean absolute error;  penalty score value.</p><p>When setting the problem, special attention should be paid to testing the resulting models. In this case, the classical approach, which consists in dividing the entire data set into training and test samples, is not acceptable. Since on the test sample we will get the estimated number of lost packets and, therefore, the estimated penalty based on the difference between the predicted and actual buffer sizes, the obtained models must be tested by simulating traffic transformation with a queue size limitation, based on the results of applying the tested model for determining the size of the queue to the sequence being converted. This task is not trivial and will not be discussed in this article.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">The solution of the problem</head><p>The simulation model presented in <ref type="bibr" target="#b2">[2]</ref> provides transformation of the input flow of packets, which is obviously self-similar, into a given distribution law, in particular, into an exponential one. The object of transformation is a one-dimensional distribution density of time intervals between packets of the input flow. Using the developed model, 11,000 tests were carried out and data were obtained for statistical analysis.</p><p>Since the amount of losses in the general case does not give any information about the efficiency of using the queue in the process of the transformation traffic, to assess the quality of the resulting model, we introduce a quality metric -the penalty score, which takes into account not only the amount of losses, but also not rational use of buffer memory.</p><p>Let define i y as the true value of the queue size in the sample, ˆi y is the predicted value of the queue size in the sample corresponding to the true value i y . If îi yy  , we will penalize the learning system by   îi yy   . If îi yy  , the amount of the penalty will depend on the value of the difference</p><formula xml:id="formula_0">î i i yy    ,</formula><p>with i    the amount of the penalty will be   i       and 0 -otherwise. Let us illustrate this with an example (Figure <ref type="figure" target="#fig_0">1</ref>). In the second case 2 ŷy      and the penalty is 0, it is assumed that ŷ is the preferred queue size for 2 y . In the third case 3 ŷ y    , the amount of the penalty is determined from the expression</p><formula xml:id="formula_1">  3 ŷy       .</formula><p>Thus, the amount of the penalty score will be determined from the equation:</p><formula xml:id="formula_2">    ˆ, if , ˆ, if , 0, otherwise. i i i i i i i i y y y y p y y y y                    </formula><p>The total penalty for all trials is determined as the arithmetic mean between the penalties for each trial:</p><formula xml:id="formula_3">1 1 n i i pp n   </formula><p>where n is the number of tests. In the process of training the model, it is necessary to ensure the minimum value of the penalty for all tests, in other words min p  . The presented system of penalties provides for the introduction of three hyperparameters: ,   and  , where 0</p><p>   and 0  . Let's set the hyperparameter values as follows:</p><p>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1.">The initial data analysis</head><p>Figure <ref type="figure" target="#fig_1">2</ref> shows a scatter plot of queue size in dependence of Hurst exponent. The figure clearly shows that there is a certain correlation between the Hurst exponent and the buffer size <ref type="bibr" target="#b4">[4]</ref>.</p><p>Let us first group the tests by the value of the Hurst exponent and then select 30 groups to estimate the spread of the queue size.  Next, we can build a box-plot for each group. It follows from Figure <ref type="figure" target="#fig_2">3</ref> that the largest amount of outliers from above is observed for the first 10 groups, which corresponds to the Hurst exponent value close to 0.5. Consequently, at these values of the Hurst exponent, losses may occur due to the fact that the required buffer size will be greater than the predicted one.</p><p>For groups from 28 to 30, there are significant outliers from the bottom, which leads to inefficient use of buffer memory.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.">The regression analysis</head><p>Machine learning is a subset of artificial intelligence that studies and explores algorithms that can learn without direct programming. Linear regression is a typical representative of machine learning algorithms <ref type="bibr" target="#b7">[7]</ref>.</p><p>There are the following tasks solved by machine learning: supervised learning, unsupervised learning, reinforcement learning, active learning, knowledge transfer, etc. Regression (as well as classification) belongs to the class of supervised learning problems, when a certain target variable must be predicted for a given set of features of the observed object. As a rule, in supervised learning problems, experience E is represented as a set of pairs of features and target variables:</p><formula xml:id="formula_4">    , 1... ii i D x y n </formula><p>. In the case of linear regression, the feature description of an object is a real vector m xR  , where R is the set of real numbers and the target variable is a scalar yR  . The simplest measure of the quality L for the regression problem is</p><formula xml:id="formula_5">    2</formula></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>ˆ,, L y y y y </head><p>where ŷ is an estimate of the real value of the target variable <ref type="bibr" target="#b7">[7,</ref><ref type="bibr" target="#b8">8]</ref>.</p><p>Let us restore the dependence shown in Figure <ref type="figure" target="#fig_1">2</ref> using the methods of regression analysis. The basis of regression analysis is the method of least squares (OLS), according to which the function   y f x  is taken as the regression equation such that the sum of the squares of the differences would satisfy</p><formula xml:id="formula_6">  2 1 ˆmin. n ii i S y y     </formula><p>Using the methods of pairwise regression analysis, we will carry out a statistical analysis of the data obtained by transforming an input self-similar flow distributed according to the Pareto law into a flow having an exponential distribution. Let us examine the methods widely used in practice, which allow finding the buffer size for the input flow with a given Hurst exponent.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.3.">The linear regression analysis</head><p>In this case, the relationship between the Hurst exponent H and queue size ŷ is determined according to the linear equation: (1) The result of the fitting for the current model is shown on Figure <ref type="figure" target="#fig_4">4</ref>. Thus we obtained the statistically significant result. In Table <ref type="table" target="#tab_0">1</ref> the quality metrics values are shown for the obtained linear regression model. The obtained value of the coefficient of determination suggests that only about 58% of cases of changes in the Hurst exponent lead to a change in the size of the queue within the framework of the linear model. The obtained result is unsatisfactory for practice, therefore, in the simplest case, it makes sense to consider other methods using the methods of linearization of nonlinear dependencies. As a result, the nonlinear dependence can be reduced to linear, and then, the least squares method can be used.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.4.">The hyperbolic regression</head><p>For the hyperbolic regression, the relationship between H and ŷ can be described as follows: We obtain the regression equation using the least squares method:</p><formula xml:id="formula_7">H y 379 . 489 438 . 875 ˆ  (2)</formula><p>The result of the fitting for the current model is shown on Figure <ref type="figure">5</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Figure 5:</head><p>The first hyperbolic model report is built using the statsmodels package of the Python programming language</p><p>Thereby, we obtained the statistically significant result. In Table <ref type="table" target="#tab_1">2</ref> the quality metrics values are shown for the obtained first hyperbolic regression model. The obtained value of the coefficient of determination suggests that about 45% of cases of changes in the Hurst exponent lead to a change in the size of the queue. This is much worse than the value of the coefficient of determination of the linear model. For this reason, it makes sense to consider a different hyperbolic regression model:  The result of the fitting for the current model is shown on Figure <ref type="figure">6</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Figure 6:</head><p>The second hyperbolic model report is built using statsmodels package of the Python programming language Accordingly we obtained the statistically significant result. In Table <ref type="table" target="#tab_2">3</ref> the quality metrics values are shown for the obtained second hyperbolic regression model. The obtained value of the coefficient of determination is about 59%, which is slightly better than the linear model.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.5.">The power regression</head><p>In the case of the power regression, the relationship between H and ŷ is: The exponential function is internally linear, therefore, estimates of the unknown parameters of its linearized form can be calculated using the classical least squares method. The regression equation is:</p><p>(4) The result of the fitting for the current model is shown on Figure <ref type="figure">7</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Figure 7:</head><p>The power model report is built using statsmodels package of the Python programming language Thus we obtained the statistically significant result. In Table <ref type="table" target="#tab_3">4</ref> the quality metrics values are shown for the obtained power regression model. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>53.042</head><p>The obtained value of the coefficient of determination is 70%, which is much better than the coefficient of determination of the linear model.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.6.">The exponential regression</head><p>For the exponential regression, the relationship between H and ŷ is: , In this way we obtained the statistically significant result. In Table <ref type="table" target="#tab_3">4</ref> the quality metrics values are shown for the obtained exponential regression model.  The obtained value of the coefficient of determination indicates that about 74% of cases of changes in the Hurst exponent lead to a change in the size of the queue in the framework of the exponential model, which is the best result when using the methods of paired regression analysis. An analysis of the amount of the penalty gives the same result.</p><p>Let us carry out a comparative analysis of the results obtained and then build graphs of the regression equations (1-5) (Figure <ref type="figure">9</ref>). It is obvious that exponential regression most closely fits the relationship between the Hurst exponent and the buffer size.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Figure 9: The comparative analysis of the results of paired regression analysis</head><p>The trivial paired regression models described above do not adequately describe the dependence of the queue size on the Hurst exponent, so we complicate the model. One possible way is the isotonic regression usage.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.7.">The isotonic regression</head><p>In statistics, isotonic regression or monotonic regression is a method of fitting a free-form line to a sequence of observations under the following constraints: the fitted free-form line should be non-decreasing (or not increasing) over the domain, and should lie as close as possible to the observations <ref type="bibr" target="#b13">[13]</ref>. In the process of constructing an isotonic curve, the following problem is solved <ref type="bibr" target="#b13">[13]</ref>:</p><formula xml:id="formula_8">  2 ˆmin, i i i i w y y  </formula><p>where the value of the weighting factor is 0 i w  . This gives a vector that consists of the non-decreasing elements that are closest in terms of the root mean square error. In practice, this list of elements forms a piecewise linear function.</p><p>Let us train the isotonic regression model using the scikit-learn package of the Python 3 programming language <ref type="bibr" target="#b9">[9]</ref> and build a graph corresponding to the model built using isotonic regression (Figure <ref type="figure" target="#fig_11">10</ref>)</p><p>In Table <ref type="table" target="#tab_5">6</ref> the quality metrics values are shown for the obtained isotonic regression model. The obtained value of the coefficient of determination suggests that about 92% of cases of changes in the Hurst exponent lead to a change in the size of the queue within the framework of this model, which is much better than the models built on the basis of pair regression methods. Moreover, the value of the penalty for isotonic regression is two times less than the corresponding value for paired regression. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.8.">The support vector regression</head><p>Support Vector Machines (SVM) is a linear algorithm used in classification and regression problems (for regression problems it is called SVR -Support Vector Regression). The main idea of the method is to construct a hyperplane that separates the sampled objects in an optimal way <ref type="bibr" target="#b10">[10]</ref><ref type="bibr" target="#b11">[11]</ref><ref type="bibr" target="#b12">[12]</ref>.</p><p>Support vector machines maximize the padding of objects, which is closely related to minimizing the likelihood of overfitting. Moreover, it makes it very easy to go to the construction of a nonlinear dividing surface due to the nuclear transition <ref type="bibr" target="#b10">[10,</ref><ref type="bibr">119]</ref>.</p><p>Let us train the model based on SVR. The nonlinear nature of the relationship between the Hurst exponent value and the queue size indicates the need to choose a radial basis kernel for the SVR model. This model was trained using the scikit-learn package of the Python 3 programming language <ref type="bibr" target="#b12">[12]</ref>. In Figure <ref type="figure" target="#fig_12">11</ref> the graph of the relationship between queue size and Hurst exponent is shown. In Table <ref type="table" target="#tab_6">7</ref> the quality metrics values are shown for the obtained support vector regression model. The obtained value of the coefficient of determination is about 90%, which is slightly worse than that of the method using isotonic regression. However, the penalty for this method is less than for isotonic regression.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.9.">Comparative analysis of models</head><p>The research results are presented in Table <ref type="table" target="#tab_7">8</ref> for estimating and choosing the best method for predicting queue size from the Hurst exponent.</p><p>Based on the data of the pivot table, it can be concluded that the best predictive ability based on the introduced quality metric is a model built using the support vector machine. Within the framework of this study, it can be concluded that the complication of SVR by transition to the rectifying space does not lead to an improvement in the quality of learning. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Conclusions</head><p>Thus, we investigated seven models that allow to predict the size of the queue when transforming an input flow with a Pareto distribution into an output flow with an exponential distribution depending on the Hurst exponent of the input flow, built on the basis of regression analysis methods.</p><p>The unacceptability of the classical approach, which consists in dividing the entire dataset into training and test samples, is shown. Since, within the framework of the set task, the obtained models must be tested by simulation modeling of traffic transformation with a limit on the queue size, based on the results of applying the tested model for determining the queue size to the sequence being converted.</p><p>Since the amount of losses in the general case does not give any information about the efficiency of using the queue in the process of converting traffic, to assess the quality of the resulting model, a penalty was introduced that takes into account not only the amount of losses, but also not rational use of buffer memory.</p><p>The best for the selected quality metrics are the isotonic regression and the support vector regression. It managed to to reduce the penalty score for these models by more than two times in comparison with the trivial linear model. The use of these models will make it possible to more efficiently use the buffer space of the RAM of telecommunication network switching nodes. The usage of these models will make it possible to more efficiently use the buffer space of the RAM in telecommunication network switching nodes. Nevertheless, the obtained models do not belong to strong machine learning models, therefore, the additional researches are required using decision trees ensembles and neural networks.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: Graph of the dependence of the amount of the penalty on the buffer volume</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figure 2 :</head><label>2</label><figDesc>Figure 2: The scatter plot of queue size in dependence of the Hurst exponent</figDesc><graphic coords="3,62.30,342.70,238.80,191.05" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>Figure 3 :</head><label>3</label><figDesc>Figure 3: The box-plot of the 30 analyzed groups</figDesc><graphic coords="3,301.10,341.05,231.80,192.70" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_3"><head></head><label></label><figDesc>We obtain the regression equation using the least squares method:</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_4"><head>Figure 4 :</head><label>4</label><figDesc>Figure 4: The linear model report is built using the statsmodels package of the Python programming language</figDesc><graphic coords="4,126.00,415.30,357.35,174.85" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_5"><head></head><label></label><figDesc>the hyperbolic equation is achieved by replacing 1 H with a new variable, which we denote by z [6]. Then the hyperbolic regression equation takes the form 01 ˆ. y b b z </figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_6"><head></head><label></label><figDesc>Using the least squares method, we obtain the regression equation for this model:</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_8"><head>1 b</head><label>1</label><figDesc>and belongs to the class of regression models that can be reduced to linear form using transformations<ref type="bibr" target="#b6">[6</ref></figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_9"><head>1 b</head><label>1</label><figDesc>This equation is non-linear with respect to the coefficient and belongs to the class of regression models, which are reduced to a linear form using transformations[6   The exponential function is internally linear; therefore, estimates of the unknown parameters of its linearized form can be calculated using the classical least squares method. The regression equation is: result of the fitting for the current model is shown on the Figure8.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_10"><head>Figure 8 :</head><label>8</label><figDesc>Figure 8: The exponential model report is built using statsmodels package of the Python programming language</figDesc><graphic coords="8,106.90,56.65,381.25,185.85" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_11"><head>Figure 10 :</head><label>10</label><figDesc>Figure 10: Plotting an isotonic curve to a dataset</figDesc><graphic coords="9,141.60,465.60,312.00,253.90" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_12"><head>Figure 11 :</head><label>11</label><figDesc>Figure 11: Plot corresponding to trained support vector machine</figDesc><graphic coords="10,143.75,226.90,307.55,250.00" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1</head><label>1</label><figDesc>Linear regression model quality metrics</figDesc><table><row><cell cols="2">Quality Metric</cell><cell>Value</cell></row><row><cell cols="2">Coefficient of determination R2</cell><cell>0.584</cell></row><row><cell cols="2">Root mean square error of regression RMSE</cell><cell>130.908</cell></row><row><cell cols="2">Mean absolute error MAE</cell><cell>96.808</cell></row><row><cell>Penalty score</cell><cell>p</cell><cell>55.710</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 2</head><label>2</label><figDesc>Quality metrics of a hyperbolic regression model</figDesc><table><row><cell cols="2">Quality Metric</cell><cell>Value</cell></row><row><cell cols="2">Coefficient of determination R2</cell><cell>0.453</cell></row><row><cell cols="2">Root mean square error of regression RMSE</cell><cell>150.218</cell></row><row><cell cols="2">Mean absolute error MAE</cell><cell>110.511</cell></row><row><cell>Penalty score</cell><cell>p</cell><cell>63.841</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>Table 3</head><label>3</label><figDesc>Quality metrics of the modified hyperbolic model</figDesc><table><row><cell cols="2">Quality Metric</cell><cell>Value</cell></row><row><cell cols="2">Coefficient of determination R2</cell><cell>0.591</cell></row><row><cell cols="2">Root mean square error of regression RMSE</cell><cell>223.798</cell></row><row><cell cols="2">Mean absolute error MAE</cell><cell>77.543</cell></row><row><cell>Penalty score</cell><cell>p</cell><cell>39.537</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_3"><head>Table 4</head><label>4</label><figDesc>Extent regression model quality metrics</figDesc><table><row><cell>Quality Metric</cell><cell>Value</cell></row><row><cell>Coefficient of determination R2</cell><cell>0.699</cell></row><row><cell>Root mean square error of regression RMSE</cell><cell>128.675</cell></row><row><cell>Mean absolute error MAE</cell><cell>72.823</cell></row><row><cell>Penalty score p</cell><cell></cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_4"><head>Table 5</head><label>5</label><figDesc>Exponential regression model quality metrics</figDesc><table><row><cell>Quality Metric</cell><cell>Value</cell></row><row><cell>Coefficient of determination R2</cell><cell>0.745</cell></row><row><cell>Root mean square error of regression RMSE</cell><cell>112.443</cell></row><row><cell>Mean absolute error MAE</cell><cell>65.199</cell></row><row><cell>Penalty score p</cell><cell>46.768</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_5"><head>Table 6</head><label>6</label><figDesc>Isotonic regression quality metrics</figDesc><table><row><cell cols="2">Quality Metric</cell><cell>Value</cell></row><row><cell cols="2">Coefficient of determination R2</cell><cell>0.928</cell></row><row><cell cols="2">Root mean square error of regression RMSE</cell><cell>54.437</cell></row><row><cell cols="2">Mean absolute error MAE</cell><cell>39.501</cell></row><row><cell>Penalty score</cell><cell>p</cell><cell>21.269</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_6"><head>Table 7</head><label>7</label><figDesc>Support vector model quality metrics</figDesc><table><row><cell>Quality Metric</cell><cell>Value</cell></row><row><cell>Coefficient of determination R2</cell><cell>0.901</cell></row><row><cell>Root mean square error of regression RMSE</cell><cell>63.868</cell></row><row><cell>Mean absolute error MAE</cell><cell>52.506</cell></row><row><cell>Penalty score p</cell><cell>18.374</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_7"><head>Table 8</head><label>8</label><figDesc>The comparison between considered regression methods for 0.5&lt;H&lt;1</figDesc><table><row><cell></cell><cell>Coefficient of determination R2</cell><cell>Root mean square error of regression RMSE</cell><cell>Mean absolute error MAE</cell><cell>Penalty score p</cell></row><row><cell>Linear regression</cell><cell>0.584</cell><cell>130.908</cell><cell>96.808</cell><cell>55.710</cell></row><row><cell>Hyperbolic regression 1</cell><cell>0.453</cell><cell>150.218</cell><cell>110.511</cell><cell>63.841</cell></row><row><cell>Hyperbolic regression 2</cell><cell>0.591</cell><cell>223.798</cell><cell>77.543</cell><cell>39.537</cell></row><row><cell>Extent regression</cell><cell>0.699</cell><cell>128.675</cell><cell>72.823</cell><cell>53.042</cell></row><row><cell>Exponential regression</cell><cell>0.745</cell><cell>112.443</cell><cell>65.199</cell><cell>46.768</cell></row><row><cell>Isotonic regression</cell><cell>0.928</cell><cell>54.437</cell><cell>39.501</cell><cell>21.269</cell></row><row><cell>Support vector machine SVR</cell><cell>0.901</cell><cell>63.868</cell><cell>52.506</cell><cell>18.374</cell></row></table></figure>
		</body>
		<back>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<monogr>
		<author>
			<persName><forename type="first">O</forename><forename type="middle">I</forename><surname>Shelukhin</surname></persName>
		</author>
		<title level="m">Fractal processes in telecommunications</title>
				<editor>
			<persName><forename type="first">I</forename><surname>Shelukhin</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><forename type="middle">M</forename><surname>Tenyakshev</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><forename type="middle">V</forename><surname>Aspen</surname></persName>
		</editor>
		<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<monogr>
		<title/>
		<author>
			<persName><forename type="first">Ed</forename><forename type="middle">O I</forename><surname>Shelukhin</surname></persName>
		</author>
		<imprint>
			<publisher>Radiotekhnika</publisher>
			<biblScope unit="page" from="2003" to="2479" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<monogr>
		<title level="m" type="main">Simulation model of asynchronous transformation of self-similar traffic in switching nodes using a queue // Infocommunication technologies</title>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">I</forename><surname>Linets</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">V</forename><surname>Govorova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">A</forename><surname>Voronkin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><forename type="middle">P</forename><surname>Mochalov</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2019">2019</date>
			<biblScope unit="volume">17</biblScope>
			<biblScope unit="page" from="293" to="303" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Security methods for a group of mobile robots according to the requirements of Russian and foreign legislation</title>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">S</forename><surname>Basan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><forename type="middle">S</forename><surname>Basan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">A</forename><surname>Lapina</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><forename type="middle">N</forename><surname>Kormakova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><forename type="middle">G</forename><surname>Lapin</surname></persName>
		</author>
		<idno type="DOI">10.1088/1757-899X/873/1/012031</idno>
	</analytic>
	<monogr>
		<title level="j">IOP Conference Series: Materials Science and Engineering</title>
		<imprint>
			<biblScope unit="volume">873</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page">12031</biblScope>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Decrease energy consumption of transport telecommunication networks due to the usage of stage-by-stage controlling procedure</title>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">I</forename><surname>Linets</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">V</forename><surname>Melnikov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">V</forename><surname>Govorova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><forename type="middle">V</forename><surname>Medenec</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">A</forename><surname>Lapina</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">CEUR Workshop Proceedings REMS 2018 -Proceedings of the 2018 Multidisciplinary Symposium on Computer Science and ICT</title>
				<imprint>
			<date type="published" when="2018">2018</date>
			<biblScope unit="page" from="181" to="190" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<monogr>
		<title level="m" type="main">A program for generating a dataset to study the statistical characteristics of a self-similar traffic transformation model</title>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">I</forename><surname>Linets</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">V</forename><surname>Govorova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Voronkin</surname></persName>
		</author>
		<imprint>
			<biblScope unit="page">2019619275</biblScope>
		</imprint>
	</monogr>
	<note>Certificate of state registration of a computer program. Register date 15</note>
</biblStruct>

<biblStruct xml:id="b6">
	<monogr>
		<idno type="DOI">10.1007/978-3-662-46221-8</idno>
		<ptr target="https://doi.org/10.1007/978-3-662-46221-8" />
		<title level="m">Handbook of Mathematics</title>
				<editor>
			<persName><forename type="first">N</forename><surname>Bronshtein</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">K</forename><forename type="middle">A</forename><surname>Semendyayev</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">G</forename><surname>Musiol</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">H</forename><surname>Mühlig</surname></persName>
		</editor>
		<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<monogr>
		<ptr target="https://habr.com/ru/company/ods/blog/322076/" />
		<title level="m">Basic principles of machine learning on the example of linear regression</title>
				<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<monogr>
		<title level="m" type="main">Building machine learning systems in Python</title>
		<author>
			<persName><forename type="first">Pedro</forename><surname>Luis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Willie</forename><surname>Coelho</surname></persName>
		</author>
		<author>
			<persName><surname>Richart</surname></persName>
		</author>
		<editor>Slinkin A.A. M.</editor>
		<imprint>
			<date type="published" when="2016">2016</date>
			<publisher>DMK Press</publisher>
			<biblScope unit="page">302</biblScope>
		</imprint>
	</monogr>
	<note>2nd edition / translate from English</note>
</biblStruct>

<biblStruct xml:id="b9">
	<monogr>
		<ptr target="https://scikit-learn.org/stable/modules/isotonic.html(датаобращения01.06" />
		<title level="m">Isotonic regression</title>
				<imprint>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<monogr>
		<author>
			<persName><forename type="first">B</forename><surname>Chardin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Massaron</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Boschetti</surname></persName>
		</author>
		<title level="m">Large-scale machine learning with Python / translate from English</title>
				<editor>
			<persName><forename type="first">A</forename><forename type="middle">V</forename><surname>Logunova</surname></persName>
		</editor>
		<imprint>
			<publisher>DMK Press</publisher>
			<date type="published" when="2018">2018</date>
			<biblScope unit="page">358</biblScope>
		</imprint>
	</monogr>
	<note>in Russian</note>
</biblStruct>

<biblStruct xml:id="b11">
	<monogr>
		<title level="m" type="main">Python and machine learning / translate from English</title>
		<author>
			<persName><forename type="first">S</forename><surname>Raska</surname></persName>
		</author>
		<editor>A.V. Logunova</editor>
		<imprint>
			<date type="published" when="2017">2017</date>
			<publisher>DMK Press</publisher>
			<biblScope unit="page">418</biblScope>
		</imprint>
	</monogr>
	<note>in Russian</note>
</biblStruct>

<biblStruct xml:id="b12">
	<monogr>
		<ptr target="https://scikit-learn.org/stable/auto_examples/svm/plot_svm_regression.html?highlight=svr" />
		<title level="m">Support Vector Regression (SVR) using linear and non-linear kernels</title>
				<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<monogr>
		<title level="m" type="main">Causal isotonic regression</title>
		<author>
			<persName><forename type="first">T</forename><surname>Westling</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Gilbert</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Carone</surname></persName>
		</author>
		<ptr target="http://arxiv.org/abs/1810.03269" />
		<imprint/>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
