<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Using of Ellipsoid Method and Linear Regression with L 1 -Regularization for Medical Data Investigation</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Petro</forename><surname>Stetsyuk</surname></persName>
							<email>stetsyukp@gmail.com</email>
							<affiliation key="aff0">
								<orgName type="department">V.M. Glushkov Institute of Cybernetics of the NASU</orgName>
								<address>
									<addrLine>Academician Glushkov Avenue, 40</addrLine>
									<postCode>03187</postCode>
									<settlement>Kyiv</settlement>
									<country key="UA">Ukraine</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Viktor</forename><surname>Stovba</surname></persName>
							<email>vik.stovba@gmail.com</email>
							<affiliation key="aff0">
								<orgName type="department">V.M. Glushkov Institute of Cybernetics of the NASU</orgName>
								<address>
									<addrLine>Academician Glushkov Avenue, 40</addrLine>
									<postCode>03187</postCode>
									<settlement>Kyiv</settlement>
									<country key="UA">Ukraine</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Ivan</forename><surname>Senko</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">V.M. Glushkov Institute of Cybernetics of the NASU</orgName>
								<address>
									<addrLine>Academician Glushkov Avenue, 40</addrLine>
									<postCode>03187</postCode>
									<settlement>Kyiv</settlement>
									<country key="UA">Ukraine</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Illya</forename><surname>Chaikovsky</surname></persName>
							<email>illya.chaikovsky@gmail.com</email>
							<affiliation key="aff0">
								<orgName type="department">V.M. Glushkov Institute of Cybernetics of the NASU</orgName>
								<address>
									<addrLine>Academician Glushkov Avenue, 40</addrLine>
									<postCode>03187</postCode>
									<settlement>Kyiv</settlement>
									<country key="UA">Ukraine</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">≤</forename><surname>𝑧 #</surname></persName>
						</author>
						<author>
							<persName><forename type="first">−</forename><surname>𝑦</surname></persName>
						</author>
						<author>
							<persName><surname>𝑎 #% 𝑥</surname></persName>
						</author>
						<author>
							<persName><surname>𝑧 # , 𝑖</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">V.M. Glushkov Institute of Cybernetics of the NASU</orgName>
								<address>
									<addrLine>Academician Glushkov Avenue, 40</addrLine>
									<postCode>03187</postCode>
									<settlement>Kyiv</settlement>
									<country key="UA">Ukraine</country>
								</address>
							</affiliation>
						</author>
						<author>
							<affiliation key="aff1">
								<address>
									<postCode>2024</postCode>
									<settlement>Cambridge</settlement>
									<region>MA</region>
									<country key="US">USA</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Using of Ellipsoid Method and Linear Regression with L 1 -Regularization for Medical Data Investigation</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">6FC165EAE79D78A29F0BC41BF37443D9</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T19:12+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>linear regression, least moduli criterion, 𝐿 ! -regularization, non-smooth optimization problem, Shor&apos;s ellipsoid method, dependent factors, data prediction 1 0000-0003-4036-2543 (P. Stetsyuk)</term>
					<term>0000-0003-3023-5815 (V. Stovba)</term>
					<term>0000-0002-2432-4582 (I. Senko)</term>
					<term>0000-0002-4152-0331 (I. Chaikovsky)</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>The problem of finding of parameters of linear regression model with 𝐿 ! -regularization and the least moduli criterion with 1 ≤ 𝑝 ≤ 2 is considered. To solve the problem the Shor's ellipsoid method is used, which is implemented as the emlmpr algorithm. A series of three computational experiments is conducted, which demonstrate solving time of the emlmpr algorithm and robustness of the least moduli criterion if 𝑝 is close to 1. The third experiment considers situation when the model contains linearly dependent features and shows the effect of 𝐿 ! -regularization on the quality of solutions obtained.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>Regression models are an extremely prevalent tool for effective prediction both in machine learning and artificial intelligence in general. Applying of linear regression models for building effective forecasting models, which describe linear relationships between factors, in such fields as statistics, medicine, economics, ecology, identification of parameters of complex systems etc. is studied and investigated. This type of models has proven themselves to be flexible in construction and to provide clear interpretation of relationships between dependent variable and model factors, sometimes even outperforming more complex nonlinear models <ref type="bibr" target="#b0">[1]</ref>.</p><p>When working with regression models, it is rather important to choose correct criteria for estimating model parameters. The most well-known and common variants are the criterion based on least squares and based on least moduli. Effectiveness of the first variant is confirmed by theoretical studies <ref type="bibr" target="#b1">[2]</ref> and numerous statistical experiments. Nevertheless, one of the most significant disadvantages of the least squares criterion is the increase of the effect of large errors when they are squared, which makes the model extremely sensitive to anomalous observations (or outliers). An important condition for using this criterion is the standard normal distribution of model errors, which is not always fulfilled in practice. A well-known and effective alternative to this criterion is the criterion based on the least moduli, which is robust to outliers <ref type="bibr" target="#b2">[3,</ref><ref type="bibr" target="#b3">4]</ref> and assumes a Laplacian distribution of model errors.</p><p>Another important aspect of work with linear regression models is the presence of dependencies between two or more factors of a model, which negatively affect the quality of the obtained parameter estimates. Usually, such dependencies are detected at the stage of data preprocessing and model building by selecting optimal set of model factors that best describe relationship between the dependent variable and the factors. However, in practice, situations often occur when a certain group of factors collectively affects the dependent variable. As a result, both the criterion based on the least squares and the least moduli incorrectly determines parameters of the model, often significantly overestimating or underestimating them. Therefore, it is expedient to develop methods and criteria that make it possible to detect such dependencies between factors and make their coefficients be close to zero. One of the most famous so-called shrinkage methods in machine learning <ref type="bibr" target="#b0">[1]</ref> is regularization approach that permits to balance the model and reduce the effect of dependent factors on the quality of parameter determination.</p><p>The article is dedicated to applying of the Shor's ellipsoid method for finding parameters of a linear regression model with 𝐿 ! -regularization and the least moduli criterion with 1 ≤ 𝑝 ≤ 2. This criterion includes the use of the least moduli (𝑝 = 1) and the least squares (𝑝 = 2) criteria, as well as allows to use any value of the parameter 𝑝. Certain work results of applying the ellipsoid method for this type of problems are given in <ref type="bibr" target="#b4">[5]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Finding of linear regression model parameters using the least moduli criterion powered to p</head><p>Let us consider a classical linear regression problem: to find 𝑛 unknown parameters 𝑥 ! , … , 𝑥 " with known observations</p><formula xml:id="formula_0">(𝐚 # , 𝑦 # ), 𝐚 # = (𝑎 #! , 𝑎 #$ , … , 𝑎 #" ) ∈ 𝑹 " , 𝑦 # ∈ 𝑹, 𝑖 = 1, 𝑚</formula><p>444444 , which are related as follows:</p><formula xml:id="formula_1">𝑦 # = 5 𝑎 #% 𝑥 % " %&amp;! + 𝜀 # , 𝑖 = 1, 𝑚 444444 ,<label>(1)</label></formula><p>where 𝑎 #% are known coefficients, 𝜀 # are unknown random variables, which have (approximately) the same distribution functions, 𝑚 &gt; 𝑛. The equation ( <ref type="formula" target="#formula_1">1</ref>) can be rewritten in matrix form</p><formula xml:id="formula_2">𝑦 = 𝐴𝑥 + 𝜀,<label>(2)</label></formula><p>where</p><formula xml:id="formula_3">𝑦 = (𝑦 ! , … , 𝑦 ' ) ( ∈ 𝑹 ' and 𝜀 = (𝜀 ! , … , 𝜀 ' ) ( ∈ 𝑹 ' are 𝑚-dimentional vectors, 𝐴 is a 𝑚 × 𝑛- matrix, 𝑥 = (𝑥 ! , … , 𝑥 " ) ( ∈ 𝑹 " is a 𝑛-dimentional</formula><p>vector that is to be evaluated. The least moduli method powered to 𝑝, which corresponds to finding the unknown vector 𝑥 ) * according to the least moduli criterion powered to 𝑝 (1 ≤ 𝑝 ≤ 2), is a mathematical programming problem:</p><formula xml:id="formula_4">𝑓 * = 𝑓=𝑥 ) * &gt; = min +∈𝑹 " B𝑓(𝑥) = 5 C𝑦 # − 5 𝑎 #% 𝑥 % " %&amp;! C ) ' #&amp;! E,<label>(3)</label></formula><p>where |•| is an absolute value of a number. The function 𝑓(𝑥) is non-smooth, if 𝑝 = 1 and smooth, if 𝑝 &gt; 1.</p><p>The problem (3) is a problem of unconditional minimization of the convex function 𝑓(𝑥), subgradient of which at the point 𝑥̅ is calculated using the following formula:</p><formula xml:id="formula_5">𝑔 . (𝑥̅ ) = ⎩ ⎪ ⎪ ⎨ ⎪ ⎪ ⎧ 𝑝 5 𝑠𝑖𝑔𝑛 O5 𝑎 #% " %&amp;! 𝑥̅ % − 𝑦 # P C5 𝑎 #% " %&amp;! 𝑥̅ % − 𝑦 # C )/! ' #&amp;! 𝑎 #! , … 𝑝 5 𝑠𝑖𝑔𝑛 O5 𝑎 #% " %&amp;! 𝑥̅ % − 𝑦 # P C5 𝑎 #% " %&amp;! 𝑥̅ % − 𝑦 # C )/! ' #&amp;! 𝑎 #" , ⎭ ⎪ ⎪ ⎬ ⎪ ⎪ ⎫<label>(4)</label></formula><p>If 𝑝 = 1, the problem (3) can be formulated as a following mathematical programming problem:</p><formula xml:id="formula_6">𝑓 ! * = min +∈𝑹 " B𝑓 ! (𝑥) = 5 C𝑦 # − 5 𝑎 #% 𝑥 % " %&amp;! C ' #&amp;! E.<label>(5)</label></formula><p>The problem ( <ref type="formula" target="#formula_6">5</ref>) is a problem of unconditional minimization of convex piecewise-linear function 𝑓 ! (𝑥), which corresponds to the least moduli method, which has proven to be robust to anomalous observations or outliers <ref type="bibr" target="#b2">[3,</ref><ref type="bibr" target="#b5">6]</ref>. Finding the best according to the least moduli criterion vector 𝑥 * , where 𝑥 * is a solution of the problem <ref type="bibr" target="#b4">(5)</ref>, can be formulated as the following LP-problem: to find</p><formula xml:id="formula_7">E.<label>(8)</label></formula><p>The problem ( <ref type="formula" target="#formula_7">8</ref>) is a problem of unconditional minimization of a convex piecewise-linear function 𝑓 ) (𝑥). Here 𝜆 is a regularization parameter, and if 𝜆 = 0 the function 𝑓 ) (𝑥) coincides with the function 𝑓(𝑥). To calculate the subgradient of the function 𝑓 ) (𝑥) at the point 𝑥̅ one can use the following formula:</p><formula xml:id="formula_8">𝑔 . # (𝑥̅ ) = 𝑔 . (𝑥̅ ) + 𝜆 𝑠𝑖𝑔𝑛(𝑥̅ ),<label>(9)</label></formula><p>where 𝑔 . (𝑥̅ ) is calculated using the expression (4).</p><p>For solving the problem (8) the Shor's ellipsoid method <ref type="bibr" target="#b6">[7,</ref><ref type="bibr" target="#b7">8,</ref><ref type="bibr" target="#b8">9]</ref> can be used, which is implemented as the emshor program <ref type="bibr" target="#b9">[10]</ref>. We will apply it for the problem of the function 𝑓 ) (𝑥) minimization, providing that its minimum point 𝑥 ) * is localized in 𝑛-dimensional ball with radius 𝑟 4 , which is centered at the point 𝑥 4 ∈ 𝑹 " , i.e. \𝑥 4 − 𝑥 ) * \ ≤ 𝑟 4 . The algorithm to be used is called emlmpr, description of which is given below.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">The emlmpr algorithm and its Octave implementation</head><p>The input parameter of the algorithm is 𝜀 . &gt; 0 -accuracy, with which 𝑓 ) * = 𝑓 ) =𝑥 ) * &gt; is to be found.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Initialization.</head><p>Let us consider 𝑛 × 𝑛-matrix 𝐵 and set 𝐵 4 ≔ 𝐼 " , where 𝐼 " is 𝑛 × 𝑛 identity matrix. We go to the first iteration with values 𝑥 4 , 𝑟 4 and 𝐵 4 . Let values 𝑥 5 ∈ 𝑹 " , 𝑟 5 , 𝐵 5 be found at the iteration 𝑘. Passing to the iteration 𝑘 + 1 consists of the following sequence of actions.</p><p>Step 1. Calculate 𝑓 ) (𝑥 5 ) and subgradient 𝑔 . # (𝑥 5 ) at the point 𝑥 5 using formula <ref type="bibr" target="#b8">(9)</ref>. If 𝑟 5 a𝐵 5 ( 𝑔 . # (𝑥 5 )a ≤ 𝜀 . , then "Stop: 𝑘 * = 𝑘 and 𝑥 ) * = 𝑥 5 ". Otherwise, go to the step 2.</p><p>Step 2. Set 𝜉 5 ≔</p><formula xml:id="formula_9">6 $ % 7 &amp; # (+ $ ) :6 $ % 7 &amp; # (+ $ ):</formula><p>.</p><p>Step </p><formula xml:id="formula_10">\𝐵 5 /! =𝑥 5 − 𝑥 ) * &gt;\ ≤ 𝑟 5 , 𝑘 = 0,1,2, … , 𝑘 * .</formula><p>On each iteration 𝑘 &gt; 0 the value of decreasing of volume of the ellipsoid 𝐸 5 = i𝑥 ∈ 𝑅 " : \𝐵 5 /! (𝑥 5 − 𝑥)\ ≤ 𝑟 5 k, which localizes point 𝑥 ) * , is constant and equal to</p><formula xml:id="formula_11">𝑞 = 𝑣𝑜𝑙(𝐸 5 ) 𝑣𝑜𝑙(𝐸 5/! ) = p 𝑛 − 1 𝑛 + 1 q 𝑛 √𝑛 $ − 1 s " &lt; 𝑒𝑥𝑝 v− 1 2(𝑛 + 1) w &lt; 1.</formula><p>Theorem implies the fact that the algorithm of finding 𝑥 ) * can be successfully run on modern computers, if 𝑛 = 10 ÷ 30 and 𝑚 = 100 ÷ 1000. Indeed, to decrease in 10 times volume of the ellipsoid localizing the point 𝑥 ) * , it is needed to perform 𝐾 iterations, where 𝐾 = =" !4</p><p>=" &gt; ≈ (2 𝑙𝑛 10)(𝑛 + 1) ≈ 4.6(𝑛 + 1). It means that in order to improve deviation of found record value of the function 𝑓 ) (𝑥) from its optimal value 𝑓 ) * by 10 times, it is necessary to perform 4.6(𝑛 + 1) $ iterations of the algorithm for finding 𝑥 ) * .</p><p>If 𝑛 = 30 and 𝜀 . = 10 /? × 𝑓(𝑥 4 ), then the maximal number of iterations of the algorithm is equal to 4.6(𝑛 + 1) $ = 46 × 961 = 44206. Therefore, even the straight-up matrix-vector implementation of calculation of the function 𝑓 ) (𝑥) value and its subgradient according to the formula (9) allows to provide fast algorithm work on modern computers.</p><p>The algorithm emlmpr for finding an approximation to the point 𝑥 ) * is implemented using Octave language. Its code is given below. Core of the emlmpr program is the for loop (rows 4-16). First, the value of the function 𝑓 (line 5) and its normalized subgradient at the point 𝑥 ) (row 10) are calculated. If the stop condition is satisfied (row 11), the algorithm stops its work. Stop in the emlmpr algorithm occurs when a condition 𝑟 5 a𝐵 5 ( 𝑔 . # (𝑥 5 )a ≤ 𝜀 . is fulfilled, which is equivalent to condition 𝑓 ) (𝑥 5 ) − 𝑓 ) * ≤ 𝜀 . . Otherwise, the next point 𝑥 5;! is calculated (row 13), the space transformation matrix 𝐵 5;! (row 14) and the radius 𝑟 5;! (row 15) are recalculated.</p><formula xml:id="formula_12"># Input parameters: #com01 # A(m,n) -observation matrix; #com02 # y(m,1) -vector of tags (output vector); #com03 # p -power for least moduli criterion, 1&lt;=p&lt;=2; #com04 # lambda -regularization rate; #com05 # x0(n,1) -starting point; #com06 # r0 -</formula></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Computational experiments without regularization</head><p>To demonstrate the effectiveness of the emlmpr algorithm work we present results of three computational experiments conducted for solving the problem <ref type="bibr" target="#b7">(8)</ref>. For the first and the second experiments parameters 𝑛 = 30 and 𝑚 = 10 × 𝑛 = 300. The purpose of the first experiment is to estimate time of solving the problem (8) for specified parameters on a personal computer with Intel Core i7-10750H processor (2.6 GHz), and 16 Gb RAM. The purpose of the second experiment is to demonstrate robustness of the least moduli method, and therefore solutions of the problem (8) without regularization (𝜆 = 0), if 𝑝 is close to one. Third experiment is dedicated to finding parameters of linear regression model using real medical data for further prediction psychological indicators.</p><p>All the calculations are performed on a computer with Intel Core i7-10750H processor (2.6 GHz), 16 Gb RAM in Windows 10/64 system using GNU Octave, version 6.3.0. For the first two experiments regularization parameter 𝜆 is chosen equal to zero.</p><p>Test example 1. For the first experiment input data for the problem (8) are matrix 𝐴 and vector 𝑦, which are generated randomly with a standard uniform distribution according to the following formulas:  It is easy to see from Table <ref type="table">1</ref> that to get solution with accuracies 10 /? ÷ 10 /!$ for different 𝑝 the emlmpr algorithm requires approximately 40 000 iterations and no more than 7 seconds of time. The least deviation 𝑑𝑥 equals 2.3e-11 and is obtained for 𝑝 = 1.</p><formula xml:id="formula_13">A =</formula><p>Test example 2. The purpose of the second experiment is to demonstrate robustness of the least moduli method, which means that the same robustness will characterize solutions of the problem (8), if 𝑝 is close to one. Here, the matrix 𝐴, the starting point 𝑥 4 , ball radius 𝑟 4 are chosen to be the same as in the first test, the vector 𝑦 is adjusted so that its odd components remain the same as in the first test, and even components are multiplied by the value q = (1.0 + 1.0*sign(0.5 -rand)). Thus, even components of the vector 𝑦 can be considered anomalous (incorrect) results of observations.  Calculation results for 𝑛 = 30 and 𝑚 = 300 are given in Table <ref type="table">2</ref>. Here, 𝑖𝑠𝑡 is an exit code of the emlmpr program, 𝑑𝑥 is a norm of deviation of found approximation to the minimum point from the point xstar. The 5th column contains values of the function 𝑓 ) at the found point 𝑥 ) , the 7th column contains the 𝑝-th root of the 5th column. For all the values of the parameter 𝑝 code 𝑖𝑠𝑡 = 1, which indicates successful completion of the program.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Table 2</head><p>Results of solving the problem (8) with 𝒏 = 𝟑𝟎, 𝒎 = 𝟑𝟎𝟎, 𝝀 = 𝟎, and different 𝒑 Results of Table <ref type="table">2</ref> show that the function value 𝑓 C grows as the parameter 𝑝 increases: from 1.34e+05 if 𝑝 = 1 to 1.09e+08 if 𝑝 = 2. Deviation 𝑑𝑥 of the solution found from the minimum point with 𝑝 = 1 is significantly smaller than if 𝑝 &gt; 1, which confirms robustness of the least moduli method corresponding to 𝑝 = 1 situation. It is important to emphasize that this situation is typical for all the values of the parameter 𝑝 close enough to 1. Time used for finding solutions for each of the parameter 𝑝 values does not exceed 4 seconds.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Computational experiments with regularization</head><p>To show effectiveness of the emlmpr algorithm applied to real data we consider the problem of prediction of psychological indicators of the patient's condition based on cardiological data obtained using complex <ref type="bibr" target="#b10">[11]</ref>. There were 90 patients studied with more than 200 features including cardiological and basic ones (like age and ordinal number). Willing to exclude choice of categorial features recoding method from analysis so we are omitting categorial feature as well as ordinal. Practically, usage of ordinal features instead of numerical could increase the quality of linear modelling, see <ref type="bibr" target="#b11">[12]</ref>, however, we need to simplify experiment in order to research only the ellipsoid method usage. While ability of the medical complex <ref type="bibr" target="#b10">[11]</ref> to create binning good enough for the linear modelling is out the scope of the current research. So, we are taking just 175 numerical features that we have. Then, we apply the feature selection procedure to test the ellipsoid method on the dataset being optimal at least at some sense.</p><p>We want to select features that describe relationship between medical and psychological data in the best way using the 𝑅 $ metric <ref type="bibr" target="#b0">[1]</ref>. While the goal of the studying the medical data includes feature interpretability, we take these data as is. In other words, we do not make transformations like PCA and similar ones to get linear independent features. Undoubtedly, it is possible to get some interpretation even after the transformations, but our approach is to take features as is. Taking into account that internal metrics for feature importance in the case of linear regression model work are the best when features are either linearly independent or have normal distributions at least, we cannot rely on internal linear regression metrics, so we try to use "wrapper" approach for the model feature selection <ref type="bibr" target="#b12">[13]</ref>. For the quality metric, we use 5-fold cross-validation <ref type="bibr" target="#b0">[1]</ref>. Since the initial dataset holds missing values, we use simple imputation via median strategy using only training subsample to avoid distortion due to the whole-set median calculation. Moreover, in our situation the initial number of features, which is 200, is greater than number of observations, which is 90, so we start from the first feature, increase number of features until the quality metric 𝑅 $ stops to grow. Also, we consider non-transformed features to decrease the number of experiments to perform and the variability of the whole scheme. Selection of the optimal transformation is an additional task, which is out of scope of the current paper. In general, the feature selection procedure is described at Figure <ref type="figure" target="#fig_2">1</ref>.</p><p>The calculations for feature selection are made in Python 3 <ref type="bibr" target="#b13">[14]</ref> using Google Colab with Sequential Feature Selection and Linear Regression classes with embedded 𝑅 $ -metric taken from Scikit-learn library <ref type="bibr" target="#b14">[15]</ref>. We also used Pandas library <ref type="bibr" target="#b15">[16]</ref> for keeping feature names during calculations.   Table <ref type="table">4</ref> shows that to solve the problem with 𝑝 = 1 with 𝜀 . = 10 /? and 𝜀 . = 10 /$4 the emlmpr program requires approximately 8 thousand operations. If we use 𝑝 = 1 for the same accuracies 11700 iterations are required, and their number is increased to 29719 iterations when using 𝜀 . = 10 /F4 . The 𝑓 ) value for fixed 𝑝 remains unchanged.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>𝑝</head><p>As it can be seen from Table <ref type="table">4</ref>, the emlmpr program successfully finds linear regression model coefficients when using 𝜆 = 0 (see Table <ref type="table" target="#tab_4">5</ref>). However, some of the coefficients are rather larger than others (bold values in Table <ref type="table" target="#tab_4">5</ref>), which can indicate presence of dependency between the following features in the observation matrix. To reduce their effect on the quality of coefficients restoration we apply 𝐿 ! -regularization, which allows to set model parameters corresponding to dependent columns to zero. In practice, it is difficult to obtain exactly zero values of the corresponding parameters, so we have to settle for values close to zero with a certain accuracy.   <ref type="table" target="#tab_4">5</ref>, as well as any changes in coefficients digits are highlighted in bold. It is easy to see that now these coefficients are rather close to zero with sufficient accuracy: 10 /$ for the feature 7 with any values of 𝑝 and 𝜀 . , 10 /G for the feature 14 with 𝑝 = 1 and 𝜀 . = 10 /? and even 10 /$H for the feature 16 with 𝑝 = 2 and 𝜀 . = 10 /F4 . The rest of the coefficients remained almost unchanged except several digits. It is also worth noting that increasing of the regularization rate leads to decreasing coefficients values of dependent features even more. It gives an instrument to adjust the impact of regularization and obtain coefficients at dependent features close enough to zero, thus improving quality of the solutions obtained.</p><p>The prediction results obtained using the model with parameters calculated with the emlmpr algorithm show that using the least moduli method (𝒑 = 𝟏) we obtain many more zero values (which means that solution is found with required accuracy) than in case of using the least square method (𝒑 = 𝟐). Thus, using 𝒑 = 𝟏 is more appropriate than 𝒑 = 𝟐. 𝑝 = 1 𝑝 = 2 𝜀 . = 10 /?</p><p>𝜀 . = 10 </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.">Conclusions</head><p>The paper investigates the problem of finding parameters of linear regression model with the least moduli criterion with 𝟏 ≤ 𝒑 ≤ 𝟐 and 𝑳 𝟏 -regularization. The problem is formulated as a problem of unconditional minimization of a convex piecewise-linear function. For solving this problem, Shor's ellipsoid method is used, which is implemented by the emlmpr program using Octave programming language.</p><p>Series of three computational experiments with the emlmpr program are considered. Results of the first experiment show that the problem of finding parameters of linear regression model with 𝒏 = 𝟑𝟎 and 𝒎 = 𝟑𝟎𝟎 can be solved within 7 seconds being run on modern laptop of average performance. The second experiment shows that the least moduli criterion is robust if 𝒑 is close to one, thus solutions of the problem are robust as well. The third experiment is dedicated to using of 𝑳 𝟏 -regularization for decreasing effect of linearly dependent features that the model can include on the solutions quality. Results of the experiment, where real cardiological data are used for prediction of psychological indicators of the patient's condition, show that the emlmpr algorithm can successfully compute linear regression model parameters with 𝒏 = 𝟏𝟔, 𝒎 = 𝟗𝟎 within 3 seconds, and set coefficients at dependent features to zero with sufficient accuracy using 𝑳 𝟏 -regularization approach. 𝑝 = 1 𝑝 = 2 𝜀 . = 10 /?</p><p>𝜀 . = 10 /$4 𝜀 . = 10 /!$ 𝜀 . = 10 </p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>3 .Step 5 .</head><label>35</label><figDesc>Calculate the next point 𝑥 5;! ≔ 𝑥 5 − ℎ 5 𝐵 5 𝜉 5 , where ℎ 5 = ! ";! 𝑟 5 . Step 4. Calculate 𝐵 5;! : = 𝐵 5 + Oe "/! ";! − 1P (𝐵 5 𝜉 5 )𝜉 5 ( and 𝑟 5;! : = 𝑟 5 " √" ' /! . Go to the iteration 𝑘 + 1 with values 𝑥 5;! , 𝑟 5;! , 𝐵 5;! . Theorem. Sequence of points {𝑥 5 } 5&amp;4 5 * satisfy the following inequalities:</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head></head><label></label><figDesc>radius of the ball centered at x0 that localizes x_p^*; #com07 # epsf, maxitn -stop parameters: #com08 # epsf -precision to stop by the value of the function fp, #com09 # maxitn -maximal number of iterations; #com10 # intp -print information for every intp iteration. #com11 # Output parameters: #com12 # xp(n,1) -approximation to x_p^*; #com13 # fp -the value of the function f_R at the point xp; #com14 # itn -the number of iterations; #com15 # ist -exit code: 1 -epsf, 4 -maxitn. #com16 function [xp,fp,itn,ist] = emlmpr(A,y,p,lambda,x0,r0, epsf,maxitn,intp); #row01 n = columns(A); xp = x0; B = eye(n); r = r0; #row02 dn = double(n); beta = sqrt((dn-1.d0)/(dn+1.d0)); #row03 for (itn = 0:maxitn) #row04 temp = A*xp-y; fp = sum(abs(temp).^p) + lambda*sum(abs(xp)); #row05 if((mod(itn,intp)==0)&amp;&amp;(intp&lt;=maxitn)) #row06 printf(" itn %4d fp %14.6e\n",itn,fp); #row07 endif #row08 g1 = p*A'*(sign(temp).*(abs(temp)).^(p-1)) + lambda*sign(xp);#row09 g = B'*g1; dg = norm(g); #row10 if(r*dg &lt; epsf) ist = 1; return; endif #row11 xi = (1.d0/dg)*g; dx = B * xi; #row12 hs = r/(dn+1.d0); xp -= hs * dx; #row13 B += (beta -1) * B * xi * xi'; #row14 r = r/sqrt(1.d0-1.d0/dn)/sqrt(1.d0+1.d0/dn);</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head># Test 1 :</head><label>1</label><figDesc>10*rand(m,n), y = A*xstar(n,1), xstar(n,1) = round(10*rand(n,1) + 0.5). Starting point is chosen according to the rule x0(n,1) = round(5*rand(n,1)), and radius of the sphere, in which the point 𝑥 ) * = 𝑥 @ABC is located, is chosen according to the rule r0 = 5*norm(x0 -xstar), i.e. 𝑟 4 = ‖𝑥 4 − 𝑥 @ABC ‖. The first experiment is implemented by the following Octave code. emlmpr running time for n = 30 and m = 300 n = 30, m = 10*n, rand("seed", 2024); A = 10*rand(m,n); xstar = round(10.0*rand(n,1) + 0.5); y = A*xstar; x0 = round(5.0*rand(n,1)); r0 = 5*norm(x0xstar), maxitn = 50000, intp = 10000, lambda = 0.0, # running the emlmpr algorithm for p=1.0;1.1.2;1.5;1.8;2.0 printf("\n Test 1: emlmpr runnning time for n = 30 and m = 300 \n"); epsf0 = 1.e-6; ntest = 5;</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_3"><head># Test 2 :</head><label>2</label><figDesc>robustness of the least moduli method for n = 30 and m = 300 n = 30, m = 10*n, rand("seed", 2024); A = 10*rand(m,n); # test example generation xstar = round(10.0*rand(n,1) + 0.5); y = A*xstar; x0 = round(5.0*rand(n,1)); r0 = 5*norm(x0xstar), m1 = m/2, for i = 1:m1 ind = (i-1)*2 + 1; y(ind) = y(ind)*(1.0 + 1.0*sign(0.5rand)); endfor # running the emlmpr algorithm for p=1.0;1.1.2;1.5;1.8;2.0 printf("\nTest 2: robustness of the Least Moduli Method \n"); maxitn = 50000, intp = 10000, lambda = 0.0, epsf0 = 1.e-6; ntest = 5;</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_4"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: Feature selection workflow</figDesc><graphic coords="8,162.03,63.55,278.80,246.00" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head></head><label></label><figDesc>Results of the emlmpr program work for the first experiment are 𝑡𝑖𝑚𝑒 required to solve the problem (8) with accuracy 𝜀 . , the number of iterations 𝑖𝑡𝑛 of the method, the minimum value of the function 𝑓 ) found, norm of deviation 𝑑𝑥 of the found approximation to the minimum point from the known minimum point xstar are given in Table1. Here 𝜀 . is chosen as follows: if 𝑝 = 1 the value 𝜀 . = 10 /? , if 𝑝 &gt; 1 we choose 𝜀 . = (10 /? ) ) .</figDesc><table><row><cell>Table 1</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell cols="8">Results of solving the problem (8) with 𝒏 = 𝟑𝟎, 𝒎 = 𝟑𝟎𝟎 and 𝝀 = 𝟎</cell></row><row><cell></cell><cell>𝑝</cell><cell>𝜀 .</cell><cell></cell><cell>time (sec)</cell><cell>itn</cell><cell></cell><cell>𝑓 )</cell><cell>𝑑𝑥</cell></row><row><cell></cell><cell>1.0</cell><cell cols="2">1.0e-06</cell><cell>5.17</cell><cell>45375</cell><cell></cell><cell>1.71062e-08</cell><cell>2.3e-11</cell></row><row><cell></cell><cell>1.2</cell><cell cols="2">3.2e-08</cell><cell>6.99</cell><cell>42148</cell><cell></cell><cell>3.58266e-10</cell><cell>1.2e-10</cell></row><row><cell></cell><cell>1.5</cell><cell cols="2">1.0e-09</cell><cell>5.39</cell><cell>40061</cell><cell></cell><cell>9.87425e-12</cell><cell>4.1e-10</cell></row><row><cell></cell><cell>1.8</cell><cell cols="2">3.2e-11</cell><cell>5.97</cell><cell>38260</cell><cell></cell><cell>1.81563e-13</cell><cell>7.2e-10</cell></row><row><cell></cell><cell>2.0</cell><cell cols="2">1.0e-12</cell><cell>3.91</cell><cell>37216</cell><cell></cell><cell>7.45098e-15</cell><cell>1.8e-09</cell></row><row><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell>dx];</cell><cell></cell><cell></cell></row><row><cell>itn, fp,</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell>endfor</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell>n,m,</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell>printf("</cell><cell>p</cell><cell>epsf</cell><cell>time</cell><cell>itn ist</cell><cell>fp</cell><cell></cell><cell>dx \n");</cell></row><row><cell cols="2">for (i = 1:ntest)</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell>printf("</cell><cell cols="2">%4.1f</cell><cell cols="2">%6.1e</cell><cell>%4.2f</cell><cell>%6d</cell><cell>%2d</cell><cell>%10.5e</cell><cell>%10.1e\n",</cell></row><row><cell cols="3">table(i, 1:7))</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell>endfor</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row></table><note>table = []; for (i = 1:ntest) p = 1.d0 + (i -1.d0)/(ntest -1.d0), epsf = epsf0**(p); time0 = time(); [xp,fp,itn,ist] = emlmpr(A,y,p,lambda,x0,r0,epsf,maxitn,intp); time1 = time() -time0, dx = norm(xpxstar); table = [table; p epsf time1 itn ist fp</note></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>table = []; for (in = 1:ntest) p = 1.d0 + (in -1.d0)/(ntest -1.d0), epsf</head><label>=</label><figDesc></figDesc><table><row><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell>];</cell><cell></cell><cell></cell></row><row><cell>itn, fp,</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell>endfor</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell>n,m,</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell>printf("</cell><cell>p</cell><cell>epsf</cell><cell>time</cell><cell>itn ist</cell><cell>fp</cell><cell>dx</cell><cell>r(fp)\n");</cell></row><row><cell cols="2">for (i = 1:ntest)</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell cols="5">printf(" %4.1f %6.1e %4.2f %6d %2d</cell><cell>%10.5e %10.1e</cell><cell cols="2">%10.5e\n", table(i, 1:8))</cell></row><row><cell>endfor</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row></table><note>= epsf0**(p); time0 = time(); [xp,fp,itn,ist] = emlmpr(A,y,p,lambda,x0,r0,epsf,maxitn,intp); time1 = time() -time0, dx = norm(xpxstar); table = [table; p epsf time1 itn ist fp dx fp^(1/p)</note></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_3"><head>Table 3</head><label>3</label><figDesc></figDesc><table><row><cell cols="5">Results of the emlmpr program work with 𝒏 = 𝟏𝟔, 𝒎 = 𝟗𝟎, 𝝀 = 𝟎, 𝒑 = 𝟏. 𝟎; 𝟐. 𝟎 and different</cell></row><row><cell>accuracies 𝜺 𝒇</cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell>𝑝</cell><cell>𝜀 .</cell><cell>time (sec)</cell><cell>itn</cell><cell>𝑓 )</cell></row><row><cell cols="2">1.0 1.0e-06</cell><cell>0.56</cell><cell>7640</cell><cell>2.61732e+02</cell></row><row><cell cols="2">1.0 1.0e-20</cell><cell>0.69</cell><cell>8383</cell><cell>2.61732e+02</cell></row><row><cell cols="2">2.0 1.0e-12</cell><cell>0.95</cell><cell>11700</cell><cell>1.38880e+03</cell></row><row><cell cols="2">2.0 1.0e-40</cell><cell>2.15</cell><cell>29719</cell><cell>1.38880e+03</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_4"><head>Table 5 Linear regression model parameters found by the emlmpr program with</head><label>5</label><figDesc>𝒑 = 𝟏. 𝟎; 𝟐. 𝟎, 𝝀 = 𝟎 and different accuracies 𝜺 𝒇</figDesc><table /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_5"><head>Table 6</head><label>6</label><figDesc>contains coefficients of linear regression model found by the emlmpr program with 𝑝 = 1.0; 2.0, different accuracies 𝜀 . and regularization rate 𝜆 = 0.1. Corresponding values to large coefficients from Table</figDesc><table /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_6"><head>Table 6 Linear regression model parameters found by the emlmpr program with</head><label>6</label><figDesc>𝒑 = 𝟏. 𝟎; 𝟐. 𝟎, 𝝀 = 𝟎. 𝟏 and different accuracies 𝜺 𝒇</figDesc><table><row><cell></cell><cell>/$4</cell><cell>𝜀 . = 10 /!$</cell><cell>𝜀 . = 10 /F4</cell></row><row><cell>-8.9453e-02</cell><cell>-8.9453e-02</cell><cell>-1.1440e-01</cell><cell>-1.1440e-01</cell></row><row><cell>-2.5798e-03</cell><cell>-2.5798e-03</cell><cell>-7.0401e-03</cell><cell>-7.0401e-03</cell></row><row><cell>-3.2033e-02</cell><cell>-3.2033e-02</cell><cell>-2.4929e-02</cell><cell>-2.4929e-02</cell></row><row><cell>2.0159e-02</cell><cell>2.0159e-02</cell><cell>2.7608e-02</cell><cell>2.7608e-02</cell></row><row><cell>2.4191e-01</cell><cell>2.4191e-01</cell><cell>2.6629e-01</cell><cell>2.6629e-01</cell></row><row><cell>6.8819e-03</cell><cell>6.8820e-03</cell><cell>3.5181e-02</cell><cell>3.5181e-02</cell></row><row><cell>1.4368e+08</cell><cell>1.4868e+08</cell><cell>7.5756e+06</cell><cell>6.3818e+06</cell></row><row><cell>-1.4482e-02</cell><cell>-1.4482e-02</cell><cell>-1.2610e-02</cell><cell>-1.2610e-02</cell></row><row><cell>3.7213e-02</cell><cell>3.7213e-02</cell><cell>3.7520e-02</cell><cell>3.7520e-02</cell></row><row><cell>-8.9244e-03</cell><cell>-8.9244e-03</cell><cell>-4.5045e-02</cell><cell>-4.5045e-02</cell></row><row><cell>1.4087e+00</cell><cell>1.4087e+00</cell><cell>2.1146e+00</cell><cell>2.1146e+00</cell></row><row><cell>-1.0175e+00</cell><cell>-1.0175e+00</cell><cell>-2.4560e+00</cell><cell>-2.4560e+00</cell></row><row><cell>-3.2982e-01</cell><cell>-3.2982e-01</cell><cell>-1.0611e-01</cell><cell>-1.0611e-01</cell></row><row><cell>-3.9263e+08</cell><cell>-3.7817e+08</cell><cell>-7.8529e+06</cell><cell>-4.2703e+07</cell></row><row><cell>1.4368e+08</cell><cell>1.4868e+08</cell><cell>7.5756e+06</cell><cell>6.3818e+06</cell></row><row><cell>5.5243e+08</cell><cell>-4.9760e+08</cell><cell>-4.5914e+08</cell><cell>9.8456e+08</cell></row></table></figure>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgements</head><p>The paper is supported by National Research Foundation of Ukraine (grants № 2021.01/0136 and №2023.04/0094), Volkswagen Foundation grant № 97775, the project of research works of young scientists №07-02/03-2023, the NASU grant for research laboratories/groups of young scientists №02/01-2024 <ref type="bibr" target="#b4">(5)</ref>, and the DTT TS KNU NASU project № 0124U002162.</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">An Introduction to Statistical Learning: with Applications in Python</title>
		<author>
			<persName><forename type="first">G</forename><surname>James</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Witten</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Hastie</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Tibshirani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Taylor</surname></persName>
		</author>
		<idno type="DOI">10.1007/978-3-031-38747-0</idno>
	</analytic>
	<monogr>
		<title level="s">Springer Texts in Statistics</title>
		<imprint>
			<date type="published" when="2023">2023</date>
			<publisher>Springer Cham</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<monogr>
		<title level="m" type="main">Mathematics for Machine Learning: textbook</title>
		<author>
			<persName><forename type="first">M</forename><surname>Deisenroth</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Faisal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">Soon</forename><surname>Ong</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2020">2020</date>
			<pubPlace>Cambridge</pubPlace>
		</imprint>
	</monogr>
	<note>1st Edition</note>
</biblStruct>

<biblStruct xml:id="b2">
	<monogr>
		<author>
			<persName><forename type="first">P</forename><forename type="middle">J</forename><surname>Huber</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><forename type="middle">M</forename><surname>Ronchetti</surname></persName>
		</author>
		<title level="m">Robust Statistics</title>
				<imprint>
			<publisher>John Wiley &amp; Sons</publisher>
			<date type="published" when="2011">2011</date>
		</imprint>
	</monogr>
	<note>2nd Edition</note>
</biblStruct>

<biblStruct xml:id="b3">
	<monogr>
		<title level="m" type="main">Optimization and Nonsmooth Analysis</title>
		<author>
			<persName><forename type="first">F</forename><forename type="middle">H</forename><surname>Clarke</surname></persName>
		</author>
		<imprint>
			<date type="published" when="1990">1990</date>
			<publisher>SIAM</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Using the Ellipsoid Method to Study Relationships in Medical Data</title>
		<author>
			<persName><forename type="first">P</forename><surname>Stetsyuk</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Budnyk</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Sen'ko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Stovba</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Chaikovsky</surname></persName>
		</author>
		<idno type="DOI">10.34229/2707-451X.23.3.3</idno>
	</analytic>
	<monogr>
		<title level="j">Cybernetics and Computer Technologies</title>
		<imprint>
			<biblScope unit="page" from="23" to="43" />
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">On curve estimation by minimizing mean absolute deviation and its implications</title>
		<author>
			<persName><forename type="first">J</forename><surname>Fan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Hall</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">The Annals of Statistics</title>
		<imprint>
			<biblScope unit="page" from="867" to="885" />
			<date type="published" when="1994">1994</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Cutting-off Method with Space Dilation for Solving Convex Programming Problems</title>
		<author>
			<persName><forename type="first">N</forename><forename type="middle">Z</forename><surname>Shor</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Cybernetics</title>
		<imprint>
			<biblScope unit="page" from="94" to="95" />
			<date type="published" when="1977">1977</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<monogr>
		<title level="m" type="main">Nondifferentiable Optimization and Polynomial Problems</title>
		<author>
			<persName><forename type="first">N</forename><forename type="middle">Z</forename><surname>Shor</surname></persName>
		</author>
		<imprint>
			<date type="published" when="1998">1998</date>
			<publisher>Kluwer</publisher>
			<pubPlace>Amsterdam</pubPlace>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<monogr>
		<title level="m" type="main">Minimization Methods for Non-Differentiable Functions</title>
		<author>
			<persName><forename type="first">N</forename><forename type="middle">Z</forename><surname>Shor</surname></persName>
		</author>
		<imprint>
			<date type="published" when="1985">1985</date>
			<publisher>Springer-Verlag</publisher>
			<pubPlace>Berlin</pubPlace>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">The Generalized Ellipsoid Method and Its Implementation</title>
		<author>
			<persName><forename type="first">P</forename><surname>Stetsyuk</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Fischer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Khomyak</surname></persName>
		</author>
		<idno type="DOI">10.1007/978-3-030-38603-0_26</idno>
	</analytic>
	<monogr>
		<title level="j">Communications in Computer and Information Science</title>
		<imprint>
			<biblScope unit="page" from="355" to="370" />
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Development and implementation into medical practice new information technologies and metrics for analysis of small changes in electromagnetic field of human heart</title>
		<author>
			<persName><forename type="first">I</forename><surname>Chaikovsky</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Primin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Kazmirchuk</surname></persName>
		</author>
		<idno type="DOI">10.15407/visn2021.02.033</idno>
	</analytic>
	<monogr>
		<title level="j">Visnyk of the National Academy of Sciences of Ukraine</title>
		<imprint>
			<biblScope unit="page" from="33" to="43" />
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<monogr>
		<author>
			<persName><forename type="first">R</forename><surname>Persson</surname></persName>
		</author>
		<ptr target="https://lup.lub.lu.se/luur/download?func=downloadFile&amp;recordOId=9066332&amp;fileOId=9067075" />
		<title level="m">Weight of evidence transformation in credit scoring models: How does it affect the discriminatory power?</title>
				<meeting><address><addrLine>Lund, Sweden</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
		<respStmt>
			<orgName>Lund university</orgName>
		</respStmt>
	</monogr>
	<note type="report_type">Master&apos;s thesis</note>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Feature Selection: A Data Perspective</title>
		<author>
			<persName><forename type="first">L</forename><surname>Jundong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Cheng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Morstatter</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">P</forename><surname>Trevino</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Tang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Tang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Liu</surname></persName>
		</author>
		<idno type="DOI">10.1145/3136625</idno>
	</analytic>
	<monogr>
		<title level="j">ACM Computing Surveys</title>
		<imprint>
			<biblScope unit="page" from="1" to="45" />
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<monogr>
		<title level="m" type="main">Python 3 Reference Manual</title>
		<author>
			<persName><forename type="first">G</forename><surname>Van Rossum</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><forename type="middle">L</forename><surname>Drake</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2009">2009</date>
			<publisher>CreateSpace</publisher>
			<pubPlace>Scotts Valley, CA</pubPlace>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">Scikit-learn: Machine Learning in Python</title>
		<author>
			<persName><forename type="first">F</forename><surname>Pedregosa</surname></persName>
		</author>
		<ptr target="http://jmlr.org/papers/v12/pedregosa11a.html" />
	</analytic>
	<monogr>
		<title level="j">Journal of Machine Learning Research</title>
		<imprint>
			<biblScope unit="volume">12</biblScope>
			<biblScope unit="page" from="2825" to="2830" />
			<date type="published" when="2011">2011</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">Data structures for statistical computing in python</title>
		<author>
			<persName><forename type="first">W</forename><surname>Mckinney</surname></persName>
		</author>
		<idno type="DOI">10.25080/Majora-92bf1922-00a</idno>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 9th Python in Science Conference</title>
				<meeting>the 9th Python in Science Conference<address><addrLine>Austin</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2010-07-03">28 June-3 July 2010</date>
			<biblScope unit="page" from="56" to="61" />
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
