<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">The Scalar Metric of Classification Algorithm Choice in Machine Learning Problems Based on the Scheme of Nonlinear Compromises</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Igor</forename><surname>Puleko</surname></persName>
							<email>pulekoigor@gmail.com</email>
							<affiliation key="aff0">
								<orgName type="institution">Zhytomyr Polytechnic State University</orgName>
								<address>
									<addrLine>103, Chydnivska srt</addrLine>
									<postCode>10005</postCode>
									<settlement>Zhytomyr</settlement>
									<country key="UA">Ukraine</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Oleksandra</forename><surname>Svintsytska</surname></persName>
							<affiliation key="aff0">
								<orgName type="institution">Zhytomyr Polytechnic State University</orgName>
								<address>
									<addrLine>103, Chydnivska srt</addrLine>
									<postCode>10005</postCode>
									<settlement>Zhytomyr</settlement>
									<country key="UA">Ukraine</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Victor</forename><surname>Chumakevych</surname></persName>
							<email>chumakevich@ukr.net</email>
							<affiliation key="aff1">
								<orgName type="institution">Lviv Polytechnic National University</orgName>
								<address>
									<addrLine>12, S. Bandery str</addrLine>
									<postCode>79013</postCode>
									<settlement>Lviv</settlement>
									<country key="UA">Ukraine</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Vadym</forename><surname>Ptashnyk</surname></persName>
							<affiliation key="aff2">
								<orgName type="institution">Lviv National Agrarian University</orgName>
								<address>
									<addrLine>1, V.Velykoho str., Dubliany-Lviv</addrLine>
									<postCode>80381</postCode>
									<country key="UA">Ukraine</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Yuliia</forename><surname>Polishchuk</surname></persName>
							<email>polishchuk.yu.ya@gmail.com</email>
							<affiliation key="aff3">
								<orgName type="institution">National Aviation University</orgName>
								<address>
									<addrLine>1, Liubomyra Huzara ave</addrLine>
									<postCode>03058</postCode>
									<settlement>Kyiv</settlement>
									<country key="UA">Ukraine</country>
								</address>
							</affiliation>
						</author>
						<author>
							<affiliation key="aff4">
								<orgName type="department">International Conference on Computational Linguistics and Intelligent Systems</orgName>
								<address>
									<addrLine>May 12-13</addrLine>
									<postCode>2022</postCode>
									<settlement>Gliwice</settlement>
									<country key="PL">Poland</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">The Scalar Metric of Classification Algorithm Choice in Machine Learning Problems Based on the Scheme of Nonlinear Compromises</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">2910CE739BC970CAD59EA191D751468E</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T12:55+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Communications, software, machine learning, classification evaluation metrics, accuracy, recall, precision, F1, AUC_ROC, nonlinear scheme of compromises 0000-0001-8875-017X (I. Puleco)</term>
					<term>0000-0002-2613-2437 (O. Svintsytska)</term>
					<term>0000-0002-5773-393X (V. Chumakevych)</term>
					<term>0000-0002-1018-1138 (V. Ptashnyk)</term>
					<term>0000-0002-0686-2328 (Yu. Polishchuk)</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>A classic example of machine learning methods is machine classification algorithms. At present, a large number of machine classification methods have been developed, which are offered in the form of ready-made software. The presence of a large number of machine classification algorithms raises the problem of choosing the best algorithm for solving a particular task. This problem is also complicated by the ambiguity with the choice of quality indicators since for the analysis of the quality of the classification there exist a number of indicators (metrics) of the classification. It is difficult for an inexperienced user to understand and choose his priorities. However, the problem of evaluating the classification algorithm's quality can be considered a problem of multicriteria decision-making. It is proposed to evaluate the algorithm's quality by means of one scalar indicator obtained by convolution of other indicators by a nonlinear scheme of compromises.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>In the sphere of information technology (IT), machine learning (ML) methods, which are used to solve a number of applied problems, have become widely used. In essence, ML is a class of methods of artificial intelligence, the characteristic feature of which is not a direct solving the problem but learning through the use of solutions to many similar problems. Typically, such methods use mathematical statistics, probability theory, mathematical analysis, optimization methods, numerical methods, graph theory, and various techniques for working with data in digital form <ref type="bibr" target="#b0">[1]</ref>.</p><p>Among the existing methods of machine learning, the most researched and developed are the methods of machine classification, which belong to the controlled type of learning or learning with a teacher (supervised learning) <ref type="bibr" target="#b1">[2]</ref>. A classification task is a task in which there are many objects (situations), divided into classes in some way. Also, a finite set of objects is defined, for which it is known to which classes they belong. This set is called a sample (training sample). The class belonging of other objects is unknown. We need to construct an algorithm that can classify an arbitrary object (specify the number or name of the class to which the object belongs) from the initial set.</p><p>Algorithms that solve the classification problem have been known for a long time, and in mathematical statistics, they are also called problems of discriminant analysis <ref type="bibr" target="#b2">[3]</ref>. In ML, the problem of classification is solved, in particular, by means of a large number of algorithms, including those with the application of methods of artificial neural networks.</p><p>Today, nearly all leading IT companies, to some extent, develop, use, or provide as a service various methods and algorithms of ML. For example <ref type="bibr" target="#b3">[4,</ref><ref type="bibr" target="#b4">5]</ref>, Microsoft's Azure Machine Learning Studio has more than a dozen classification algorithms, each of which can perform the set task (see Figure <ref type="figure" target="#fig_0">1</ref>). The variety of supply raises the problem of choice. Which of the algorithms is better to choose to perform a task? The answer to this question is far from unambiguous since the existing metrics for evaluating the classification quality do not provide an unambiguous result. It is especially challenging for beginners to make such a choice. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Related Works</head><p>The most common and frequently used classification algorithms today are <ref type="bibr" target="#b5">[6]</ref>: Naive Bayes, Decision Trees, Logistic Regression, K-Nearest Neighbors, Support Vector Machines and others.</p><p>According to classical theory, several indicators (classification metrics) are used to evaluate the quality of such classification algorithms. Consider them in more detail.</p><p>The confusion matrix is a table used to describe the effectiveness of a classifier. It is usually extracted from a test data set for which basic true values are known <ref type="bibr" target="#b6">[7]</ref>.</p><p>Here the results of assignment to each class are analyzed and the share of incorrectly assigned classes is determined. In the process of constructing the above table, we are dealing with several key metrics that play a very important role in machine learning.</p><p>For the classification problem, given the actual label and the predicted label, the first thing we can do is divide our samples into 4 segments <ref type="bibr" target="#b6">[7]</ref>:  A number of other characteristics are built on the basis of this matrix (Table <ref type="table" target="#tab_0">1</ref>). Consider each of them in more detail.</p><formula xml:id="formula_0"></formula></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1.">Accuracy</head><p>Accuracy <ref type="bibr" target="#b7">[8]</ref> is the proportion of accurate predictions relative to the total number of predictions, i.e., it is the probability that the class will be predicted correctly <ref type="bibr" target="#b0">(1)</ref>.</p><p>𝑇𝑃 + 𝑇𝑁 𝑇𝑃 + 𝑇𝑁 + 𝐹𝑃 + 𝐹𝑁 .</p><p>(1)</p><p>Thus, accuracy is the fraction of the correct answers of the algorithm.</p><p>Although accuracy is a quick and informative indicator of model performance, we cannot rely on it alone. This is due to the fact that it hides the presence of a shift in the model, which is common if the data set is unbalanced, i.e., the negative aspects are much more than the positive ones, or vice versa. That is, this metric is useless in problems with unequal classes, which, as an option, can be corrected using sampling algorithms. Sampling (data sampling) is a method of adjusting the training sample in order to balance the distribution of classes in the original data set <ref type="bibr" target="#b8">[9]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2.">Precision</head><p>Precision is the proportion of the correct answers of the model within the class, i.e., the proportion of objects that really belong to this class relative to the total number of objects that the system has assigned to this class <ref type="bibr" target="#b7">[8]</ref>.</p><formula xml:id="formula_1">𝑃𝑟𝑒𝑐𝑖𝑠𝑖𝑜𝑛 = 𝑇𝑃 𝑇𝑃 + 𝐹𝑃 .<label>(2)</label></formula><p>The introduction of precision does not allow us to assign all objects to one class since in this case we have an increase in the level of FP.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.3.">Recall</head><p>Recall is the share of true positive rate (TPR) <ref type="bibr" target="#b7">[8]</ref>. Recall shows what share of objects that actually belong to the positive class we have predicted correctly. In other words, this is the proportion of options classified as positive that actually proved to be positive.</p><formula xml:id="formula_2">𝑅𝑒𝑐𝑎𝑙𝑙 = 𝑇𝑃 𝑇𝑃 + 𝐹𝑁 .<label>(3)</label></formula><p>Recall demonstrates the ability of the algorithm to detect this class in general.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.4.">F-score</head><p>Precision and recall do not depend on the relationship of classes (as opposed to accuracy) and, therefore, can be used in unbalanced samples. Often in real practice, the task is to find the optimal (for the customer) balance between these two metrics. It is obvious that the higher the precision and recall, the better. However, in real life, maximum precision and recall are unattainable simultaneously, and, therefore, we have to search for some balance. Thus, it would be convenient to have some metrics that would combine information about the precision and recall of our algorithm. In this case, it will be easier for us to decide which implementation to launch into production. The F-score serves precisely these needs.</p><p>The F-score is the harmonic mean between precision and recall. It tends to zero if accuracy or completeness tends to zero.</p><formula xml:id="formula_3">𝐹1 = 2 × 𝑝𝑟𝑒𝑐𝑖𝑠𝑖𝑜𝑛 × 𝑟𝑒𝑐𝑎𝑙𝑙 𝑝𝑟𝑒𝑐𝑖𝑠𝑖𝑜𝑛 + 𝑟𝑒𝑐𝑎𝑙𝑙 . (<label>4</label></formula><formula xml:id="formula_4">)</formula><p>This formula gives the same weight to precision and recall, so the F-score will fall equally with decreasing precision and recall. It is possible to calculate the F-score by giving different weights to precision and recall if you consciously prioritize one of these metrics when developing an algorithm.</p><p>The F-score is a good candidate for a formal classifier quality evaluation metric. It reduces to one value these two basic metrics: precision and recall. Having the F-score is much easier to answer the question: "Has the algorithm changed for the better or not?"</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.5.">ROC-curve</head><p>Receiver operating characteristics (ROC) curve is used to analyze the behavior of classifiers at different thresholds <ref type="bibr" target="#b9">[10]</ref>. ROC-curve allows us to consider all threshold values for this classifier. It demonstrates the proportion of false positive rate (FPR) compared to the proportion of true positive rate (TPR) (see Figure <ref type="figure" target="#fig_2">2</ref>).</p><formula xml:id="formula_5">𝑇𝑃𝑅 = 𝑇𝑃 𝑇𝑃 + 𝐹𝑁 = 𝑅𝑒𝑐𝑎𝑙𝑙 ,<label>(5)</label></formula><formula xml:id="formula_6">FPR = FP FP + TN . (6)</formula><p>The share of FPR is the proportion of negative samples that were incorrectly classified as positive. 𝐹𝑃𝑅 = 1 − 𝑇𝑁𝑅 , <ref type="bibr" target="#b6">(7)</ref> where TNR is the share of true negative rate (TNR), which is the proportion of negative samples that were correctly classified as negative.</p><p>The TNR fraction is also called specificity. Thus, the ROC-curve depicts sensitivity, i.e., recall, compared to the difference: 1 -specificity. The scalar indicator that follows from this indicator and allows us to compare classifiers is the value of the area under the curve (AUC). A perfect classifier will have an area under the ROC-curve (AUC-ROC) equal to 1, while a random classifier will have an area of 0.5.</p><p>The ROC chart helps us decide where to impose a classification threshold for maximizing a true positive rate or minimizing a pseudo-positive rate, which is ultimately a business decision.</p><p>The scalar indicator that follows from this indicator and allows us to compare classifiers is the value of the area under the curve (AUC). A perfect classifier will have an area under the ROC-curve (AUC-ROC) equal to 1, while a random classifier will have an area of 0.5.</p><p>The ROC chart helps us decide where to impose a classification threshold for maximizing a true positive rate or minimizing a pseudo-positive rate, which is ultimately a business decision.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.6.">PR-curve</head><p>The precision-recall curve determines the sensitivity to the ratio of classes. If the positive class is significantly smaller, the AUC-ROC may provide an inadequate estimate of the algorithm's quality, as it measures the proportion of incorrectly accepted objects in relation to the total number of negative ones.</p><p>We can eliminate this problem with unbalanced classes by passing from the ROC-curve to the precision-recall (PR) curve. The PR-curve is determined similarly to the ROC-curve; the only difference is that on the axes, we lay not FPR and TPR but recall (abscissa) and precision (ordinate). The criterion for the quality of the family of algorithms is the area under the PR-curve (AUC-PR).</p><p>The quality indicators of the classification considered here are far from all. Categorical Crossentropy or Log Loss / Binary Crossentropy and others may also be used in specific cases <ref type="bibr" target="#b6">[7]</ref>. Even from the above list, it can be seen that comparing classification algorithms and choosing the best one is a difficult task, especially for beginners. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Methods</head><p>Quite often, when using different classification algorithms, similar quality indicators are obtained and it is difficult for a user to choose one of them. This is especially true for the criterion of "highest efficiency", which is calculated for each system separately and depends on business objectives. This problem can be considered as a multicriteria optimization problem <ref type="bibr" target="#b10">[11]</ref>.</p><p>In practice, such multicriteria problems are solved quite successfully, but a strict mathematical solution of multicriteria optimization problems still does not exist. In practice, several approaches are used, each of which has advantages and disadvantages. Here the authors propose to apply the method for solving multicriteria problems based on a nonlinear scheme of compromises presented in work by Voronin A. M. <ref type="bibr" target="#b11">[12]</ref> having the form:</p><formula xml:id="formula_7">𝑥 * = arg min 𝐼 [𝐼 − 𝐼 (𝑥)] . (<label>8</label></formula><formula xml:id="formula_8">)</formula><p>where Imi is the upper limit for the partial criterion Ii.</p><p>If necessary, we can introduce the weight coefficients Ci into the nonlinear convolution (8) <ref type="bibr" target="#b12">[13]</ref>.</p><formula xml:id="formula_9">𝑥 * = arg min 𝐶 𝐼 [𝐼 − 𝐼 (𝑥)] .<label>(9)</label></formula><p>The introduction of coefficients allows us to give preference to one or another criterion being better adapted to the specific business task.</p><p>Since the best quality of the classification algorithm is needed, the efficiency criterion must be maximized, and then the calculation formula takes the following form:</p><formula xml:id="formula_10">𝑁𝑆𝐶 = arg max 𝐼 [𝐼 − 𝐼 (𝑥)] , (<label>10</label></formula><formula xml:id="formula_11">)</formula><p>where NSC is the scalar indicator on a nonlinear scheme of compromises.</p><p>As partial criteria of the classification quality it is suggested to use known indicators of quality: accuracy, recall, precision and F1. Then the calculation formula has the final form:</p><formula xml:id="formula_12">𝑁𝑆𝐶 = arg max 1 1 − 𝐴𝑐𝑐 + 1 1 − 𝑃𝑟 + 1 1 − 𝑅𝑒𝑐 + 1 1 − 𝐹1 (<label>11</label></formula><formula xml:id="formula_13">)</formula><p>where 𝐴𝑐𝑐 is accuracy; Pr is precision; 𝑅𝑒𝑐 is recall; 𝐹1 -F1-score.</p><p>The obtained scalar number will not have any physical meaning. Its values can also vary from dozens to dozens of thousands. The highest scalar value of the indicator NSC will determine the best algorithm for implementing a particular classification task.</p><p>The advantages of the method of nonlinear compromise scheme <ref type="bibr" target="#b13">[14]</ref> are, first of all, that this method is quite simple in terms of computational costs and allows us to obtain solutions from the Pareto set taking into account constraints on the principle of "as far from constraints as possible". Second, the scalar convolution <ref type="bibr" target="#b9">(10)</ref>, under convexity of partial criteria, has the property of unimodality (i.e., the problem becomes one-extreme one). Moreover, the nonlinear scheme of compromises has the property of continuous adaptation to different situations in which it is necessary to accept a multi-criteria solution. In the tense situations (when one or more partial criteria are in dangerous proximity to constraints) it acts equivalent to the minimax model; in the fairly calm situations, the convolution <ref type="bibr" target="#b9">(10)</ref> or <ref type="bibr" target="#b10">(11)</ref> acts equivalent to the model of integrated optimality (i.e., economic scheme of compromises). In the interval between the two poles, the nonlinear convolution gives different degrees of alignment of the partial criteria. Thus, the application of the nonlinear scheme of compromises allows us to increase the accuracy of the decision due to continuity of adaptation <ref type="bibr" target="#b14">[15]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Experiment</head><p>To verify the efficiency and evaluate the efficiency, experiments have been conducted on the application of the proposed metric together with the calculation of known indicators. The experiments were conducted using the Python programming language and a number of its libraries, such as scikitlearn, pandas and others <ref type="bibr" target="#b15">[16]</ref>.</p><p>The scikit-learn library has many classification algorithms that can be used to build a machine learning model. All scikit-learn machine learning models are implemented in their own classes, which are called Estimators .</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head></head><p>The following learning models have been created:  Logistic regression or logit model (LR);  Linear discriminant analysis (LDA);  K-nearest neighbors method (KNN);  Classification and regression with using trees (CART);  Naive Bayes classifier (NB);  Support vector machines (SVM);</p><p>We used a mixture of simple linear (LR and LDA) and nonlinear (KNN, CART, NB and SVM) algorithms.</p><p>As data for research, the "Iris" <ref type="bibr" target="#b16">[17]</ref>, well-known classical data set in machine learning and statistics, was used. It is included in the datasets module of the scikit-learn library and is run by the command: url = https://raw.githubusercontent.com/jbrownlee/Datasets/master/iris.csv</p><p>Based on the dataset, a machine learning model has been built, which predicts iris varieties for a new set of measurements. Before we apply our model to the new set, we need to make sure that the model actually works and its predictions can be trusted.</p><p>Unfortunately, we cannot use the data we took to build the model in order to evaluate the quality of the model. This is because our model memorizes the entire training set, and therefore it will always predict the correct label for any data point in the training set. This "memorization" tells us nothing about the quality of the model (in other words, we do not know if this model works properly on a new dataset).</p><p>To evaluate the effectiveness of the model, we present it with new labeled data. This is usually done by splitting the collected data (in this case, 150 flowers) into two parts. One piece of data is used to build our machine learning model and is called training data or training set. Other data will be used to evaluate the quality of the model, and they are called test data.</p><p>We will use stratified 10-fold cross-validation to improve the accuracy of the model <ref type="bibr" target="#b17">[18]</ref>. This is an additional procedure, and, in general, we do not need to use it when the amount of input data is great. The dataset of our case is 150 lines (50 of each type), which is relatively small. Therefore, it is needed to increase the accuracy of the model.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Results</head><p>Scikit-learn Python contains many built-in features for analyzing the performance of models. In this task, we use some of these metrics and have written our own quality assessment functions from scratch to compare them with known ones <ref type="bibr">[19 -23]</ref>.</p><p>The following indicators from sklearn.metrics are programmed:  confusion_matrix (matrix of errors or matrix of inaccuracies or confusion);  accuracy_score (accuracy);  recall_score (recall);  precision_score (precision);  F1_score (F-score);  roc_curve (ROC-curve);  roc_auc_score (AUC-ROC). Additionally, we propose our own indicator of the quality of classification algorithms -nonlinear scheme of compromises (NSC).</p><p>The libraries presented in Figure <ref type="figure" target="#fig_4">4</ref> were used for software development.</p><p>The obtained research results are presented in Tables <ref type="table" target="#tab_3">2-4</ref>.   Since the experiments were performed for a well-known data set, which was well balanced and tested, the additional experiments were performed for a two-class classification of a real data set with 15,758 copies. The results obtained are summarized in Table <ref type="table" target="#tab_4">5</ref>. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.">Discussions</head><p>Analysis of the results of experiment 1 (Tables <ref type="table" target="#tab_3">2-4</ref>) shows that in this case, we have a balanced date set for which all algorithms show high-quality values, close to the standard.</p><p>The comparison table for the class Iris-setosa (Table <ref type="table" target="#tab_1">2</ref>) shows that if all other indicators are equal, it is possible to determine based on a single indicator (accuracy) and it makes no sense to calculate the NSC . With respect to a single indicator of accuracy, the highest accuracy was found for the SVM algorithm.</p><p>We should note that in tense moments, when the partial indicator attains the maximum '1', the calculation of NSC is not possible. In such cases, it is necessary to increase the accuracy of calculations, not to use rounding of digital values, or not to take into account such an indicator in the convolution. Another option may be to use instead of '1' the nearest number of required accuracy, for example, 0.99999.</p><p>Analysis of Table <ref type="table" target="#tab_2">3</ref> shows that in the classification of Iris-versicolor, the best NSC indicator is shown by the SVM algorithm.</p><p>Analysis of Table <ref type="table" target="#tab_3">4</ref> shows that in the classification of Iris-virginica, the best NSC indicator is also shown by the SVM algorithm.</p><p>For the second data set, approximately the same quality assessment results are observed for the algorithms of CART and SVM. Due to the calculation of the NSC, it is possible to reasonably determine the best algorithm, namely for this case of the CART classification.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: Classification Algorithms Azure Machine Learning Studio</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head></head><label></label><figDesc>True Positive (TP): actual = 1, predicted = 1;  False Positive (FP): actual = 0, predicted = 1;  False Negative (FN): actual = 1, predicted = 0;  True Negative (TN): actual = 0, predicted = 0.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>Figure 2 :</head><label>2</label><figDesc>Figure 2: ROC-curve</figDesc><graphic coords="4,121.80,391.56,351.48,256.80" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_3"><head>Figure 3 :</head><label>3</label><figDesc>Figure 3: PR-curve</figDesc><graphic coords="5,121.80,344.88,351.48,251.52" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_4"><head>Figure 4 :</head><label>4</label><figDesc>Figure 4: Used software libraries</figDesc><graphic coords="8,114.96,71.76,364.56,377.52" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1</head><label>1</label><figDesc></figDesc><table><row><cell>Classification error matrix</cell><cell></cell><cell></cell></row><row><cell></cell><cell>y = 1</cell><cell>y = 0</cell></row><row><cell>a(x) = 1</cell><cell>TP</cell><cell>FP</cell></row><row><cell>a(x) = 0</cell><cell>FN</cell><cell>TN</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 2</head><label>2</label><figDesc></figDesc><table><row><cell cols="2">Comparative table for Iris-setosa class</cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell>Classification</cell><cell>Accuracy</cell><cell>Recall</cell><cell>Precision</cell><cell>F1</cell><cell>NSC</cell></row><row><cell>algorithm</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell>LR</cell><cell>0,942</cell><cell>1,00</cell><cell>1,00</cell><cell>1,00</cell><cell>Non</cell></row><row><cell>LDA</cell><cell>0,975</cell><cell>1,00</cell><cell>1,00</cell><cell>1,00</cell><cell>Non</cell></row><row><cell>KNN</cell><cell>0,958</cell><cell>1,00</cell><cell>1,00</cell><cell>1,00</cell><cell>Non</cell></row><row><cell>CART</cell><cell>0,95</cell><cell>1,00</cell><cell>1,00</cell><cell>1,00</cell><cell>Non</cell></row><row><cell>NB</cell><cell>0,95</cell><cell>1,00</cell><cell>1,00</cell><cell>1,00</cell><cell>Non</cell></row><row><cell>SVM</cell><cell>0,983</cell><cell>1,00</cell><cell>1,00</cell><cell>1,00</cell><cell>Non</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>Table 3</head><label>3</label><figDesc>Comparative table for Iris-versicolor class Classification algorithm</figDesc><table><row><cell></cell><cell>Accuracy</cell><cell>Recall</cell><cell>Precision</cell><cell>F1</cell><cell>NSC</cell></row><row><cell>LR</cell><cell>0,942</cell><cell>0,918</cell><cell>1,00</cell><cell>0,958</cell><cell>10053</cell></row><row><cell>LDA</cell><cell>0,975</cell><cell>0,92</cell><cell>1,00</cell><cell>0,96</cell><cell>10078</cell></row><row><cell>KNN</cell><cell>0,958</cell><cell>0,92</cell><cell>1,00</cell><cell>0,96</cell><cell>10069</cell></row><row><cell>CART</cell><cell>0,95</cell><cell>0,92</cell><cell>1,00</cell><cell>0,96</cell><cell>10058</cell></row><row><cell>NB</cell><cell>0,95</cell><cell>0,92</cell><cell>1,00</cell><cell>0,958</cell><cell>10056</cell></row><row><cell>SVM</cell><cell>0,983</cell><cell>0,92</cell><cell>1,00</cell><cell>0,96</cell><cell>10096</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_3"><head>Table 4 table</head><label>4</label><figDesc></figDesc><table><row><cell cols="3">for Iris-virginica class</cell><cell></cell><cell></cell><cell></cell></row><row><cell>Classification</cell><cell>Accuracy</cell><cell>Recall</cell><cell>Precision</cell><cell>F1</cell><cell>NSC</cell></row><row><cell>algorithm</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell>LR</cell><cell>0,942</cell><cell>0,86</cell><cell>1,00</cell><cell>0,92</cell><cell>10037</cell></row><row><cell>LDA</cell><cell>0,975</cell><cell>0,83</cell><cell>1,00</cell><cell>0,91</cell><cell>10058</cell></row><row><cell>KNN</cell><cell>0,958</cell><cell>0,86</cell><cell>1,00</cell><cell>0,92</cell><cell>10043</cell></row><row><cell>CART</cell><cell>0,95</cell><cell>0,88</cell><cell>1,00</cell><cell>0,93</cell><cell>10040</cell></row><row><cell>NB</cell><cell>0,95</cell><cell>0,86</cell><cell>1,00</cell><cell>0,92</cell><cell>10039</cell></row><row><cell>SVM</cell><cell>0,983</cell><cell>0,9</cell><cell>1,00</cell><cell>0,94</cell><cell>10083</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_4"><head>Table 5</head><label>5</label><figDesc>Comparative table of classification algorithms for real data set</figDesc><table><row><cell>Classification</cell><cell>Accuracy</cell><cell>Recall</cell><cell>Precision</cell><cell>F1</cell><cell>NSC</cell></row><row><cell>algorithm</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell>LR</cell><cell>0,57</cell><cell>0,61</cell><cell>0,68</cell><cell>0,64</cell><cell>10,79</cell></row><row><cell>LDA</cell><cell>0,61</cell><cell>0,62</cell><cell>0,71</cell><cell>0,66</cell><cell>11,58</cell></row><row><cell>KNN</cell><cell>0,67</cell><cell>0,63</cell><cell>0,69</cell><cell>0,66</cell><cell>11,9</cell></row><row><cell>CART</cell><cell>0,67</cell><cell>0,65</cell><cell>0,71</cell><cell>0,68</cell><cell>12,46</cell></row><row><cell>NB</cell><cell>0,65</cell><cell>0,63</cell><cell>0,68</cell><cell>0,65</cell><cell>11,54</cell></row><row><cell>SVM</cell><cell>0,67</cell><cell>0,66</cell><cell>0,70</cell><cell>0,68</cell><cell>12,43</cell></row></table></figure>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="7.">Conclusions</head><p>Thus, in the paper, we propose for evaluating the classification quality to use the developed scalar quality indicator, which is a scalar nonlinear convolution of the known quality indicators, such as accuracy, recall, precision, F1, AUC_ROC.</p><p>This indicator allows us to give preference to one or another classification algorithm when the values of typical indicators are almost the same or have contradictions.</p><p>Studies have confirmed the usefulness of the proposed indicator NSC.</p><p>In the future, it is advisable to study the scalar indicator of quality in more detail, determine its limits and develop recommendations for the use of the NSC.</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<monogr>
		<author>
			<persName><forename type="first">M</forename><surname>Mitchell</surname></persName>
		</author>
		<title level="m">Artificial Intelligence. A Guide for Thinking Humans</title>
				<meeting><address><addrLine>London</addrLine></address></meeting>
		<imprint>
			<publisher>Penguin</publisher>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<monogr>
		<title level="m" type="main">Engineering MLOps</title>
		<author>
			<persName><forename type="first">E</forename><surname>Raj</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2021">2021</date>
			<publisher>Packt Publishing</publisher>
			<pubPlace>Birmingham</pubPlace>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Evaluation of Classification Models in Machine Learning</title>
		<author>
			<persName><forename type="first">J</forename><surname>Dj</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Novakovi´c</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">S</forename><surname>Veljovic´</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Ilic´</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Papic´</surname></persName>
		</author>
		<author>
			<persName><surname>Tomovic´</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Theory and Applications of Mathematics &amp; Computer Science</title>
		<imprint>
			<biblScope unit="volume">7</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page" from="39" to="46" />
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<monogr>
		<title level="m" type="main">Mastering Azure Machine Learning. Perform large-scale end-to-end advanced machine learning in the cloud with Microsoft Azure Machine Learning</title>
		<author>
			<persName><forename type="first">C</forename><surname>Körner</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Waaijer</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2020">2020</date>
			<publisher>Packt Publishing</publisher>
			<pubPlace>Birmingham</pubPlace>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Method of Machine Learning Based on Discrete Orthogonal Polynomials of Chebyshev</title>
		<author>
			<persName><forename type="first">I</forename><surname>Puleko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Kravchenko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Chumakevych</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Ptashnyk</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 4th International Conference on Computational Linguistics and Intelligent Systems, COLINS 2020</title>
				<meeting>the 4th International Conference on Computational Linguistics and Intelligent Systems, COLINS 2020<address><addrLine>Lviv</addrLine></address></meeting>
		<imprint>
			<publisher>CEUR</publisher>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="67" to="76" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<monogr>
		<ptr target="https://developers.google.com/machine-learning/crash-course/classification/true-false-positive-negative" />
		<title level="m">Machine Learning Crash Course, Classification: True vs. False and Positive vs. Negative</title>
				<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<monogr>
		<author>
			<persName><forename type="first">S</forename><surname>Minaee</surname></persName>
		</author>
		<ptr target="https://towardsdatascience.com/20-popular-machine-learning-metrics-part-1-classification-regression-evaluation-metrics-1ca3e282a2ce" />
		<title level="m">20 Popular Machine Learning Metrics. Part 1: Classification &amp; Regression Evaluation Metrics</title>
				<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<monogr>
		<title level="m" type="main">Beyond Accuracy: Precision and Recall</title>
		<author>
			<persName><forename type="first">W</forename><surname>Koehrsen</surname></persName>
		</author>
		<ptr target="https://towardsdatascience.com/beyond-accuracy-precision-and-recall-3da06bea9f6c" />
		<imprint>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<monogr>
		<title level="m" type="main">Assessing and Comparing Classifier Performance with ROC Curves</title>
		<author>
			<persName><forename type="first">J</forename><surname>Brownlee</surname></persName>
		</author>
		<ptr target="https://machinelearningmastery.com/assessing-comparing-classifier-performance-roc-curves-2/" />
		<imprint>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Classification Model Evaluation Metrics</title>
		<author>
			<persName><forename type="first">Ž</forename><forename type="middle">Đ</forename><surname>Vujović</surname></persName>
		</author>
		<idno type="DOI">10.14569/IJACSA.2021.0120670</idno>
	</analytic>
	<monogr>
		<title level="j">International Journal of Advanced Computer Science and Applications</title>
		<imprint>
			<biblScope unit="volume">12</biblScope>
			<biblScope unit="issue">6</biblScope>
			<biblScope unit="page" from="599" to="606" />
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<monogr>
		<title level="m" type="main">Vector Optimization of Dynamical Systems</title>
		<author>
			<persName><forename type="first">A</forename><surname>Voronin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Yu</forename><surname>Ziatdinov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Kozlov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Chabaniuk</surname></persName>
		</author>
		<imprint>
			<date type="published" when="1999">1999</date>
			<publisher>Tehnika</publisher>
			<pubPlace>Kyiv</pubPlace>
		</imprint>
	</monogr>
	<note>in russian</note>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Nonlinear Tradeoff Scheme in Multicriteria Estimation and Optimization Problems</title>
		<author>
			<persName><forename type="first">A</forename><surname>Voronin</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Сybernetics and Systems Analysis</title>
		<imprint>
			<biblScope unit="volume">45</biblScope>
			<biblScope unit="issue">4</biblScope>
			<biblScope unit="page" from="106" to="115" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Methods of Simplifying the Problem of Nonlinear Programming on the Basis of Classification of Limitations</title>
		<author>
			<persName><forename type="first">A</forename><surname>Zasjadko</surname></persName>
		</author>
		<idno type="DOI">10.30748/soi.2020.161.07</idno>
	</analytic>
	<monogr>
		<title level="j">Information Processing Systems</title>
		<imprint>
			<biblScope unit="volume">161</biblScope>
			<biblScope unit="page" from="59" to="70" />
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<monogr>
		<author>
			<persName><forename type="first">A</forename><surname>Voronin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Yu</forename><surname>Ziatdinov</surname></persName>
		</author>
		<title level="m">Theory and Practice of Multicriteria Decisions: Models, Methods, Implementation</title>
				<imprint>
			<publisher>Lambert Academic Publishing</publisher>
			<date type="published" when="2013">2013</date>
		</imprint>
	</monogr>
	<note>in russian</note>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">Non-Linear Trade-off Scheme in Multicriteria Decision-Making Problems</title>
		<author>
			<persName><forename type="first">A</forename><surname>Voronin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Yu</forename><surname>Ziatdinov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Varlamov</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">International Journal &quot;Information Technologies &amp; Knowledge</title>
		<imprint>
			<biblScope unit="volume">11</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page" from="3" to="22" />
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">Hands-On Machine Learning with Scikit-Learn and TensorFlow</title>
		<author>
			<persName><forename type="first">Aurelien</forename><surname>Geron</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Concepts, Tools, and Techniques to Build Intelligent Systems</title>
				<meeting><address><addrLine>Boston</addrLine></address></meeting>
		<imprint>
			<publisher>O&apos;REILLY</publisher>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<monogr>
		<title level="m" type="main">Practical Machine Learning For Data Analysis Using Python</title>
		<author>
			<persName><forename type="first">A</forename><surname>Subasi</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2020">2020</date>
			<publisher>Academic Press is an imprint of Elsevier</publisher>
			<pubPlace>London</pubPlace>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<monogr>
		<title level="m" type="main">Classic Computer Science Problems in Python</title>
		<author>
			<persName><forename type="first">D</forename><surname>Kopec</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2019">2019</date>
			<publisher>Manning Publications Co</publisher>
			<pubPlace>New York</pubPlace>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<monogr>
		<title level="m" type="main">Data Science Fundamentals for Python and MongoDB</title>
		<author>
			<persName><forename type="first">D</forename><surname>Paper</surname></persName>
		</author>
		<idno type="DOI">10.1007/978-1-4842-3597-3</idno>
		<imprint>
			<date type="published" when="2018">2018</date>
			<publisher>Apress</publisher>
			<pubPlace>New York</pubPlace>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<monogr>
		<title level="m" type="main">Mastering Machine Learning with Python in Six Steps</title>
		<author>
			<persName><forename type="first">M</forename><surname>Swamynathan</surname></persName>
		</author>
		<idno type="DOI">10.1007/978-1-4842-4947-5</idno>
		<imprint>
			<date type="published" when="2019">2019</date>
			<publisher>Apress</publisher>
			<pubPlace>New York</pubPlace>
		</imprint>
	</monogr>
	<note>second edition</note>
</biblStruct>

<biblStruct xml:id="b20">
	<monogr>
		<title level="m" type="main">Python Code for Evaluation Metrics in ML/AI for Classification Problems</title>
		<author>
			<persName><forename type="first">R</forename><surname>Lakshmanamoorthy</surname></persName>
		</author>
		<ptr target="https://analyticsindiamag.com/evaluation-metrics-in-ml-ai-for-classification-problems-wpython-code/" />
		<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b21">
	<monogr>
		<title level="m" type="main">Metrics to Evaluate Machine Learning Algorithms in Python</title>
		<author>
			<persName><forename type="first">J</forename><surname>Brownlee</surname></persName>
		</author>
		<ptr target="https://machinelearningmastery.com/metrics-evaluate-machine-learning-algorithms-python/" />
		<imprint>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b22">
	<monogr>
		<ptr target="https://scikit-learn.org/stable/modules/model_evaluation.html" />
		<title level="m">Metrics and Scoring: Quantifying the Quality of Predictions</title>
				<imprint>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
	<note>Scikit-learn</note>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
