<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Learning of Multi-valued Multithreshold Neural Units</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author role="corresp">
							<persName><forename type="first">Vladyslav</forename><surname>Kotsovsky</surname></persName>
							<email>vladyslav.kotsovsky@uzhnu.edu.ua</email>
							<affiliation key="aff0">
								<orgName type="institution">State University &quot;Uzhhorod National University&quot;</orgName>
								<address>
									<addrLine>Narodna Square 3</addrLine>
									<postCode>88000</postCode>
									<settlement>Uzhhorod</settlement>
									<country key="UA">Ukraine</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Learning of Multi-valued Multithreshold Neural Units</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">DAD2045A560AD2E5E1D32514192C0C70</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T18:15+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Multithreshold neuron</term>
					<term>multi-valued neuron</term>
					<term>machine learning</term>
					<term>neural network</term>
					<term>classification 1</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>The issues related to the use of multithreshold neural units in multiclass classification are treated in the paper. Two models of multi-valued k-threshold neurons are considered. Online and offline modifications of the learning algorithm are designed to train multithreshold neuron to solve multiclass classification tasks using simple and fast learning techniques. The conditions are found ensuring the finiteness of the training. The experiment results demonstrate the performance of multithreshold multiclass classifier on real-world datasets compared to some popular classifiers.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>Neural-like networks and systems have numerous applications in artificial intelligence <ref type="bibr" target="#b0">[1]</ref> and intelligent data analysis <ref type="bibr" target="#b1">[2]</ref>. They are used in modern hardware <ref type="bibr" target="#b2">[3]</ref> and software <ref type="bibr" target="#b3">[4]</ref> tools and products <ref type="bibr" target="#b4">[5,</ref><ref type="bibr" target="#b5">6]</ref>. The amazing capacities of artificial neural networks (ANN) are provided by the appropriate use of the network architecture <ref type="bibr" target="#b6">[7]</ref> and related learning techniques <ref type="bibr" target="#b7">[8,</ref><ref type="bibr" target="#b8">9]</ref>.</p><p>The synergy between the network architecture, the kind of network nodes and the network learning (or synthesis) procedures is very important in the practice of neural computations <ref type="bibr" target="#b9">[10]</ref>. Linear neural units with threshold activation functions <ref type="bibr" target="#b10">[11]</ref>, binary inputs and output were used in early models <ref type="bibr" target="#b11">[12]</ref>. This kind of computation units was inspired by the models of biological neurons from the brain study <ref type="bibr" target="#b10">[11]</ref>. But both the theoretical studies and practical applications showed the strong limitations of the basic neuron model of McCulloch and Pitts <ref type="bibr" target="#b11">[12,</ref><ref type="bibr" target="#b12">13]</ref> as well as difficulties related to the learning of threshold ANN <ref type="bibr" target="#b13">[14,</ref><ref type="bibr" target="#b14">15]</ref>. In order to overcome abovementioned limitations and difficulties, many more complicated models of neural devices were proposed <ref type="bibr" target="#b10">[11,</ref><ref type="bibr" target="#b11">12]</ref>. The overall majority of these models employed two ways to increase the network capacities by enhancing the power of network neurons <ref type="bibr" target="#b9">[10]</ref>. The first is based on the use of more sophisticated models of the aggregation of the input signals of the neural unit instead of the classical weighted sum of inputs <ref type="bibr" target="#b11">[12]</ref>, e.g., polynomial threshold units <ref type="bibr" target="#b11">[12,</ref><ref type="bibr" target="#b12">13]</ref>. The second approach consists in the use of more complicated activation functions instead of the step function <ref type="bibr" target="#b11">[12]</ref> from the Rosenblatt model <ref type="bibr" target="#b15">[16,</ref><ref type="bibr" target="#b16">17]</ref>. Both approaches have their pros and cons discussed in <ref type="bibr" target="#b9">[10]</ref><ref type="bibr" target="#b10">[11]</ref><ref type="bibr" target="#b11">[12]</ref><ref type="bibr" target="#b12">[13]</ref><ref type="bibr" target="#b13">[14]</ref> The multithreshold models were developed under the second approach <ref type="bibr" target="#b17">[18]</ref>. One of the earliest among them was the multithreshold threshold element <ref type="bibr" target="#b18">[19]</ref>. Binary multithreshold neuron with weight vector </p><formula xml:id="formula_0">        2 1 2 2 2 1 1, if , 1, / 2 , 0, if , 0,1, / 2 , jj jj t t j k y t t j k − +       =        wx wx (1)</formula><p>where ( )  <ref type="bibr" target="#b17">[18,</ref><ref type="bibr" target="#b19">20]</ref>, because they are activated when the sum of weighted inputs is within the one if given disjoint half-open intervals, which are specified by the ordered sequence of their thresholds <ref type="bibr" target="#b20">[21]</ref>.</p><p>But the increase in the recognition capability of multithreshold is not gratuitous. One must pay a high price for this, which consists in the difficulty of the learning of such units <ref type="bibr" target="#b6">[7,</ref><ref type="bibr" target="#b21">22]</ref>, because the respective learning task is NP-hard even in the case of a unit with two thresholds. The research has two main goals:</p><p>•</p><p>The study of the model of multi-valued multithreshold neuron that should effectively use the advantages of multiple thresholds, be suitable for the multiclass classification and admits fairly simple training techniques.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>•</head><p>The development of the learning algorithm for such units and the study of its fitness for intended applications in classification. The paper has the following structure. First, the works related to the topic of the study will be reviewed. Then, two models of multithreshold neural units will be considered: binary-valued and multi-valued, respectively. We will discuss its advantages and consider some downsides related to the complexity of their learning. In the next section two learning algorithms will be described, which are designed for the learning of a single k-threshold neuron. For both algorithms the conditions on the learning rate will be stated, which satisfy the finiteness of the learning in the case of their application to the learning of strongly k-separable sets. Next, the simulation results will be treated of the performance of trained multiclass k-threshold neural classifiers in the comparison with some other popular classifiers provided by Sklearn library <ref type="bibr" target="#b10">[11]</ref>. Finally, two last sections contain the discussion of obtained results and conclusions.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Related works</head><p>The study of multithreshold neural units has a long history <ref type="bibr" target="#b18">[19,</ref><ref type="bibr" target="#b22">23,</ref><ref type="bibr" target="#b23">24]</ref>. Multithreshold neural elements were introduced in the early studies in threshold logic <ref type="bibr" target="#b18">[19,</ref><ref type="bibr" target="#b24">25]</ref>. As mentioned above, the additional thresholds were proposed with intention to increase the capacities of basic singlethreshold element <ref type="bibr" target="#b18">[19,</ref><ref type="bibr" target="#b25">26]</ref>. Some properties of multithreshold neurons were stated in <ref type="bibr" target="#b21">[22,</ref><ref type="bibr" target="#b24">25,</ref><ref type="bibr" target="#b25">26]</ref>. These works mostly dealt with the recognition capacity of multithreshold elements <ref type="bibr" target="#b1">[2]</ref>. Issues related to the synthesis of multithreshold devices remained almost untouched, because few algorithms for training such multithreshold units and networks had been developed <ref type="bibr" target="#b17">[18,</ref><ref type="bibr" target="#b23">24]</ref>. Therefore, the applications of devices using multithreshold approach were almost unknown <ref type="bibr" target="#b26">[27]</ref> despite the better capabilities of multithreshold units compared to the classical linear threshold units <ref type="bibr" target="#b19">[20,</ref><ref type="bibr" target="#b25">26]</ref>. The hardness results from <ref type="bibr" target="#b14">[15,</ref><ref type="bibr" target="#b21">22]</ref> can explain these difficulties for the practical application of bithreshold systems to some extent. Nevertheless, as stated in <ref type="bibr" target="#b7">[8,</ref><ref type="bibr" target="#b9">10,</ref><ref type="bibr" target="#b27">28]</ref>, the lack of learning techniques for multithreshold systems caused the decline of interest in their study.</p><p>But recent advances in multithreshold logic changed the situation <ref type="bibr" target="#b6">[7,</ref><ref type="bibr" target="#b13">14]</ref>. One of the reasons were new approaches in the synthesis ANN with hidden layers consisting of neurons with bithreshold activation functions <ref type="bibr" target="#b13">[14,</ref><ref type="bibr" target="#b19">20]</ref>. They were developed on the base of the generalization of the Baum's synthesis algorithm <ref type="bibr" target="#b28">[29]</ref> for threshold networks in the case of bithreshold nodes <ref type="bibr" target="#b13">[14,</ref><ref type="bibr" target="#b27">28]</ref>.</p><p>The advance in the application of so-called bithreshold networks was stated in <ref type="bibr" target="#b0">[1,</ref><ref type="bibr" target="#b9">10]</ref>, where such networks were considered as the effective tools, which are capable to solve typical problems of intellectual data processing and computational intelligence. The limitations and downsides of the basic bithreshold ANN from <ref type="bibr" target="#b6">[7,</ref><ref type="bibr" target="#b13">14]</ref> were stated in <ref type="bibr" target="#b27">[28]</ref>. Hybrid models of the multiclass classifier with heterogenous hidden layers were proposed in <ref type="bibr" target="#b27">[28]</ref>, where other kinds of neural units (e.g., WTA and single-threshold) units were used in order to enhance network performance and reduce its drawbacks. It should be noted that bithreshold ANN can be useful not only in classifiers. Their potential applications are considerably wider <ref type="bibr" target="#b1">[2,</ref><ref type="bibr" target="#b5">6,</ref><ref type="bibr" target="#b7">8,</ref><ref type="bibr" target="#b8">9]</ref>. E.g., they were mentioned</p><formula xml:id="formula_1">0 t = − 1 k t + = +</formula><p>in design of powerful deep ANN providing the exponential improvement of the memorization capacity <ref type="bibr" target="#b15">[16]</ref>. The bithreshold approach primary was employed for the solution of real-valued problems <ref type="bibr" target="#b9">[10]</ref>. But it admits the generalization to the complex domain <ref type="bibr" target="#b13">[14]</ref>. The complex analogs of bithreshold activation could be proposed <ref type="bibr" target="#b29">[30]</ref> that extend the capacity of complex-valued threshold neural units. This allows the multithreshold approach in the proceeding of data in the complex domain <ref type="bibr" target="#b16">[17,</ref><ref type="bibr" target="#b27">28]</ref>.</p><p>It should be noted that the above-mentioned advance in the application of multithreshold systems is actually related to only bithreshold models <ref type="bibr" target="#b6">[7]</ref>. The examples of successful application of general multithreshold models with an arbitrary number of thresholds are unknown <ref type="bibr" target="#b13">[14,</ref><ref type="bibr" target="#b29">30]</ref>. It became evident that the additional study is necessary before such models can be employed in machine learning systems <ref type="bibr" target="#b9">[10]</ref>. One of them was the paper <ref type="bibr" target="#b21">[22]</ref>, where general k-threshold neural units were treated in the case 2 k  . As was observed in [22], the parity of k has the great influence to the properties of multithreshold neurons. Moreover, every multithreshold unit can be realized using a small threshold circuit, and, consequently, every multithreshold network can be replaced by the equivalent networks consisting solely of bithreshold and threshold nodes <ref type="bibr" target="#b29">[30]</ref>. Notice also that unlike the learning of a single threshold linear unit, the learning of a multithreshold unit proved to be NP-hard <ref type="bibr" target="#b21">[22]</ref> confirming the similar result of the intractability of the learning of a single bithreshold unit <ref type="bibr" target="#b13">[14]</ref>.</p><p>Notice that all mentioned applications of bithreshold and k-threshold neurons have the binary outputs <ref type="bibr" target="#b27">[28]</ref>. Thus, their employment in the classifiers requires the special shape of the network output layer with a separate neuron for every class and the using of "one versus all" approach in the learning or synthesis <ref type="bibr" target="#b10">[11]</ref>. In some cases, a single output multi-valued neuron is preferable <ref type="bibr" target="#b11">[12]</ref>, because its application results in the network having fewer nodes and weight coefficients.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Models and methods</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1.">Two models of multithreshold neural units</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1.1.">Model of binary-valued k-threshold neuron</head><p>Let us consider again a model of k-threshold binary-valued neuron with the weight vector w and (ordered) threshold vector t, which output is given by <ref type="bibr" target="#b0">(1)</ref>. Note that its performance can be described as follows: </p><formula xml:id="formula_2">) / 2, if . k kk k k t tt y tt t −         =   + −      + −    wx wx wx wx (2)<label>1 ( 1)</label></formula><p>Model (2) has a simple geometrical interpretation <ref type="bibr" target="#b21">[22,</ref><ref type="bibr" target="#b25">26]</ref>. The family of parallel hyperplanes</p><formula xml:id="formula_3">:, jj Ht = wx   1, ..., jk </formula><p>divides the space n R by 1 k + parts, which can be successively labeled by numbers 0, 1, …, k. All points belonging to "even" parts are attributed as "negative" ones. Remaining parts are considered as "positive" <ref type="bibr" target="#b21">[22]</ref>. The illustration is shown in Figure <ref type="figure" target="#fig_2">1</ref>, where the case 2, 3 nk == is considered. Figure <ref type="figure" target="#fig_2">1</ref> can also illustrate the nature of difficulties related to the application of binary-valued multithreshold neuron. Its value can alternate many times. It ensures the great capability of the multithreshold unit on the one hand, but results in the hardness of its training on the other hand. Note that the strict proof of the NP-hardness of the learning of a single binary-valued multithreshold neuron can be found in <ref type="bibr" target="#b21">[22]</ref>. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1.2.">Model of multi-valued k-threshold neuron</head><p>The multi-valued modification of the model ( <ref type="formula">2</ref>) can be considered <ref type="bibr" target="#b17">[18,</ref><ref type="bibr" target="#b22">23]</ref> that keeps the capacity of the base model and is easier in the training <ref type="bibr" target="#b23">[24]</ref>. This multithreshold model uses the same weight vector w and threshold vector t, but differs in the output range of the neuron. To be more precise, the range set of k-threshold multi-valued neuron is</p><formula xml:id="formula_4">  1 0,1,,..., k k + = Z</formula><p>, and the neuron output y satisfies the following condition:</p><p>( )</p><formula xml:id="formula_5">yf = t wx,<label>(3)</label></formula><p>where ( ) </p><formula xml:id="formula_6">      =   −       t (4)</formula><p>Consider again the geometrical illustration, now, for the k-threshold multi-valued neuron (3), (4). As it is shown in Figure <ref type="figure" target="#fig_3">2</ref>, the performance of the neuron is also defined by parallel hyperplanes :</p><formula xml:id="formula_7">jj Ht = wx ,   1, ..., jk </formula><p>, which make partition of the space n R by 1 k + parts. These parts also are labeled by indices 0, 1, …, k corresponding to the output value of the (multi-valued) neuron whose activation is given by (4). Notice that same points are used in both Figure <ref type="figure" target="#fig_2">1</ref> and Figure <ref type="figure" target="#fig_3">2</ref>, but their partition by classes differs, because there are only two classes for binary-valued k-threshold neuron and 1 k + -for its many-valued counterpart [22].</p><p>The pair (w, t) completely defines the multi-valued multithreshold neuron and is called its structure pair. Let A be an arbitrary set in n R . Then every multi-valued k-threshold neuron with structure pair ( ) </p><formula xml:id="formula_8">( )   | , 0,1,..., i A A f i i k =   = = t x w x<label>(5)</label></formula><p>This partition is called an ordered k-threshold partition of the set A, whereas sets 01 , ,..., k A A A are called strongly k-separable (compare with <ref type="bibr" target="#b21">[22]</ref>). Note that the order matters for the strongly separated sets. Sets 01 , ,...,</p><formula xml:id="formula_9">k A A A are called k-separable, if there exists a permutation 1 : k  + → Z 1 k + Z such that sets ( ) ( ) ( ) 01 , ,..., k A A A   </formula><p>are strongly k-separable <ref type="bibr" target="#b21">[22]</ref>. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.">Learning algorithms</head><p>,..., ( ,..., ,0,...,0,1,0,...,0).</p><formula xml:id="formula_11">j n n j k j x x x x −− = a<label>(7)</label></formula><p>The chained inequality</p><formula xml:id="formula_12">1 jj tt +    wx is equal to the system 1 0, 0. j j t t +  −     −  +   </formula></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>wx wx</head><p>The last system can be rewritten in the following way:</p><formula xml:id="formula_13">( ) ( ) 1 0, 0. j j+     −     a x v a x v<label>(8)</label></formula><p>Thus, we can reduce system (6) to the following system: A A A = − − ( X denotes the cardinality of the set X) and vectors bi are obtained using <ref type="bibr" target="#b6">(7)</ref> and <ref type="bibr" target="#b7">(8)</ref>. Note that there are algorithms solving (9) in polynomial time <ref type="bibr" target="#b12">[13]</ref>. Thus, the task of the learning of k-threshold multi-valued neuron (3)-( <ref type="formula">4</ref>) is not NP-complete.</p><p>The reduction process can be described using the following pseudocode:  A from the previous subsection and an adopted version of the relaxa- tion algorithm from <ref type="bibr" target="#b30">[31,</ref><ref type="bibr" target="#b31">32]</ref>. The pseudocode of the algorithm is shown in the function Online-Multithreshold: A -an ordered partition corresponding to strongly k-separable sets, r-the number of learning epochs, 0 v -initial approxima- tion,  -the schedule function that defines the behavior of the learning rate. The above algorithm uses three internal counters: i that is responsible for learning epochs, j-responsible for learning corrections, and err-responsible for the unit errors during the current epoch of learning. The goal of algorithm is the search of a vector</p><formula xml:id="formula_14">ReduceSet ( ) 01 , , , k A A A 1 B  2 for x in 0 A : 3 add ( ) 1 −ax into B 4 for i in   1,..., 1 k − : 5 for x in</formula></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.2.">Online learning algorithm</head><formula xml:id="formula_15">OnlineMultithreshold 1 0 0 ) ( , , , ,,, k r A A A  v 1 B  ReduceSet ( ) 01 , , , k A A A 2 0  vv 3 ( ) ( ) , ,<label>0</label></formula><formula xml:id="formula_16">nk +  vR such that for all B  b</formula><p>the inequality 0  vb holds. If such vector is already found, then the learning process terminates. Otherwise, the weight correction occurs in step 13 at least once per epoch. Note that this correction is successful only in the case 0 s  . Thus, a random initial approximation should be used for 0 v to avoid the situation 0 s = during the learning. The following proposition states conditions ensuring the successful completion of the online learning using above algorithm.</p><formula xml:id="formula_17">Proposition 1. If 01 ... k A A A A =    , sets 01 , ,..., k A A A finite and strongly k-separable, (<label>) ( ) ( ) ( )</label></formula><formula xml:id="formula_18">j jj sj     =+ ,</formula><p>where j is a correction step, s(j) is the dot product obtained in step 8 before jth correction, ( )</p><formula xml:id="formula_19">02 j   , (<label>)</label></formula><formula xml:id="formula_20">min max 0 j          ,<label>(10)</label></formula><p>then there exists r such that OnlineMultithreshold produces a structure pair ( ) , wt of multi-valued k-threshold neuron, which satisfies <ref type="bibr" target="#b5">(6)</ref> and performs desired partition of the set A.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.3.">Offline learning algorithm</head><p>Let us describe the offline approach to the learning of k-threshold multi-valued neural unit. It is designed using the modification of offline spectral algorithm from <ref type="bibr" target="#b31">[32]</ref> adopted to solving the system <ref type="bibr" target="#b8">(9)</ref>. Let   1 ,..., m B = bb be a finite subset of nk + R , and nk +  vR . We will need the following notations:</p><formula xml:id="formula_21">( ) ( )<label>( ) ( ) ( ) ( ) ( ) ( ) ( ) 1 11</label></formula><p>, sgn , ,..., ,</p><formula xml:id="formula_22">mm i m i i ii B g B g g B g == = =  = =  v v v v v v s b b v b g b b s b b ,<label>(1,...,1) nk +</label></formula><formula xml:id="formula_23">= 1 .</formula><p>Note that both ( ) B s and ( ) </p><formula xml:id="formula_24">  : 1,0,1 B →− v g</formula><p>. Consider the following algorithm:</p><formula xml:id="formula_25">OfflineMultithreshold 1 0 0 ) ( , , , ,,, k r A A A  v 1 B  ReduceSet ( ) 01 , , , k A A A 2 0  vv 3 compute s(B) 4 compute ( ) B v g 5 0 j  6 while jr  and ( ) B  v g1 : 7 1 jj + 8 compute ( ) B v s 9 ( ) ( ) ( ) ( ) ( ) ( ) ( ) ( ) ( ) 2 i B B BB BB  −  − − − v v v s s v v v s s ss 10 compute ( ) B v g 11 ( ) 1 ,..., n vv  w 12 ( ) 1 ,..., n n k vv ++  − − t 13 return , wt Note that OfflineMultithreshold 1 0 0 ) ( , , , ,,, k r A A A  v</formula><p>has identical input parameters as its online counterpart from the previous subsection.</p><p>The following proposition states conditions ensuring the successful completion of the offline learning using above algorithm.  Proofs of both propositions are omitted. They can be obtained using reasons similar to <ref type="bibr" target="#b31">[32]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Experiment and results</head><p>Consider the capability of our learning algorithms from the previous section to train a multivalued multithreshold-based classifier to solve the classification problems on some benchmarks. Let us compare their performance with well-known classification methods, such as classical perceptron, nearest neighbor classifier, random forest and feed-forward ANN (multilayer perceptron). Classifiers were compared on the following two real-world datasets: "balance-scale" (Balance Scale Weight &amp; Distance Database) and "dry-bean" (Dry Bean Dataset) <ref type="bibr" target="#b32">[33,</ref><ref type="bibr" target="#b33">34]</ref> provided by UC Irvine Machine Learning Repository <ref type="bibr" target="#b34">[35]</ref>. The datasets contain 625 and 13611 learning instances from 3 and 7 classes, respectively <ref type="bibr" target="#b32">[33,</ref><ref type="bibr" target="#b34">35]</ref>. The first dataset has 5 features, the second one-16 <ref type="bibr" target="#b32">[33]</ref>. 25% instances of every dataset were used as the test set, and the rest 75%-as the training set. In order to obtain consistent results <ref type="bibr" target="#b11">[12]</ref>, the repeated random subsampling validation <ref type="bibr" target="#b10">[11,</ref><ref type="bibr" target="#b35">36]</ref> was used. The learning experiments were repeated 500 times for every dataset and then obtained results were averaged concerning the accuracy on the training and test sets.</p><p>Default values of parameters recommended by Scikit-Learn library were used during training experiments for first four classical classifiers: 5 neighbors for nearest neighbor classifier, 1000 iterations for linear perceptron classifier, unbounded depth for random forest, one hidden layer with 100 nodes and 200 iterations for multilayer perceptron <ref type="bibr" target="#b35">[36]</ref>. The constant learning rate 2  = was used for both MultiThreshold algorithm as well as random initial approximations 00 , wt. Datasets are not provided with an ordered partition into classes <ref type="bibr" target="#b34">[35]</ref>. So, the classes were ordered using the alphabetical order induced by their labels. The following table contains results of experiments. By analyzing data from Table <ref type="table" target="#tab_4">1</ref>, we can conclude that:</p><p>• Both multithreshold algorithms performed well on the relatively easy small 3-class classification task on balance-scale dataset and the online modification had the second-best accuracy on the test set. • Classification on the dry-bean dataset was more difficult task for almost all classifiers considered during simulation. Learning for both linear perceptron and multilayer perceptron failed completely. Multi-valued multithreshold neuron yielded by OfflineMultithreshold performed better than neuron produced by online algorithm and had the best accuracy among all neural-like models, which were considered. But its accuracy was considerably worse than in the case of the use of random forest classifier.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Discussions</head><p>Two versions of the learning algorithm for multi-valued multithreshold neurons have been proposed. The simulation results prove that both algorithms are capable to yield networks, which are suitable for the solution of classification problems in the case when the number of classes is relatively small. But the performance of both algorithms decreases in the case when the number of classes increases. It seems that it is due to at least two reasons.</p><p>The first one is the small number of parameters of the multithreshold model compared to other classifiers, which often use "one versus all" scheme <ref type="bibr" target="#b10">[11,</ref><ref type="bibr" target="#b35">36]</ref>. It seems that above drawback can be overcome by using multithreshold networks <ref type="bibr" target="#b28">[29]</ref> or more powerful neuron models with multithreshold activation, e.g., polynomial neurons <ref type="bibr" target="#b22">[23,</ref><ref type="bibr" target="#b29">30,</ref><ref type="bibr" target="#b31">32]</ref>.</p><p>The second reason is caused by the nature of the datasets related to majority of classification problems. They contain training pairs, each of which consists of a pattern and its class label. In terms of the partition, we deal with an unordered partition while proposed learning algorithms are designed to process with strongly k-separable sets corresponding to an ordered partition. The question arises how to convert an unordered partition to an ordered one. The brute force is not effective due to fast growth of factorial. Numerous heuristics can be used in order to increase the performance of the multithreshold neurons. This is a problem that deserves a separate consideration.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.">Conclusions</head><p>The problem of the application of multithreshold multi-valued neural units has been considered. These units separate the sets of patterns in n-dimensional vector space using parallel hyperplanes. This ability allows them to become candidates for computational nodes of multiclass ANN classifiers. Thus, the development of learning methods for such networks is important.</p><p>The simplest case of this learning problem has been treated, namely, issues concerning the learning of a single multi-valued multithreshold neuron. Two approaches to the training of multithreshold neuron have been developed. Both of them require the simple preliminary patterns transformation in order to reduce a given multiclass task to corresponding binary classification task. The online version of the learning algorithm is simpler and often faster. The offline modification performs single correction during each learning epoch, usually is more expensive but often yield the neuron having a somewhat better accuracy of classification. The conditions have been stated ensuring the finiteness of the learning process in the case of application of both algorithms to the training of k-separable sets.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head></head><label></label><figDesc>compu- tation unit with n inputs 1 ,, n xx whose single binary output y is calculated by the following rule:</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head></head><label></label><figDesc>.............................................</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: Illustration of the performance of binary-valued 3-threshold neuron</figDesc><graphic coords="4,206.00,72.00,182.55,166.30" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_3"><head>Figure 2 :</head><label>2</label><figDesc>Figure 2: Illustration of the performance of multi-valued 3-threshold neuron</figDesc><graphic coords="4,206.65,571.16,181.70,170.55" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_4"><head>,</head><label></label><figDesc>set А, where:</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_6"><head></head><label></label><figDesc>Notice that the transformation (7) is used in steps 3, 6, 7, 9 ensuring the filling of the output set B.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_7"><head></head><label></label><figDesc>Consider the training of the multi-valued k-threshold neural unit to separate finite strongly kus describe the online learning algorithm for a multi-valued neural unit that uses ReduceSet ( )</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_9"><head></head><label></label><figDesc> , sets 01 , ,..., k A A A are finite and strongly k-separable, ), v(j) is a value of vector v after jth correction, then there exists r such that OfflineMultithreshold</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_10"><head>,</head><label></label><figDesc>wt of a multi-valued k-threshold neuron, which performs desired k-threshold partition of the set A.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>for multithreshold neurons 3.2.1. Initial reduction of the task</head><label></label><figDesc></figDesc><table><row><cell cols="2">Let</cell><cell cols="2">01 ,</cell><cell>,...,</cell><cell></cell><cell></cell></row><row><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell>, wt that performs the desired partition</cell></row><row><cell>(</cell><cell cols="2">01 ,</cell><cell>,...,</cell><cell>)</cell><cell></cell><cell></cell><cell>01 , A A</cell><cell>,...,</cell><cell>k A , which satisfies (5).</cell></row><row><cell></cell><cell cols="7">Consider how one can reduce the above task to the solution of the homogenous system of</cell></row><row><cell cols="7">linear inequalities in</cell><cell>nk + variables</cell><cell>mk 11 ,... , ,..., w w t t . It is possible to rewrite (3)-(5) as follows:</cell></row><row><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell>  w x</cell><cell>, tA if  x</cell><cell>,</cell></row><row><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell>10 1 , if j j t A +     j t   w x x</cell><cell>( 1</cell><cell>j k  </cell><cell>)</cell><cell>,</cell><cell>(6)</cell></row><row><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell>    w x</cell><cell>, kk if tA  x</cell><cell>.</cell></row><row><cell></cell><cell cols="4">Since sets</cell><cell>01 , A A</cell><cell cols="2">,...,</cell><cell>k A are finite and strongly k-separable, system (6) has solutions, which</cell></row><row><cell cols="8">compose n-dimensional convex set. If all non-strict inequalities in (6) were replaced by strict</cell></row><row><cell cols="8">ones, then resulting system would also have solutions. Let</cell><cell>v</cell><cell>=</cell><cell>(</cell><cell>nk 11 ,..., , ,..., w w t t − −</cell><cell>)</cell><cell>,</cell></row><row><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell>(</cell><cell>)</cell></row></table><note>k A A A be strongly k-separable finite sets. Consider the task of the search of a multivalued k-threshold neuron with structure pair ( ) k A A A of the set A that is the union of (disjoint) sets</note></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_4"><head>Table 1</head><label>1</label><figDesc>Simulation results on two real-world datasets</figDesc><table><row><cell>Classifier</cell><cell cols="4">Accuracy on training set (in %) Accuracy on test set (in %) balance-scale dry-bean balance-scale dry-bean</cell></row><row><cell>Perceptron</cell><cell>84.25</cell><cell>20.91</cell><cell>82.23</cell><cell>16.59</cell></row><row><cell>5-Nearest Neighbor</cell><cell>88.07</cell><cell>81.07</cell><cell>81.85</cell><cell>72.34</cell></row><row><cell>Random Forest</cell><cell>100</cell><cell>99.98</cell><cell>82.93</cell><cell>90.03</cell></row><row><cell>MLP Classifier</cell><cell>95.28</cell><cell>53.51</cell><cell>92.74</cell><cell>49.60</cell></row><row><cell>OnlineMultithreshold</cell><cell>88.84</cell><cell>57.23</cell><cell>83.89</cell><cell>51.22</cell></row><row><cell>OfflineMultithreshold</cell><cell>88.03</cell><cell>66.02</cell><cell>82.16</cell><cell>59.83</cell></row></table></figure>
		</body>
		<back>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">High-Performance artificial intelligence recommendation of quality research papers using effective collaborative approach</title>
		<author>
			<persName><forename type="first">V</forename><forename type="middle">K</forename><surname>Venkatesan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">T</forename><surname>Ramakrishna</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Batyuk</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Barna</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Havrysh</surname></persName>
		</author>
		<idno type="DOI">10.3390/systems11020081</idno>
	</analytic>
	<monogr>
		<title level="j">Systems</title>
		<imprint>
			<biblScope unit="volume">11</biblScope>
			<biblScope unit="issue">2</biblScope>
			<biblScope unit="page">81</biblScope>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Machine learning for predicting energy efficiency of buildings: a small data approach</title>
		<author>
			<persName><forename type="first">I</forename><surname>Izonin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Tkachenko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">A</forename><surname>Mitoulis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Faramarzi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Tsmots</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Mashtalir</surname></persName>
		</author>
		<idno type="DOI">10.1016/j.procs.2023.12.173</idno>
	</analytic>
	<monogr>
		<title level="j">Procedia Computer Science</title>
		<imprint>
			<biblScope unit="volume">231</biblScope>
			<biblScope unit="page" from="72" to="77" />
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Synthesis of the integer neural elements</title>
		<author>
			<persName><forename type="first">F</forename><surname>Geche</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Kotsovsky</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Batyuk</surname></persName>
		</author>
		<idno type="DOI">10.1109/STC-CSIT.2015.7325432</idno>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the International Conference on Computer Sciences and Information Technologies, CSIT 2015</title>
				<meeting>the International Conference on Computer Sciences and Information Technologies, CSIT 2015<address><addrLine>Lviv, Ukraine</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2015">2015</date>
			<biblScope unit="page" from="121" to="136" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Identification of authorship of Ukrainian-language texts of journalistic style using neural networks</title>
		<author>
			<persName><forename type="first">M</forename><surname>Lupei</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Mitsa</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Repariuk</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Sharkan</surname></persName>
		</author>
		<idno type="DOI">10.15587/1729-4061.2020.195041</idno>
	</analytic>
	<monogr>
		<title level="j">Eastern-European Journal of Enterprise Technologies</title>
		<imprint>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="issue">2</biblScope>
			<biblScope unit="page" from="30" to="36" />
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Investigation of PNN optimization methods to improve classification performance in transplantation medicine</title>
		<author>
			<persName><forename type="first">M</forename><surname>Havryliuk</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Hovdysh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Tolstyak</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Chopyak</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Kustra</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">CEUR Workshop Proceedings</title>
				<imprint>
			<date type="published" when="2023">2023</date>
			<biblScope unit="volume">3609</biblScope>
			<biblScope unit="page" from="338" to="345" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Ethnocultural, educational and scientific potential of the interactive dialects map</title>
		<author>
			<persName><forename type="first">O</forename><surname>Mitsa</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Sharkan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Maksymchuk</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Varha</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Shkurko</surname></persName>
		</author>
		<idno type="DOI">10.1109/SIST58284.2023.10223544</idno>
	</analytic>
	<monogr>
		<title level="m">Proceedings of 2023 IEEE International Conference on Smart Information Systems and Technologies (SIST)</title>
				<meeting>2023 IEEE International Conference on Smart Information Systems and Technologies (SIST)<address><addrLine>Astana</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2023">2023</date>
			<biblScope unit="page" from="226" to="231" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Representational capabilities and learning of bithreshold neural networks</title>
		<author>
			<persName><forename type="first">V</forename><surname>Kotsovsky</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Batyuk</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Advances in Intelligent Systems and Computing</title>
				<editor>
			<persName><forename type="first">S</forename><surname>Babichev</surname></persName>
		</editor>
		<meeting><address><addrLine>Cham</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2021">2021</date>
			<biblScope unit="volume">1246</biblScope>
			<biblScope unit="page" from="499" to="514" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">An integral software solution of the SGTM neural-like structures implementation for solving different Data Mining tasks</title>
		<author>
			<persName><forename type="first">R</forename><surname>Tkachenko</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Lecture Notes on Data Engineering and Communications Technologies</title>
				<editor>
			<persName><forename type="first">S</forename><surname>Babichev</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">V</forename><surname>Lytvynenko</surname></persName>
		</editor>
		<meeting><address><addrLine>Cham</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2022">2022</date>
			<biblScope unit="volume">77</biblScope>
			<biblScope unit="page" from="696" to="713" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Analyzing Ukrainian media texts by means of support vector machines: aspects of language and copyright</title>
		<author>
			<persName><forename type="first">M</forename><surname>Lupei</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Mitsa</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Sharkan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Vargha</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Lupei</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Advances in Computer Science for Engineering and Education VI. ICCSEEA 2023</title>
		<title level="s">Lecture Notes on Data Engineering and Communications Technologies</title>
		<editor>
			<persName><forename type="first">Z</forename><surname>Hu</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">I</forename><surname>Dychka</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">M</forename><surname>He</surname></persName>
		</editor>
		<meeting><address><addrLine>Cham</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2023">2023</date>
			<biblScope unit="volume">181</biblScope>
			<biblScope unit="page" from="173" to="182" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Soft computing techniques for biomedical data analysis: open issues and challenges</title>
		<author>
			<persName><forename type="first">E</forename><forename type="middle">H</forename><surname>Houssein</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">E</forename><surname>Hosney</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">M</forename><surname>Emam</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><forename type="middle">M</forename><surname>Younis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">A</forename><surname>Ali</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><forename type="middle">M</forename><surname>Mohamed</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Artificial Intelligence Review</title>
		<imprint>
			<biblScope unit="volume">56</biblScope>
			<biblScope unit="page" from="2599" to="2649" />
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<monogr>
		<author>
			<persName><forename type="first">A</forename><surname>Géron</surname></persName>
		</author>
		<title level="m">Hands-On Machine Learning with Scikit-Learn, Keras, and TensorFlow: Concepts, Tools, and Techniques to Build Intelligent Systems</title>
				<meeting><address><addrLine>Sebastopol, CA</addrLine></address></meeting>
		<imprint>
			<publisher>O&apos;Reilly Media</publisher>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<monogr>
		<author>
			<persName><forename type="first">P</forename><surname>Setoodeh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Habibi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Haykin</surname></persName>
		</author>
		<title level="m">Nonlinear Filters: Theory and Applications</title>
				<meeting><address><addrLine>New York, NY</addrLine></address></meeting>
		<imprint>
			<publisher>Wiley</publisher>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Large-width machine learning algorithm</title>
		<author>
			<persName><forename type="first">M</forename><surname>Anthony</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Ratsaby</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Progress in Artificial Intelligence</title>
		<imprint>
			<biblScope unit="volume">9</biblScope>
			<biblScope unit="page" from="275" to="285" />
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">Feed-forward neural network classifiers with bithreshold-like activations</title>
		<author>
			<persName><forename type="first">V</forename><surname>Kotsovsky</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Batyuk</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of IEEE 17th International Scientific and Technical Conference on Computer Sciences and Information Technologies, CSIT 2022</title>
				<meeting>IEEE 17th International Scientific and Technical Conference on Computer Sciences and Information Technologies, CSIT 2022<address><addrLine>Lviv, Ukraine</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2022">2022</date>
			<biblScope unit="page" from="9" to="12" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">Training a 3-node neural network is NP-complete</title>
		<author>
			<persName><forename type="first">A</forename><surname>Blum</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Rivest</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Neural Networks</title>
		<imprint>
			<biblScope unit="volume">5</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page" from="117" to="127" />
			<date type="published" when="1992">1992</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">An exponential improvement on the memorization capacity of deep threshold networks</title>
		<author>
			<persName><forename type="first">S</forename><surname>Rajput</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Sreenivasan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Papailiopoulos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Karbasi</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Advances in Neural Information Processing Systems</title>
		<imprint>
			<biblScope unit="volume">16</biblScope>
			<biblScope unit="page" from="12674" to="12685" />
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">Unitary learning in conditional models for deep optics neural networks</title>
		<author>
			<persName><forename type="first">Z.-G</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y.-L</forename><surname>Xiao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Zhong</surname></persName>
		</author>
		<idno>volu- me 12565</idno>
	</analytic>
	<monogr>
		<title level="m">Proceedings of SPIE -The International Society for Optical Engineering</title>
				<meeting>SPIE -The International Society for Optical Engineering</meeting>
		<imprint>
			<date type="published" when="2023">2023</date>
			<biblScope unit="page">1256543</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">Analysis of nonseparable property of multi-valued multi-threshold neuron</title>
		<author>
			<persName><forename type="first">N</forename><surname>Jiang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Ma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Yang</surname></persName>
		</author>
		<idno type="DOI">10.1109/IJCNN.2008.4633825</idno>
	</analytic>
	<monogr>
		<title level="m">Proceedings of 2008 IEEE International Joint Conference on Neural Networks (IEEE World Congress on Computational Intelligence)</title>
				<meeting>2008 IEEE International Joint Conference on Neural Networks (IEEE World Congress on Computational Intelligence)<address><addrLine>Hong Kong, China</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2008">2008</date>
			<biblScope unit="page" from="413" to="419" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<analytic>
		<title level="a" type="main">Multi-threshold threshold elements</title>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">R</forename><surname>Haring</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Transactions on Electronic Computers EC</title>
		<imprint>
			<biblScope unit="volume">15</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page" from="45" to="65" />
			<date type="published" when="1966">1966</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<analytic>
		<title level="a" type="main">Characterization of multiple-valued threshold functions in the Vilenkin-Chrestenson basis</title>
		<author>
			<persName><forename type="first">I</forename><surname>Prokíc</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Multiple-Valued Logic and Soft Computing</title>
		<imprint>
			<biblScope unit="volume">34</biblScope>
			<biblScope unit="issue">3-4</biblScope>
			<biblScope unit="page" from="223" to="238" />
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<analytic>
		<title level="a" type="main">The separating capacity of a multithreshold threshold element</title>
		<author>
			<persName><forename type="first">R</forename><surname>Takiyama</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Transactions on Pattern Analysis and Machine Intelligence. PAMI</title>
		<imprint>
			<biblScope unit="volume">7</biblScope>
			<biblScope unit="page" from="112" to="116" />
			<date type="published" when="1985">1985</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b21">
	<analytic>
		<title level="a" type="main">Multithreshold neural units and networks</title>
		<author>
			<persName><forename type="first">V</forename><surname>Kotsovsky</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Batyuk</surname></persName>
		</author>
		<idno type="DOI">10.1109/CSIT61576.2023.10324129</idno>
	</analytic>
	<monogr>
		<title level="m">Proceedings of IEEE 18th International Conference on Computer Sciences and Information Technologies, CSIT 2023</title>
				<meeting>IEEE 18th International Conference on Computer Sciences and Information Technologies, CSIT 2023<address><addrLine>Lviv, Ukraine</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2023">2023</date>
			<biblScope unit="page" from="1" to="5" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b22">
	<analytic>
		<title level="a" type="main">Using three layer neural network to compute multi-valued functions</title>
		<author>
			<persName><forename type="first">N</forename><surname>Jiang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><forename type="middle">X</forename><surname>Yang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><forename type="middle">M</forename><surname>Ma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><forename type="middle">Z</forename><surname>Zhang</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Fourth International Symposium on Neural Networks</title>
				<meeting><address><addrLine>Nanjing, P.R. China</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2007-06-03">2007. June 3-7, 2007. 2007</date>
			<biblScope unit="volume">4493</biblScope>
			<biblScope unit="page" from="1" to="8" />
		</imprint>
	</monogr>
	<note>Part III</note>
</biblStruct>

<biblStruct xml:id="b23">
	<monogr>
		<title level="m" type="main">Learning multivalued multithreshold functions</title>
		<author>
			<persName><forename type="first">M</forename><surname>Anthony</surname></persName>
		</author>
		<idno>LSE- CDMA-2003-03</idno>
		<imprint>
			<date type="published" when="2003">2003</date>
		</imprint>
		<respStmt>
			<orgName>London School of Economics</orgName>
		</respStmt>
	</monogr>
	<note type="report_type">CDMA Research Report No</note>
</biblStruct>

<biblStruct xml:id="b24">
	<analytic>
		<title level="a" type="main">Incorporation of energy efficient computational strategies for clustering and routing in heterogeneous networks of smart city</title>
		<author>
			<persName><forename type="first">V</forename><forename type="middle">K</forename><surname>Venkatesan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Izonin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Periyasamy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Indirajithu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Batyuk</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">T</forename><surname>Ramakrishna</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Energies</title>
		<imprint>
			<biblScope unit="volume">15</biblScope>
			<biblScope unit="issue">20</biblScope>
			<biblScope unit="page">7524</biblScope>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b25">
	<analytic>
		<title level="a" type="main">The capacity of multilevel threshold function</title>
		<author>
			<persName><forename type="first">S</forename><surname>Olafsson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><forename type="middle">S</forename><surname>Abu-Mostafa</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Transactions on Pattern Analysis and Machine Intelligence</title>
		<imprint>
			<biblScope unit="volume">10</biblScope>
			<biblScope unit="issue">2</biblScope>
			<biblScope unit="page" from="277" to="281" />
			<date type="published" when="1988">1988</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b26">
	<analytic>
		<title level="a" type="main">Towards data normalization task for the efficient mining of medical data</title>
		<author>
			<persName><forename type="first">I</forename><surname>Izonin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Ilchyshyn</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Tkachenko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Gregus</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Shakhovska</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Strauss</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of 12th International Conference on Advanced Computer Information Technologies, ACIT 2022</title>
				<meeting>12th International Conference on Advanced Computer Information Technologies, ACIT 2022<address><addrLine>Ruzomberok, Slovakia</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2022">2022</date>
			<biblScope unit="page" from="480" to="484" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b27">
	<analytic>
		<title level="a" type="main">Hybrid 4-layer bithreshold neural network for multiclass classification</title>
		<author>
			<persName><forename type="first">V</forename><surname>Kotsovsky</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">CEUR Workshop Proceedings</title>
				<imprint>
			<date type="published" when="2023">2023</date>
			<biblScope unit="volume">3387</biblScope>
			<biblScope unit="page" from="212" to="223" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b28">
	<analytic>
		<title level="a" type="main">On the capabilities of multilayer perceptrons</title>
		<author>
			<persName><forename type="first">E</forename><forename type="middle">B</forename><surname>Baum</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Complexity</title>
		<imprint>
			<biblScope unit="volume">4</biblScope>
			<biblScope unit="issue">3</biblScope>
			<biblScope unit="page" from="193" to="215" />
			<date type="published" when="1988">1988</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b29">
	<analytic>
		<title level="a" type="main">On the size of weights for bithreshold neurons and networks</title>
		<author>
			<persName><forename type="first">V</forename><surname>Kotsovsky</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Batyuk</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Voityshyn</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of IEEE 16th International Conference on Computer Sciences and Information Technologies, CSIT 2021</title>
				<meeting>IEEE 16th International Conference on Computer Sciences and Information Technologies, CSIT 2021<address><addrLine>Lviv, Ukrain</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2021">2021</date>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="page" from="13" to="16" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b30">
	<analytic>
		<title level="a" type="main">Robust learning from discriminative feature feedback</title>
		<author>
			<persName><forename type="first">S</forename><surname>Dasgupta</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Sabato</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of Machine Learning Research</title>
				<meeting>Machine Learning Research</meeting>
		<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="volume">108</biblScope>
			<biblScope unit="page" from="973" to="982" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b31">
	<analytic>
		<title level="a" type="main">On-line relaxation versus off-line spectral algorithm in the learning of polynomial neural units</title>
		<author>
			<persName><forename type="first">V</forename><surname>Kotsovsky</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Batyuk</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Communications in Computer and Information Science</title>
				<editor>
			<persName><forename type="first">S</forename><surname>Babichev</surname></persName>
		</editor>
		<meeting><address><addrLine>Cham</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2020">2020</date>
			<biblScope unit="volume">1158</biblScope>
			<biblScope unit="page" from="3" to="21" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b32">
	<monogr>
		<title level="m" type="main">OpenML: A worldwide machine learning lab</title>
		<ptr target="https://openml.org" />
		<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b33">
	<analytic>
		<title level="a" type="main">Development of an interactive map within the implementation of actual state and public directions</title>
		<author>
			<persName><forename type="first">M</forename><surname>Lupei</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Shlahta</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Mitsa</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Horoshko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Tsybko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Gorbachuk</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 12th International Conference on Advanced Computer Information Technologies, ACIT 2022</title>
				<meeting>the 12th International Conference on Advanced Computer Information Technologies, ACIT 2022<address><addrLine>Ruzomberok, Slovakia</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2022">2022</date>
			<biblScope unit="page" from="384" to="387" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b34">
	<monogr>
		<title level="m" type="main">The UCI machine learning repository</title>
		<author>
			<persName><forename type="first">M</forename><surname>Kelly</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Longjohn</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Nottingham</surname></persName>
		</author>
		<ptr target="http://archive.ics.uci.edu" />
		<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b35">
	<analytic>
		<title level="a" type="main">Scikit-survival: A library for time-to-event analysis built on top of Scikit-learn</title>
		<author>
			<persName><forename type="first">S</forename><surname>Pölsterl</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Machine Learning Research</title>
		<imprint>
			<biblScope unit="volume">21</biblScope>
			<biblScope unit="page" from="1" to="6" />
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
