<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Matrix Deep Neural Network and Its Rapid Learning in Data Science Tasks</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Iryna</forename><surname>Pliss</surname></persName>
							<email>iryna.pliss@nure.ua</email>
							<affiliation key="aff0">
								<orgName type="laboratory">Control Systems Research Laboratory</orgName>
								<orgName type="institution" key="instit1">Kharkiv National University of Radio Electronics</orgName>
								<orgName type="institution" key="instit2">UKRAINE</orgName>
								<address>
									<addrLine>Nauky ave</addrLine>
									<postCode>14</postCode>
									<settlement>Kharkiv</settlement>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Olena</forename><surname>Boiko</surname></persName>
							<email>olena.boiko@ukr.net</email>
							<affiliation key="aff1">
								<orgName type="laboratory">Control Systems Research Laboratory</orgName>
								<orgName type="institution" key="instit1">Kharkiv National University of Radio Electronics</orgName>
								<orgName type="institution" key="instit2">UKRAINE</orgName>
								<address>
									<addrLine>Nauky ave</addrLine>
									<postCode>14</postCode>
									<settlement>Kharkiv</settlement>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Valentyna</forename><surname>Volkova</surname></persName>
							<email>v.volkova@samsung.com</email>
							<affiliation key="aff2">
								<orgName type="department">LLC R&amp;D (SRK)</orgName>
								<orgName type="institution" key="instit1">Samsung Electronics Ukraine Company</orgName>
								<orgName type="institution" key="instit2">UKRAINE</orgName>
								<address>
									<addrLine>Lva Tolstogo St</addrLine>
									<postCode>57</postCode>
									<settlement>Kyiv</settlement>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Yevgeniy</forename><surname>Bodyanskiy</surname></persName>
							<email>yevgeniy.bodyanskiy@nure.ua</email>
							<affiliation key="aff3">
								<orgName type="laboratory">Control Systems Research Laboratory</orgName>
								<orgName type="institution" key="instit1">Kharkiv National University of Radio Electronics</orgName>
								<orgName type="institution" key="instit2">UKRAINE</orgName>
								<address>
									<addrLine>Nauky ave</addrLine>
									<postCode>14</postCode>
									<settlement>Kharkiv</settlement>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Matrix Deep Neural Network and Its Rapid Learning in Data Science Tasks</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">2C18569ABCAA8D11DD369856F93C87ED</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T02:21+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>deep learning</term>
					<term>multilayer network</term>
					<term>data mining</term>
					<term>2D network</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>The matrix deep neural network and its learning algorithm are proposed. This system allows reducing the number of tunable weights due to the rejection of the operations of vectorizationdevectorization. It also saves the information between rows and columns of 2D inputs.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>I. INTRODUCTION</head><p>Nowadays, artificial neural networks (ANNs) are widely used to solve many problems arising in Data Science. Here, multilayer perceptron (MLP) <ref type="bibr" target="#b0">[1,</ref><ref type="bibr" target="#b13">[13]</ref><ref type="bibr" target="#b14">[14]</ref><ref type="bibr" target="#b15">[15]</ref><ref type="bibr" target="#b16">[16]</ref><ref type="bibr" target="#b17">[17]</ref><ref type="bibr" target="#b18">[18]</ref> is the most widely used. On the basis of MLP deep neural networks (DNNs) <ref type="bibr" target="#b1">[2]</ref><ref type="bibr" target="#b2">[3]</ref><ref type="bibr" target="#b3">[4]</ref><ref type="bibr" target="#b19">19,</ref><ref type="bibr" target="#b21">21]</ref> were developed, that have improved characteristics in comparison with their prototypes, namely traditional shallow neural networks.</p><p>In the general case, a multilayer perceptron that contains L information processing layers (  1 L − hidden layer and one output layer) realizes a nonlinear transformation that can be written in the form</p><formula xml:id="formula_0">( ) ( ) ( ) [ ] [ ] ( ) [ ] ( [ ] ( ) [ ] [ ] [ ] ( ) ( ) ( ) ( ) ( ) 1 1 2 1 1 ˆ1 1 ... 1 L L L L L Y k X k W k W k W k X k − − − = Ψ = Ψ − Ψ ×  × − Ψ Ψ −  </formula><p>where: - is discrete time index. In the DNN family, the most popular are the convolutional neural networks (CNNs) <ref type="bibr" target="#b20">[20,</ref><ref type="bibr" target="#b22">[22]</ref><ref type="bibr" target="#b23">[23]</ref><ref type="bibr" target="#b24">[24]</ref><ref type="bibr" target="#b25">[25]</ref> that are mainly designed to process images represented in the form of ( )</p><formula xml:id="formula_1">1 2 n n × -matrices ( ) ( ) { } 1 2 i i X k x k = (where 1 1 1, 2,..., i n = and 2 2 1, 2,..., i n =</formula><p>), which must be vectorized before submission to the network, i. e. they must be presented in the form of vectors <ref type="bibr" target="#b9">[10]</ref>, the dimension of which can be quite large, that leads to the effect of "curse of dimensionality".</p><p>This effect can be avoided by processing the original matrix using convolution, pooling and encoding operations.</p><p>As a result a vector of dimension smaller than ( )</p><formula xml:id="formula_2">1 2 1 n n × is fed to the perceptron's input.</formula><p>Although DNNs provide high quality of the information processing, their training time is too long, and the training process itself may require considerable computing resources. However, it is possible to speed up the information processing by bypassing the operations of vectorizationdevectorization, i.e. by storing information that will be processed not in the form of a vector, but in the form of a matrix.</p><p>The abovementioned problem is solved by the matrix neural networks <ref type="bibr" target="#b4">[5,</ref><ref type="bibr" target="#b5">6,</ref><ref type="bibr" target="#b10">11,</ref><ref type="bibr" target="#b11">12]</ref>, that are quite complex from the computational point of view.</p><p>In this connection, it seems expedient to develop architecture and algorithms for tuning a deep matrix neural network that is characterized by the simplicity of the numerical realization and high speed of its synaptic weights learning.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>II. ADAPTIVE BILINEAR MODEL</head><p>The proposed matrix DNN is based on the adaptive matrix bilinear model introduced earlier by the authors <ref type="bibr" target="#b6">[7,</ref><ref type="bibr" target="#b7">8]</ref> ( ) { } ( ) ( ) ( )</p><formula xml:id="formula_3">1 2 1 1 2 2 ˆˆ1 1 ,</formula><p>1, 2,..., ;</p><p>1, 2,...,</p><formula xml:id="formula_4">j j Y k y A k X k B k j n j n = = − − = =<label>(1)</label></formula><p>where ( )</p><formula xml:id="formula_5">1 A k − , ( ) 1 B k − are ( ) 1 1 n n × , (<label>)</label></formula><formula xml:id="formula_6">2 2 n n × -matrices ( ) ( ) ( ) ( ) ( ) ( ) ( ) ( ) ( ) ( ) ( ) ( ) T T 1 1 , 1 1 A B T T A A k A k k E k B k X k B k B k k X k A k E k η η = − + ×   × −   = − + ×   × −  (2)</formula><p>is used or its version optimized by speed <ref type="bibr" target="#b6">[7]</ref> that can be written as</p><formula xml:id="formula_7">( ) ( ) ( ) ( ) ( ( ) ( ) ( ) ( )) ( ) ( ) ( ( ) ( ) ( ) ( ) ( ) ( ) ( ) ( )) ( ) ( ) ( ) ( ) ( ) ( ) ( ) ( ) ( ( ) ( ) ( )) ( ) ( ( ) ( ) ( ) ( ) ( ) ( ) ( ) ( ) ( )) ( ) ( ) ( ) T T T T T T T 1 T T T T T T T T T 1 T T T T 1 1 1 1 1 1 1 1 , 1 1 , A A A A A A k A k Tr E k B k X k X k B k E k Tr E k B k X k X k B k B k X k X k B k E k E k B k X k B k B k Tr E k A k X k X k A k E k Tr A k X k X k A k E k E k A k X k X k A k X k A k E k − −  = − + − ×   × − ×   × − ×   × − − ×   × − ×   × −   = − + ×   × ×   × ×   × ×   × −   (3)</formula><p>that is the matrix generalization of the Kaczmarz-Widrow-Hoff learning algorithm (here</p><formula xml:id="formula_8">( ) A k η , ( ) B k η are learning rate parameters, ( ) ( ) ( ) ( ) ( ) ( ) ( ) ( ) ( ) ( )<label>1 1 , 1 ,</label></formula><formula xml:id="formula_9">A E k Y k A k X k B k E k Y k A k X k B k  = − − −   = − −   ( ) Y k is reference matrix signal).</formula><p>The learning algorithm in Eq. ( <ref type="formula">3</ref>) can be given additional filtering properties if the learning rate parameters in Eq. ( <ref type="formula">2</ref>) are calculated using the recurrence relations that can be written in the form</p><formula xml:id="formula_10">( ) ( ) ( ) ( ) ( ) ( ( ) ( ) ( ) ( ) ( ) ( ) ( ) ( )) 1 T T T T T 1 1 1 1 1 A A A k r k r k Tr E k B k X k X k B k B k X k X k B k E k η β − = = − + + − × × − × × − × × − and ( ) ( ) ( ) ( ) ( ) ( ) ( ( ) ( ) ( ) ( ) ( ) ( ) ( )) 1 T T T T T 1 B B B A A k r k r k Tr A k X k X k A k E k E k A k X k X k A k η β − = = − + × × × × ×</formula><p>where 0 1 β ≤ ≤ is smoothing parameter [9].</p><p>On the basis of the model from Eq. ( <ref type="formula" target="#formula_4">1</ref>), it is easy to introduce its nonlinear modification that can be written in the following form:</p><formula xml:id="formula_11">( ) { } ( ) ( ) ( ) ( ) ( ) 1 2 ˆˆ1 1 , j j Y k y A k X k B k U k = = Ψ − − = = Ψ  <label>(4)</label></formula><p>which is in fact the matrix generalization of the transformation that is realized by any of the layers of a multilayer perceptron.</p><p>In Eq. ( <ref type="formula" target="#formula_11">4</ref>) Ψ denotes a ( )</p><formula xml:id="formula_12">1 2</formula><p>n n × -matrix of activation functions, that acts elementwise on the matrix of internal activation signals of the system that are denoted by</p><formula xml:id="formula_13">( ) ( ) { } 1 2 j j U k u k = .</formula><p>In this case, the adjustment of the parameters of the nonlinear matrix model in Eq. ( <ref type="formula" target="#formula_11">4</ref>) can be realized on the basis of the modified δ -rule</p><formula xml:id="formula_14">( ) ( ) ( ) ( ) ( ) ( ) ( ) ( ) ( ) ( ) ( ) ( ) ( ) ( ) ( ) ( ) ( ) ( ) ( ) ( ) ( ) ( ) ( ) ( ) ( ) ( ) ( ) ( ) ( ) 1 2 1 2 1 2 2 1 2 1 2 1 2 2 1 2 1 2 1 2 1 1 2 1 2 1 1 2 1 2 1 2 1 1 2 1 2 1 2 1 1 2 1 2 1 1 1 1 1 ˆ1 ˆ, 1 1 1 j j j j A j j n j j j j i i i j j A j j j j i j j A j j i j j j j B A j j n A j j j j i i i j j B A j j a k a k k e k u k b k x k a k k e k u k x k a k k k x k b k b k k e k u k a k x k b k k e k η ψ η ψ η δ η ψ η ψ = = = − + × ′ × − = = − + × ′ × = − + + = − + × ′ × − = = − + × ′ × ∑ ∑ ( ) ( ) ( ) ( ) ( ) ( ) ( ) 1 2 2 1 2 1 2 2 ˆ1 ˆ. A j j i j j B A j j i u k x k b k k k x k η δ                    = − +   +   (5)</formula><p>On the basis of Eq. ( <ref type="formula" target="#formula_11">4</ref>) it is easy to introduce into consideration a multilayer matrix neural network that realizes the transformation</p><formula xml:id="formula_15">( ) [ ] ( ) [ ] ( ) ( ( ( [ ] ( ) ( ) [ ] ( ) ( ) ( ) [ ] ( ) )) [ ] ( ) ) 1 1 1 1 ˆ1 1 ... 1 1 ... 1 1 L L L L Y k A k A k A k X k B k B k B k − − = Ψ − Ψ − × × Ψ − − × × − −   <label>(6)</label></formula><p>Using the learning algorithm from Eq. ( <ref type="formula">5</ref>) and error backpropagation, it is possible to obtain the adaptive procedure for tuning all parameters of the matrix DNN in Eq. ( <ref type="formula" target="#formula_15">6</ref>): -for the output layer:</p><formula xml:id="formula_16">[ ] ( ) [ ] ( ) ( ) [ ] ( ) [ ] ( ) [ ] ( ) [ ] ( ) ( ) [ ] ( ) [ ] ( ) 1 2 1 2 1 2 1 1 2 1 2 1 2 2 1 1 1 , 1 L L L L j j j j A j j i L L L L j j j j B A j j A i a k a k k k o k b k b k k k o k η δ η δ − −  = − +   = − +   where [ ] ( ) [ ] ( ) ( ) ( ) 1 2 1 2 1 2 L L j j j j j j k u k e k δ ψ′ = , [ ] ( ) [ ] ( ) [ ] ( ) 2 1 1 2 1 2 2 1 1 1 ˆ1 n L L L i j j i i i o k b k o k − − = = − ∑ , [ ] ( ) [ ] ( ) ( ) ( ) 1 2 1 2 1 2 L L A j j A j j A j j k u k e k δ ψ′ = , [ ] ( ) [ ] ( ) [ ] ( ) 1 2 1 2 1 2 1 1 1 1 ˆn L L L A i j j i i i o k a k o k − − = = ∑ ; -for the l th hidden layer, 1 l L &lt; &lt; : [ ] ( ) [ ] ( ) ( ) [ ] ( ) [ ] ( ) [ ] ( ) [ ] ( ) ( ) [ ] ( ) [ ] ( ) 1 2 1 2 1 2 1 1 2 1 2 1 2 2 1 1 1 , 1 l l l l j j j j A j j i l l l l j j j j B A j j A i a k a k k k o k b k b k k k o k η δ η δ − −  = − +   = − +   where [ ] ( ) [ ] ( ) ( ) [ ] [ ] ( ) 1 1 2 1 2 1 2 1 2 1 1 1 1 n l l l l j j j j j j j j i k u k a k δ ψ δ + + = ′ = ∑ , [ ] ( ) [ ] ( ) [ ] ( ) 2 1 1 2 1 2 2 1 1 1 ˆ1 n l l l i j j i i i o k b k o k − − = = − ∑ , [<label>] ( ) [ ] ( ) ( )</label></formula><formula xml:id="formula_17">[ ] ( ) [ ] ( ) 2 1 2 1 2 1 2 1 2 2 1 1 1 n l l l l A j j A j j A j j j j i k u k k b k δ ψ δ + + = ′ = ∑ , [ ] ( ) [ ] ( ) [ ] ( ) 1 2 1 2 1 1 2 1 1 1 1 ˆn l l l A i j j i i i i o k a k o k − − = = ∑</formula><p>; -for the first hidden layer: Each observation is an image of size 28x28 pixels that represents a handwritten digit. In general the dataset has 10 classes (digits from 0 to 9). Some examples of the images from this dataset are presented in Fig. <ref type="figure" target="#fig_2">1</ref>.</p><formula xml:id="formula_18">[ ] ( ) [ ] ( ) ( ) [ ] ( ) [ ] ( ) [ ] ( ) [ ] ( ) ( ) [ ] ( ) [ ] ( ) 1 2 1 2 1 2 1 1 2 1 2 1 2 2 1 1 1 0 1 1 1 0 1 , 1 j j j j A j j i j j j j B A j j A i a k a k k k o k b k b k k k o k η δ η δ  = − +   = − +   where [ ] ( ) [ ] ( ) ( ) [ ] [ ] ( ) 1 1 2 1 2 1 2 1 2 1 1 1 2 2 1 n j j j j j j j j i k u k a k δ ψ δ = ′ = ∑ , [ ] ( ) [ ] ( ) ( ) 2 1 1 2 1 2 2 0 1 1 ˆ1 n i j j i i i o k b k x k = = − ∑ , [ ] ( ) [ ] ( ) ( ) [ ] ( ) [ ] ( ) 2 1 2 1 2 1 2 1 2 2 1 1 2 2 1 n A j j j j A j j j j i k u k k b k δ ψ δ = ′ = ∑ , [ ] ( ) [ ] ( ) ( )</formula><p>The elements of an image are represented by pixel values from 0 to 255, where 0 means white pixel (background) and 255 means black pixel (foreground). These values were preprocessed before training using normalization. The inputs for the network were ( )</p><formula xml:id="formula_19">1 2</formula><p>n n × -matrices, where 1 2 28 n n = = .</p><p>Every hidden layer also had size of 1 2 28 28 n n × = × .</p><p>The results of the computational experiments are presented in Table <ref type="table">1</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>TABLE 1. EXPERIMENTAL RESULTS</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Number of layers in the network</head><p>Error on test set, % </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>IV. CONCLUSION</head><p>In this paper the matrix deep neural network and its learning algorithm are proposed. They allow significantly to reduce the number of adjustable weights due to the rejection of the vectorization-devectorization operations of 2D input signals.</p><p>One of the main advantages of the proposed system is that it also preserves the information between rows and columns of 2D inputs of the system.</p><p>The considered DNN in comparison with traditional multilayer perceptrons has increased speed, determined by reduced number of adjustable parameters and optimization of the learning algorithm, and the simplicity of numerical implementation.</p><p>The proposed system can be used to solve a wide range of machine learning tasks, particularly connected with the problems of image processing, where input signals are presented to the system for data processing in the form of a matrix.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head></head><label></label><figDesc>,... k =</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head></head><label></label><figDesc>the proposed system and learning methods was demonstrated on the classification task. A number of experiments was carried out on the MNIST dataset that was introduced by Yann LeCun and Corinna Cortes [26]. This dataset is widely used for training and testing in machine learning, namely in the classification task. This dataset contains 60000 training observations and 10000 test observations.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>Fig. 1 .</head><label>1</label><figDesc>Fig.1. Examples of the images from the MNIST dataset.</figDesc><graphic coords="3,84.95,461.05,425.40,258.95" type="bitmap" /></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" xml:id="foot_0">ACIT 2018, June 1-3, 2018, Ceske Budejovice, Czech Republic</note>
		</body>
		<back>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<monogr>
		<title level="m" type="main">Neural Networks for Pattern Recognition</title>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">M</forename><surname>Bishop</surname></persName>
		</author>
		<imprint>
			<date type="published" when="1995">1995</date>
			<publisher>Clarendon Press</publisher>
			<pubPlace>Oxford</pubPlace>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Deep Learning</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Lecun</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Bengio</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Hinton</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Nature</title>
		<imprint>
			<biblScope unit="volume">521</biblScope>
			<biblScope unit="page" from="436" to="444" />
			<date type="published" when="2015">2015</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Deep Learning in neural networks: An overview</title>
		<author>
			<persName><forename type="first">J</forename><surname>Schmidhuber</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Neural Networks</title>
		<imprint>
			<biblScope unit="volume">61</biblScope>
			<biblScope unit="page" from="85" to="117" />
			<date type="published" when="2015">2015</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<monogr>
		<title level="m" type="main">Deep Learning</title>
		<author>
			<persName><forename type="first">I</forename><surname>Goodfellow</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Bengio</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Courville</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2016">2016</date>
			<publisher>MIT Press</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Neural networks with matrix inputs</title>
		<author>
			<persName><forename type="first">P</forename><surname>Daniušis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Vaitkus</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Informatica</title>
		<imprint>
			<biblScope unit="volume">19</biblScope>
			<biblScope unit="page" from="477" to="486" />
			<date type="published" when="2008">2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Matrix neural networks</title>
		<author>
			<persName><forename type="first">J</forename><surname>Gao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Guo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Wang</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the14th International Symposium on Neural Networks (ISNN), Part II</title>
				<meeting>the14th International Symposium on Neural Networks (ISNN), Part II<address><addrLine>Sapporo, Japan</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2017">2017</date>
			<biblScope unit="page" from="1" to="10" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Discrete adaptive identification and extrapolation of two-dimensional fields</title>
		<author>
			<persName><forename type="middle">V</forename><surname>Ye</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><forename type="middle">P</forename><surname>Bodyanskiy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><forename type="middle">A</forename><surname>Pliss</surname></persName>
		</author>
		<author>
			<persName><surname>Timofeev</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Pattern Recognition and Image Analysis</title>
		<imprint>
			<biblScope unit="volume">5</biblScope>
			<biblScope unit="page" from="410" to="416" />
			<date type="published" when="1995">1995</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<monogr>
		<title level="m" type="main">Neural Networks: A Comprehensive Foundation</title>
		<author>
			<persName><forename type="first">S</forename><surname>Haykin</surname></persName>
		</author>
		<imprint>
			<date type="published" when="1999">1999</date>
			<publisher>Prentice Hall, Inc</publisher>
			<pubPlace>Upper Saddle River, N. J.</pubPlace>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">On a non-parametric algorithm for smoothing parameter control in adaptive filtering</title>
		<author>
			<persName><forename type="first">S</forename><surname>Vorobyov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Ye</forename><surname>Bodyanskiy</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Engineering Simulation</title>
		<imprint>
			<biblScope unit="volume">16</biblScope>
			<biblScope unit="page" from="314" to="320" />
			<date type="published" when="1999">1999</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Deep learning for visual understanding: A review</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Guo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Oerlemans</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Lao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Wu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">S</forename><surname>Lew</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Neurocomputing</title>
		<imprint>
			<biblScope unit="volume">187</biblScope>
			<biblScope unit="page" from="27" to="48" />
			<date type="published" when="2016">2016</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">A vector matrix real time backpropagation algorithm for recurrent neural networks that approximate multi-valued periodic functions</title>
		<author>
			<persName><forename type="first">P</forename><surname>Stubberud</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">International Journal on Computational Intelligence and Application</title>
		<imprint>
			<biblScope unit="volume">8</biblScope>
			<biblScope unit="issue">4</biblScope>
			<biblScope unit="page" from="395" to="411" />
			<date type="published" when="2009">2009</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<monogr>
		<title level="m" type="main">New 2D Matrix-Based Neural Network for</title>
		<author>
			<persName><forename type="first">M</forename><surname>Mohamadian</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Afarideh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Babapour</surname></persName>
		</author>
		<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Image Processing Applications</title>
	</analytic>
	<monogr>
		<title level="j">IAENG (International Association of Engineers) International Journal of Computer Science</title>
		<imprint>
			<biblScope unit="volume">42</biblScope>
			<biblScope unit="issue">3</biblScope>
			<biblScope unit="page" from="265" to="274" />
			<date type="published" when="2015">2015</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<monogr>
		<author>
			<persName><forename type="first">K</forename><surname>Suzuki</surname></persName>
		</author>
		<title level="m">Artificial Neural Networks: Architectures and Applications</title>
				<meeting><address><addrLine>NY</addrLine></address></meeting>
		<imprint>
			<publisher>InTech</publisher>
			<date type="published" when="2013">2013</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<monogr>
		<title level="m" type="main">Neural Networks and Statistical Learning</title>
		<author>
			<persName><forename type="first">K</forename><forename type="middle">L</forename><surname>Du</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Swamy</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2014">2014</date>
			<publisher>Springer-Verlag</publisher>
			<pubPlace>London</pubPlace>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">Principles of Artificial Neural Networks</title>
		<author>
			<persName><forename type="first">D</forename><surname>Graupe</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Advanced Series in Circuits and Systems</title>
				<meeting><address><addrLine>Singapore</addrLine></address></meeting>
		<imprint>
			<publisher>World Scientific Publishing Co. Pte. Ltd</publisher>
			<date type="published" when="2007">2007</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<monogr>
		<author>
			<persName><forename type="first">L</forename><surname>Rutkowski</surname></persName>
		</author>
		<title level="m">Computational intelligence. Methods and techniques</title>
				<meeting><address><addrLine>Berlin-Heidelberg</addrLine></address></meeting>
		<imprint>
			<publisher>Springer-Verlag</publisher>
			<date type="published" when="2008">2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<monogr>
		<author>
			<persName><forename type="first">R</forename><surname>Kruse</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Borgelt</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Klawonn</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Moewes</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Steinbrecher</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Held</surname></persName>
		</author>
		<title level="m">Computational intelligence</title>
				<meeting><address><addrLine>Berlin</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2013">2013</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<monogr>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">T</forename><surname>Pham</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Liu</surname></persName>
		</author>
		<title level="m">Neural Networks for Identification, Prediction and Control</title>
				<meeting><address><addrLine>London</addrLine></address></meeting>
		<imprint>
			<publisher>Springer-Verlag</publisher>
			<date type="published" when="1995">1995</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<analytic>
		<title level="a" type="main">Deep machine learning -a new frontier in artificial intelligence research</title>
		<author>
			<persName><forename type="first">I</forename><surname>Arel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Rose</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Karnowski</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Computational Intelligence Magazine</title>
		<imprint>
			<biblScope unit="volume">5</biblScope>
			<biblScope unit="issue">4</biblScope>
			<biblScope unit="page" from="13" to="18" />
			<date type="published" when="2010">2010</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<analytic>
		<title level="a" type="main">Learning Convolutional Feature Hierachies for Visual Recognition</title>
		<author>
			<persName><forename type="first">K</forename><surname>Kavukcuoglu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Sermanet</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y-L</forename><surname>Boureau</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Gregor</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Mathieu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Lecun</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 23rd International Conference on Neural Information Processing Systems</title>
				<meeting>the 23rd International Conference on Neural Information Processing Systems</meeting>
		<imprint>
			<date type="published" when="2010">2010</date>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="page" from="1090" to="1098" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b21">
	<analytic>
		<title level="a" type="main">Multi-column deep neural networks for image classification</title>
		<author>
			<persName><forename type="first">C</forename><surname>Dan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">U</forename><surname>Meier</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Schmidhuber</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)</title>
				<meeting>the 2012 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)</meeting>
		<imprint>
			<date type="published" when="2012">2012</date>
			<biblScope unit="page" from="3642" to="3649" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b22">
	<analytic>
		<title level="a" type="main">ImageNet classification with deep convolutional neural networks</title>
		<author>
			<persName><forename type="first">A</forename><surname>Krizhevsky</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Sutskever</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">E</forename><surname>Hinton</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 25th International Conference on Neural Information Processing Systems (NIPS&apos;12)</title>
				<meeting>the 25th International Conference on Neural Information Processing Systems (NIPS&apos;12)</meeting>
		<imprint>
			<date type="published" when="2012">2012</date>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="page" from="1097" to="1105" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b23">
	<analytic>
		<title level="a" type="main">Deep Residual Learning for Image Recognition</title>
		<author>
			<persName><forename type="first">K</forename><surname>He</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Ren</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Sun</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)</title>
				<imprint>
			<biblScope unit="page" from="770" to="778" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b24">
	<analytic>
		<title level="a" type="main">Convolutional networks and applications in vision</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Lecun</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Kavukcuoglu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Farabet</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of 2010 IEEE International Symposium on Circuits and Systems (ISCAS)</title>
				<meeting>2010 IEEE International Symposium on Circuits and Systems (ISCAS)</meeting>
		<imprint>
			<date type="published" when="2010">2010</date>
			<biblScope unit="page" from="253" to="256" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b25">
	<analytic>
		<title level="a" type="main">Going deeper with convolutions</title>
		<author>
			<persName><forename type="first">C</forename><surname>Szegedy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Jia</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Sermanet</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Reed</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Anguelov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Erhan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Vanhoucke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Rabinovich</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition</title>
				<meeting>the IEEE Conference on Computer Vision and Pattern Recognition</meeting>
		<imprint>
			<date type="published" when="2015">2015</date>
			<biblScope unit="page" from="1" to="9" />
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
