<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Organizational and Legal Aspects of Managing the Process of Recognition of Objects in the Image</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Nataliya</forename><surname>Boyko</surname></persName>
							<email>nataliya.i.boyko@lpnu.ua</email>
							<affiliation key="aff0">
								<orgName type="institution">Lviv Polytechnic National University</orgName>
								<address>
									<postCode>Lviv79013</postCode>
									<country key="UA">Ukraine</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Lesia</forename><surname>Mochurad</surname></persName>
							<email>lesia.i.mochurad@lpnu.ua</email>
							<affiliation key="aff0">
								<orgName type="institution">Lviv Polytechnic National University</orgName>
								<address>
									<postCode>Lviv79013</postCode>
									<country key="UA">Ukraine</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Organizational and Legal Aspects of Managing the Process of Recognition of Objects in the Image</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">42C6E32A360921FF7DFFAEF379B5246C</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T13:09+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>object recognition</term>
					<term>ANN</term>
					<term>YOLO</term>
					<term>TensorFlow</term>
					<term>Keras</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>The issue of object recognition using ANN models is considered. The object of training is the YOLO approach for recognizing objects in an image. The subject of training is Keras, TensorFlow, and the ability to create and explore ANN models using them. The purpose of the work is to write a program that recognizes certain objects in the images and learn how it works according to the YOLO approach. The analysis of the YOLO approach of object recognition on images is carried out. An example of its use for recognizing objects on a student card: a barcode, a seal and a signature is given. The high-level Keras API and the TensorFlow library were used to build the ANN architecture to build and work with the computation graph. An analysis of LeNet-5, AlexNet, GoogleNet architectures was performed, while building ANN's own architecture and analysis of the YOLO approach for object recognition, a program for object recognition in an image was written using Python, TensorFlow, Keras, and TensorBoard for visualization of training and architecture artificial neural network. YOLO's approach to image recognition is explored. I have better studied TensorFlow, Keras for constructing and exploring ANN and TensorBoard models to visualize the training process of ANN, the graph of calculations. Gained practical skills in writing ANN, and their practical application. Has deepened his knowledge in the field of machine learning. The hardest part of the network was learning to recognize object sizes.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>There are many problems that have different arithmetic saturation. Some of them are easier to tell to computers: performing arithmetic operations, while others can be solved mentally by language recognition, image analysis, object classification and more. The advantage of solving tasks using computers is that they can perform known arithmetic and trigonometric operations sequentially and without fail. In addition to the usual processing of algebraic tasks, object recognition tasks are added. Known object recognition algorithms for an image usually have two parts: localization -de-termining the location of the object in the image, and classification -determining what it is for the object, ie to which class it belongs. This paper describes how to use a CNN convolutional neural network and the YOLO algorithm to identify individual objects in a static image. To analyze the task at hand, it is first scanned for computer vision.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">Setting the Task</head><p>The task is to explore the YOLO approach for recognizing objects in images and writing implementation using API TensorFlow and Keras in Python. The main task is to train the model from scratch, using the YOLO approach to identify individual objects, for example: a barcode, a seal and a signature on a student card. Achieve at least 80% average accuracy on test data. To solve it, you need to solve the following tasks:</p><p> explore the YOLO approach;  learn the basics of Keras, TensorFlow;  write your own implementation of the YOLO approach from scratch and the ANN model architecture;  create a dataset of different student card photos;  mark the dataset: barcode, seal and signature on the images;  write an algorithm for automatic generation of augmented data in YOLO format;  train the model to an accuracy of at least 80% on the test data.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">Methods of Solving</head><p>Conditionally, the recognition algorithm can be divided into 2 components: 1) Localizationdetermine the coordinates and dimensions of the object:</p><p>Input is an image, first there are important features of the image, then there is a function of dependence between the image features and the coordinates of the center, the height and width of the object. 2) Classificationthe definite on to which class an object be longs: Input is a localized object, first the features of the object are found, then there is a function of dependency between the features of the object and the class to which it belongs.</p><p>We chose the YOLO (you only look once) architecture because it combines 2 steps of recognitionlocalization and classification. Due to the fact that all recognition is performed by one net work, it can be optimized specifically for recognition efficiency: "A single neural network predicts bounding boxes and class probabilities directly from full image simonies valuation. Since the whole detection pipe line is a single network, it can be optimized end-to-end directly on detection performance" <ref type="bibr" target="#b7">[8]</ref><ref type="bibr" target="#b8">[9]</ref><ref type="bibr" target="#b9">[10]</ref><ref type="bibr" target="#b10">[11]</ref><ref type="bibr" target="#b11">[12]</ref><ref type="bibr" target="#b12">[13]</ref>.</p><p>This approach optimizes the speed of the algorithm, because the attributes of object classes are determined by one network with the attributes of its location and size: "Our unified architecture is extremely fast. Our base YOLO model processes images in real-time at 45 frames per second. A smaller version of the network, Fast YOLO, processes an astounding 155 frames per second while still achieving double the mAP of other real-time detectors" <ref type="bibr" target="#b0">[1]</ref><ref type="bibr" target="#b1">[2]</ref><ref type="bibr" target="#b2">[3]</ref><ref type="bibr" target="#b3">[4]</ref>.</p><p>"Finally, YOLO learns very general representations of objects. It outer -forms all other detection methods, including DPM and R-CNN, by a wide margin when generalizing from natural images to artwork on both the Picasso Dataset and the People-Art Dataset" <ref type="bibr" target="#b5">[6]</ref>.</p><p>We changed some parts of the algorithm to accomplish this task: since there is only one copy of each student card per class, there is no need to use non max suppression to determine one final prediction among others. It is enough to take 1 bbox for each class with the maximum level of confidence for that class.</p><p>Our task is easier than recognizing any objects -the objects are in static positions relative to each other, only the position of the student card changes, 1 image of the student card has only 1 copy of each class. So we decided to use a model with far fewer parameters than the YOLO model. In addition, it is my task to train the network from scratch, and training such a large architecture as the YOLO model is a very complex process that requires a lot of data, time and computing resources <ref type="bibr" target="#b9">[10]</ref><ref type="bibr" target="#b10">[11]</ref><ref type="bibr" target="#b11">[12]</ref><ref type="bibr" target="#b12">[13]</ref><ref type="bibr" target="#b13">[14]</ref>. Features are the images that you want to recognize. Used to calculate network output -location, size, and feature class predictions.</p><p>Filed in [N x W x H x channels], where: N is the number of images; W, H is the length and height of the image in pixels; channels -number of image channels (3 for RGB images) <ref type="bibr" target="#b17">[18]</ref><ref type="bibr" target="#b18">[19]</ref><ref type="bibr" target="#b19">[20]</ref><ref type="bibr" target="#b20">[21]</ref>.</p><p>Labels are the true coordinates of the object center, their size, and the class indexes to which they belong. They are used by the error, accuracy, and other functions that are intended to evaluate the operation of the algorithm by determining the difference between the correct data and those predicted by the algorithm.</p><p>According to the YOLO approach, the input image is divided by a grid into a grid of [S x S] cells, each cell having the same size, and is responsible for recognizing a specific area of the input image: Only one object in a class can recognize a cell -the probability of which among the C probabilities of all classes is the highest. Bouldering box is a vector of format numbers (x, y, w, h, confidence) where: (x, y) -the coordinates of the bbox center are calculated as the offset from the coordinates of the lower left corner of a particular cell of the image, so they take values between 0 and 1; (w, h) are the length and width of the bbox, normalized to the size of the input image, so they take values between 0 and 1. Confidence -the level of confidence that the bbox really recognized the true object.</p><p>The confidence value reflects how confidently you can say that a given bbox has an object, and how true the output of that bbox is. The goal of dream faience is to show how true the results of the prediction are, with no ready answers like in training mode. If there are no objects in this bbox, confidence = 0, if any, confidence = IoU(intersection over union) between this bbox (predicted) and true bbox. confidence = probability (obj) * iou (pred, label) -formal definition of confidence <ref type="bibr" target="#b3">[4]</ref><ref type="bibr" target="#b4">[5]</ref><ref type="bibr" target="#b5">[6]</ref>.</p><p>Labels are fed to the algorithm's input in format [N x Sw x Sh x (B*5+C)] , where: N -number of images; Sw -the number of cells in the width of the grid; Sh -the number of cells in the height of the grid; B -the number of bboxes per cell; C -number of object classes.</p><p>To identify the barcode, stamp and signature on the student card I used: Sw = Sh =10, B = 1, C = 3. Foresight: features. Calculations: features, labels.</p><p>Example of normalization of coordinates and size of bbox (Fig. <ref type="figure" target="#fig_3">4</ref>): Prediction: Output is the end product of the algorithm provided for each cell of an input image of an object's location and its class. The output dimension has the same format as the Labels input:</p><formula xml:id="formula_0">  ) 5 * ( C B h S w S N    <label>(1)</label></formula><p>Due to the fact that only one copy of each class is guaranteed to recognize objects on one student card, there is no need to use non max suppression. For each asset class, a bbox with a maximum confidence level greater than a certain confidence level of all bboxes in that class is selected. A class is defined as the index of the maximum value among the class predictions for a given grid cell, all the bboxes in the cell belong to that class (Fig. <ref type="figure">5</ref>). Calculation: Output is the value of certain metrics designed to evaluate the accuracy, object recognition error algorithm (IoU, max IoU, mean IoU, probability). Probability is a metric that combines the prediction of an object's coordinates with its class's predictions. This can be done by the following formula:</p><formula xml:id="formula_1">Propability= Pr (Class i | Object)* Pr (Object)* ruth tpred IoU = Pr (Class i )* ruth tpred IoU (2)</formula><p>Pr (Class i | Object) -the probability of a particular class for an object in bbox, provided that it has that object. Pr (Object) -the probability that a particular bbox has an object. This gives class-dependent confidence points for each bbox. Shows the full probability that an object of a particular class is in a particular bbox.</p><p>IoU is a feature that reflects the intersection of true and predicted bbox to merge. It is used to estimate the accuracy of the object algorithm localization. I use this feature because I need to evaluate the accuracy of object localization by an algorithm to understand how well it works. bounding box predictor to be responsible for each object. We assign one predictor to be "responsible" for predicting an object based on which prediction has the highest current IOU with the ground truth. This leads to specialization between the bounding box predictors. Each predictor gets better at predicting certain sizes, aspect ratios, or classes of object, improving overall recall". Example of calculating IoU: </p><p>Nthe number of images in the mini-package; truthtrue bbox; pred-prediction bbox. The YOLO error is a measure of how far network predictions differ from true values. The error function enables the network to learn. That is, it directly influences the learning process, thanks to which the optimization algorithm trains the network, changing its parameters in the direction of reducing the error. Without it, the network cannot learn to recognize objects, because without knowing what and how wrong the network will not be able to fix it.</p><p>Modified standard deviation between network prediction and true values is used to train the network. It is easy to optimize, but it is not well suited to the main goal of the network -maximize average accuracy because it equally evaluates localization and classification errors, which can be very different. Also, in each image, many cells are not responsible for object recognition. This causes the confidence equation for the bbox of this cell to 0, often over saturating the cell gradients responsible for recognition. It can lead to instability of the network model, and cause early training discrepancies, the model will find a poor local minimum.</p><p>To prevent this, the error for bbox coordinates is increased, and for cells with no objects the confidence error is reduced, 2 parameters are added for this: L coord = 5 and L noobj = 0.5.</p><p>The standard deviation also equates to errors in large and small bboxes. The error should reflect that small deviations in large bboxes are less significant than in small bboxes. For this, the square root is taken from the length and height of the bbox.</p><p>"We optimize for sum-squared error in the output of our model. We use sumsquared error because it is easy to optimize, however it does not perfectly align with our goal of maximizing average precision. It weights localization error equally with classification error which may not be ideal.</p><p>Also, in every image many grid cells do not contain any object. This pushes the "confidence" scores of those cells towards zero, often overpowering the gradient from cells that do contain objects. This can lead to model instability, causing training to diverge early on.</p><p>To remedy this, we increase the loss from bounding box coordinate predictions and decrease the loss from confidence predictions for boxes that don't contain objects. We use two parameters, λ coord and λ noobj to accomplish this. We setv λ coord =5 and λ noobj = 0.5.</p><p>Sum-squared error also equally weights errors in large boxes and small boxes. Our error metric should reflect that small deviations in large boxes matter less than in small boxes. To partially address this we predict the square root of the bounding box width and height instead of the width and height directly".</p><p>In YOLO, the error function consists of 4 parts, since the network output consists of several parts:  Coordinates of the bbox center.  Height and width bbox.  Confidence for bbox.  The likelihood that an object in a given cell belongs to a particular class if its center hits the cell. Consider the individual parts of the error and an example of how to calculate it for my task: Parameters for example (Fig. <ref type="figure" target="#fig_7">8</ref>). x . Its purpose is to locate the object on the image by a network, since it requires the coordinates of its center. It is calculated only for the bbox that are responsible for recognition (if their iou is maximal among other bboxes in this grid cell):"It also only penalizes bounding box coordinate error if that predictor is "responsible" for the ground truth box (i.e. has the highest IOU of any predictor in that grid cell)".</p><formula xml:id="formula_3">  2 ) ( 2 ) ( * * i y i y i x i x S i B j obj ij k coord l      (5) i i y</formula><p>x , -true coordinates of center bbox;</p><formula xml:id="formula_4">i i y x , ˆ</formula><p>-predicted coordinates of center bbox; obj ij k -coefficient equal to 1 or 0: l -when j -bbox is responsible for its recogni- tion, ie its value for IoU is maximal among other j -bboxes in this i -cell. In addition, the cell should be responsible for recognition (if it falls into the center of the object). 0 -otherwise. Its purpose is to locate the object on the image by the network, since it requires its length and width. Size and coordinate errors are only calculated for the bbox that are responsible for the recognition.</p><formula xml:id="formula_5">  2 ) ( 2 ) ( * * i h i h i w i w S i B j obj ij k coord l     <label>(6)</label></formula><p>) , ( The error should reflect smaller deviations in large bboxes less than in small ones, so it takes root from height and width: "Our error metric should reflect that small deviations in large boxes matter less than in small boxes". This is well illustrated in the example below:</p><p>Height of the first and second, width of the third object reduced by 0. "Each grid cell predicts B bounding boxes and confidence scores for those boxes. These confidence scores reflect how confident the model is that the box contains an object and also how accurate it thinks the box is that it predicts".</p><p>"If no object exists in that cell, the confidence scores should be zero. Otherwise we want the confidence score to equal the intersection over union (IoU) between the predicted box and the ground truth".</p><p>Therefore, the error reflects the difference between the real confidence value (iou) and the predicted network (c). The purpose of her finding is to understand how true are the results of the prediction of not having ready answers as in training. Their confidence for him should be 0. I think the coefficients for confidence errors are smaller than the coefficients for sizes and coordinates, since this error depend son the IOU value that changes dynamically in the learning process. So that the network does not learn when iou values are small to predict only small values of confidence, and when iou grows up sharply not to predict only high values of confidence the learning factories reduced. loss_confidence = ( 0*... + 0*(0.  The final error is the sum of four errors: 1) coordinate errors, 2) width and height, 3) confidence level, 4) classification. For the task -localization of bar code on the student card, classification is not required, so the classification error is not included in the general error: Loss = 0.01054 + 0.00171 + 0.088 + 0.00229 ~ 0.10254 Therefore, the final recognition error consists of the error of coordinate recognition, size recognition, prediction of how true this recognition is, and errors of object class definition. The purpose of the training is to increase the overall accuracy of object recognition by reducing the overall recognition error. output for the Labels algorithm by pre-converting them to YOLO (not changes). I used <ref type="bibr" target="#b15">[16]</ref> to mark the data: We trained a network with the following parameters: The total size of the training data sets 80 images. During training, I used data augmentationcertain changes to the input data in order to ensure that the algorithm works for noisy data, as well as to better study the patterns in the data. For example, in order for an algorithm to be able to recognize objects on a student card, if it sin the rotate deposition, pseudo random rotations in the training data are used. For better recognition of images made at an angle -slopes, images that are not in the center of the image -offsets, for images with different levels of brightness -fluctuations in brightness and hsv -transformation. After augmentation training set size -720 images After each epoch, the initial 80 images are augmented up to 720 by pseudorandom translating, rotating, scaling, shearing, hsv image transformations. For this I used api <ref type="bibr" target="#b15">[16]</ref>. Batch size -80 images, the training era consists of 9 mini-packages (9 step / epoch). The graph shows that at the initial stage of training, the error dropped rapidly, gradually its rate of decline decreased. In my opinion, this is due to the fact that certain features of the objects are easier to learn, others harder. It is also a result of the gradient damping phenomenon -in the first stages, the initial layers are rapidly training and the distant layers are slower because they have smaller gradients. Then the distant layers are slowly refinished, and the initial layers are adjusted, partly because the error is constantly fluctuating, but as a result it is reduced. The graph shows that the network quickly learned to recognize the coordinates of the bbox center. I think this is because all objects look like simple geometric shapes, and they are the same color. However, when resizing go rotating the image during augmentation, the center of the object does not change. The graph shows that the size recognition error decreases a there slowly and is unstable compared too there parts of the YOLO error. I think this is due to the fact that as a result of augmentation, the shape and size of the objects in the input data is constantly changing, so the network needs more time to study these patterns. The chart shows that the confidence level error from the outset was relative to other parts of the YOLO error. I believe that this is due to several factors -it depends only on one value (confidence), and the size error (w, h), coordinates (x, y) -two, and its part is multiplied by a factor of 0.5. Also, since all recognized objects are similar to simple geometric shapes in a network, it is easier to understand if an object is in a particular cell. The graph shows that the confidence level error decreases rapidly at the beginning of training. Further, the rate of decrease is slowed. I think this is due to the fact that the position of different classes of objects does not change relative to each other (left signature, then seal, then barcode) so it is easy to distinguish them by this feature. And slowing down the error reduction rate is because when other parts of the YOLO error are much larger, the optimization algorithm changes the network parameters more strongly to reduce them. The graph shows that the maximum IoU value is growing faster and more than the average IoU value. I think that because some objects area sire to localize than others, an object whose harpies very similar to a simple geometric shape (rectangle) is easier to localize than an object with a complex shapebecause any complex shape can be represented as a combination of simple forms. It takes a time for the network to learn to do this. The graph shows that at the beginning of training, the mean value of IoU increased rapidly, and gradually the rate of increase of iou decreased. In my opinion, this is due to the fact that the localization error initially decreased significantly, and the average IoU increased accordingly. Then the rate of decrease of the error gradually decreased, respectively, and the rate of increase of the average IoU decreased. It can be concluded that the localization error and the mean IoU are inversely proportional.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Fig. 1 .</head><label>1</label><figDesc>Fig. 1. ANN architecture First two digits -kernel size, third digit -amount of filters, s -strides (horizontal, vertical). Typically, three modes are used for ANN research:  Training -to increase the accuracy of object recognition.  Predictions -for object recognition.  Calculation -to calculate certain functions that evaluate the operation of the algorithm. This is the input to the algorithm for further processing. Input data vary depending on the mode of operation. Training: The training input has two components (Features, Labels).Features are the images that you want to recognize. Used to calculate network output -location, size, and feature class predictions.Filed in [N x W x H x channels], where: N is the number of images; W, H is the length and height of the image in pixels; channels -number of image channels (3 for RGB images)<ref type="bibr" target="#b17">[18]</ref><ref type="bibr" target="#b18">[19]</ref><ref type="bibr" target="#b19">[20]</ref><ref type="bibr" target="#b20">[21]</ref>.</figDesc><graphic coords="3,126.20,363.40,345.55,123.74" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Fig. 2 .</head><label>2</label><figDesc>Fig. 2. Divide the input image into cells Each cell consists of [B] bounding boxes, to locate the object and [C] the probabilities of the object belonging to a particular class Pr (Class i | Object) for its classification (C is the number of classes).Only one object in a class can recognize a cell -the probability of which among the C probabilities of all classes is the highest. Bouldering box is a vector of format numbers (x, y, w, h, confidence) where: (x, y) -the coordinates of the bbox center are calculated as the offset from the coordinates of the lower left corner of a particular cell of the image, so they take values between 0 and 1; (w, h) are the length and width of the bbox, normalized to the size of the input image, so they take values between 0 and 1.</figDesc><graphic coords="4,209.90,529.30,177.00,141.00" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>Fig. 3 .</head><label>3</label><figDesc>Fig. 3. Bbox one of the cells of the image (bbox border is red, cell grid is yellow)</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_3"><head>Fig. 4 .</head><label>4</label><figDesc>Fig. 4. Provided by the bbox object in the image (bbox borders are marked in red) Coordinates of two predicted bboxes for 1 object (barcode) in an image size 128x128px (img_w, img_h = 128px, S=4): [x, y, w, h]p1 = [47, 38, 66, 30] (px) [x, y, w, h]p2 = [51, 39, 60, 29] (px) Each cell size [cell_w x cell_h]: you should adjust the size of the image so that (img_w, img_h) is divided exactly by the number of cells (S) so that the cells cover the entire image. cell_w = img_w//Sweight of cell cell_h = img_h/Shigh of cell cell_w = 128/4 = 32(px) cell_h = 128/4 = 32(px) Coordinate normalization:</figDesc><graphic coords="5,194.90,315.40,207.00,184.10" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_4"><head></head><label></label><figDesc>of intersection to combining true bbox and predicted. Pr (Object)* ruth tpred IoU -calculated as the confidence of a particular bbox.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_5"><head>Fig. 5 . 5 Fig. 6 .</head><label>556</label><figDesc>Fig. 5. Choosing the best predictions From left to right: all predictions, bbox with confidence&gt; 0.5, max bbox with confidence&gt; 0.5</figDesc><graphic coords="7,242.57,372.12,123.00,74.25" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_6"><head>Fig. 7 .</head><label>7</label><figDesc>Fig. 7. Provided and true bbox facility Green rectangle -true, red -provided. IOU calculations for one of the predicted bbox and true bbox: [x, y, w, h]p1 = [47, 38, 66, 30](px) -The coordinates of a true rectangle [x, y, w, h]t = [48, 38, 67, 34](px) -The coordinates of the predicted rectangle x1 = x-w/2; y1 = y-h/2 -bottom left bbox point x2 = x+w/2; y2 = y+h/2upper right bbox point x1p1 = 47-66//2 = 14 y1p1 = 38-30//2 = 23 x2p1 = 47+66//2 = 80 y2p1 = 38+30//2 = 53 x1t = 48-67//2 = 15 y1t = 38-34//2 = 21 x2t = 48+67//2 = 81 y2t = 38+34//2 = 55 Intersection coordinates: (x1i , y1i); (x2i , y2i) -lower left and upper right point of intersection rectangle x1i = max(x1p1 , x1t) y1i = max(y1p1 , y1t) x2i = min(x2p1 , x2t) y1i = min(y2p1 , y2t) x1i , y1i = 15, 23 x2i , y2i = 80, 53 intersection area: intersection = max(0, (x2i -x1i ))*max(0, (y2i -y1i)) union area: union = Wp1 * hp1 + Wt* ht-intersection IoU = intersection/union</figDesc><graphic coords="8,210.77,147.40,186.60,180.72" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_7"><head>Fig. 8 .</head><label>8</label><figDesc>Fig. 8. Image for example of calculation of YOLO error green is indicated by true bbox, others include different classes Image height and width = 160, 120 px respectively, S = 10, 3 classes of objects, consider b = 2 to account for a case with multiple bboxes in one cell. Also, for simplicity, let's consider only those cells that should have an object. (values are rounded for visual clarity). True bboxes: [ 0.375, 0.53125, 0.09167, 0.26875, 0.1, 1, 0, 0] -bbox1 bar code [ 0.375, 0.53125, 0.09167, 0.26875, 0.935, 1, 0, 0] -bbox2 bar code [ 0.0833, 0.25, 0.15, 0.1125, 0.95, 0, 1, 0] -bbox1 seal [ 0.0833, 0.25, 0.15, 0.1125, 0.234, 0, 1, 0] -bbox2 seal [ 0.5, 0.9375, 0.167, 0.075, 0, 0, 0, 1] -bbox1 signature [ 0.5, 0.9375, 0.167, 0.075, 0.873, 0, 0, 1] -bbox2 signature Prediction bboxes: [ 0.363, 0.39, 0.409, 0.0414, 0.064, 1.01, 0.00491, 0.023] -bbox1 bar code [ 0.36, 0.552, 0.097, 0.267, 0.757, 1.01, 0.00491, 0.023] -bbox2 bar code [ 0.11, 0.258, 0.15, 0.11, 0.75, 0.0094, 0.994, 0] -bbox1 seal [ 0.18, 0.196, 0.089, 0.4, 0.248, 0.0094, 0.994, 0] -bbox2 seal [ 0.156, 0.136, 0.0657, -0.046, 0.0102, -0.0183, -0.0095, 1.033] -bbox1signature [ 0.513, 0.96, 0.176, 0.0814, 0.69, -0.0183, -0.0095, 1.033] -bbox2 signature The value of -0.046 issued by the network as the height of bbox1 for the signature is converted to 0 because the height cannot be negative. IoU: [0.1, 0.935, 0.95, 0.234, 0.0, 0.873] is the IoU value of true and predicted bbox for each object (used as true confidence). 1) Error for coordinates This is the value that characterizes the difference between the coordinates of the bbox center ) , ( i i y x predicted by the network, and true</figDesc><graphic coords="11,207.15,147.40,193.85,191.24" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_8"><head>= 5 - 2 )</head><label>52</label><figDesc>the magnitude of the increase in error for the coordinates and size Error for width and height This is the value that characterizes the difference between the size of a bbox, its width and the length provided by the network )</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_9"><head></head><label></label><figDesc>((0.091671/2 -0.4091/2)2 + (0.268751/2 -0.04141/2)2) + 1*((0.091671/2-0.0971/2)2 + (0.268751/2-0.2671/2)2) + 0*... + 1*((0.151/2 -0.151/2)2 + (0.11251/2 -0.111/2)2) + 0*((0.151/2-0.0891/2)2 + (0.11251/2-0.41/2)2) + 0*... + 0*((0.1671/2 -0.06571/2)2 + (0.0751/2 -01/2)2) + 1*((0.1671/2-0.1761/2)2 + (0.0751/2-0.08141/2)2) + 0*... ) ~ 0.00171</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_10"><head></head><label></label><figDesc>max(bbox) icells, because true confidence equals iou; 5 -trust coefficient for bbox not responsible for object recognition.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_11"><head></head><label></label><figDesc>the center of the object enters the i -cell. 0 -otherwise. "Note that the loss function only penalizes classification error if an object is present in that grid cell" loss_class= ( 0*... + (1 -1.01)^2 + (0 -0.00491)^2 + (0 -0.023)^2 + 0*... + (0 -0.0094)^2 + (1 -0.994)^2 + (0 -0)^2 + 0*... + (0 -(-0.0183))^2 + (0 -(-0.0095))^2 + (1 -1.033)^2 + 0*... ) ~ 0.00229 This is the value that characterizes the overall deviation of the present network performance from the desired results. It has the effect of further optimizing the network. The next step in the network operation is to calculate the dependence of the error function on the network parameters, then the parameters change in the direction of reducing the error.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_12"><head>Fig. 9 .</head><label>9</label><figDesc>Fig. 9. Example of image markup</figDesc><graphic coords="16,126.20,171.40,341.96,205.60" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_13"><head>Fig. 10 .</head><label>10</label><figDesc>Fig. 10. General error function</figDesc><graphic coords="17,126.20,147.40,341.89,144.85" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_14"><head>Fig. 11 .</head><label>11</label><figDesc>Fig. 11. Coordinate recognition error function</figDesc><graphic coords="17,126.20,400.60,341.97,143.15" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_15"><head>Fig. 12 .</head><label>12</label><figDesc>Fig. 12. Size recognition error function</figDesc><graphic coords="18,126.20,147.40,352.05,144.00" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_16"><head>Fig. 13 .</head><label>13</label><figDesc>Fig. 13. Trust recognition error function</figDesc><graphic coords="18,126.20,363.75,341.22,141.25" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_17"><head>Fig. 14 .</head><label>14</label><figDesc>Fig. 14. Class recognition error function</figDesc><graphic coords="19,126.20,147.40,341.15,140.60" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_18"><head>Fig. 15 .</head><label>15</label><figDesc>Fig. 15. Max value of IoU</figDesc><graphic coords="19,126.20,396.35,341.15,143.25" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_19"><head>Fig. 16 .</head><label>16</label><figDesc>Fig. 16. Average value of IoU</figDesc><graphic coords="20,126.20,147.40,344.91,142.15" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0"><head></head><label></label><figDesc></figDesc><graphic coords="1,0.00,190.95,595.32,460.02" type="bitmap" /></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="4" xml:id="foot_0">ExperimentsTo mark the data, we applied bbox boundary markers as rectangles of different colors for each class, and saved the coordinates, sizes, classes of these rectangles as true</note>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">Results</head><p>We managed to reach mean IoU of ~ 85% on test data over 400 training periods. At the end of each training period, the values and metrics described above were calculated and stored to represent the state of the network and the training process as a whole. For graphing I used <ref type="bibr" target="#b13">[14]</ref>. Results for 400 training periods:</p><p>The x-axis is the index of the era, the y-axis is the value of the function</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6">Conclusion</head><p>We achieved the task (Achieve at least 80% average accuracy on the test data) using the algorithm describe dab vein 4h 10min using the Intel (R) Core (TM) i5-8300H CPU @ 2.30GHz 2.30 GHz to train the network. With the YOLO approach, you can train you net work to recognize objects using a relatively small amount of computing power within a reasonable time. The speed of training depends strongly on the objects of recognition and image parameters: the complexity of the shape of the object, the possibility of the presence of several objects of the same class in the image, the number of classes of objects, or easily distinguish them, camera angle, change of illumination, distance to the object , placing the object in the image. After all, training a network for object recognition under different conditions requires a larger data set to train and test, to train the network on a larger set takes longer. Also, a more complex ANN model may be required to analyze more complex data. Therefore, such details should be determined in the first phase of building the object recognition system. In order to evaluate the network effectively, it is necessary to select the appropriate metrics (in this case IoU, yolo_loss). Selecting an error function by minimizing which will maximize recognition accuracy. Combining localization and classification steps enables the recognition of 1-step ANN calculations and simplifies the design of algorithm input. At the same time, the recognition process is optimized, since the same features of the object studied by the same network are used for localization and classification.</p><p>Research materials: labeled data set, trained model, references to useful resources -can be used for future research, and research in general -an example for future generations, which they can refine, refine. The trained model can be used to recognize barcodes, prints, signatures on an image, and then transfer them to other data processing algorithms. For example, a barcode can be transmitted to a barcode reader to obtain student information, and a stamp and signature can be transmitted to document validation algorithms.</p><p>Thanks to the development of machine learning, we have received real-time object recognition tools with fairly high precision. This facilitates process automation, as work related to the analysis of visual images by humans can be partially translated to a computer. Further research is promising, as it can increase the amount of computertranslated work and thus free people from monotonous work. To do this, the accuracy and speed of the algorithms must be at least as human as possible.</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">P2P Spatial query processing by Delaunay triangulation</title>
		<author>
			<persName><forename type="first">H.-Y</forename><surname>Kang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B.-J</forename><surname>Lim</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K.-J</forename><surname>Li</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="s">Lecture notes in computer science</title>
		<imprint>
			<biblScope unit="volume">3428</biblScope>
			<biblScope unit="page" from="136" to="150" />
			<date type="published" when="2005">2005</date>
			<publisher>Springer</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Density connected clustering with local subspace preferences</title>
		<author>
			<persName><forename type="first">C</forename><surname>Boehm</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Kailing</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Kriegel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Kroeger</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. of the 4th IEEE Intern. conf. on data mining</title>
				<meeting>of the 4th IEEE Intern. conf. on data mining<address><addrLine>Los Alamitos</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2004">2004</date>
			<biblScope unit="page" from="27" to="34" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Heterogeneous spatial data mining based on grid</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Wu</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="s">Lecture notes in computer science</title>
		<imprint>
			<biblScope unit="volume">4683</biblScope>
			<biblScope unit="page" from="503" to="510" />
			<date type="published" when="2007">2007</date>
			<publisher>Springer</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Synthesis control system physiological state of a soldier on the battlefield</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Kryvenchuk</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Boyko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Helzynskyy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Helzhynska</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Danel</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">CEUR</title>
		<imprint>
			<biblScope unit="volume">2488</biblScope>
			<biblScope unit="page" from="297" to="306" />
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Clustering spatial data using random walks</title>
		<author>
			<persName><forename type="first">D</forename><surname>Harel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Koren</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. of the 7th ACM SIGKDD Intern. conf. on knowledge discovery and data mining</title>
				<meeting>of the 7th ACM SIGKDD Intern. conf. on knowledge discovery and data mining<address><addrLine>San Francisco, California</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2000">2000</date>
			<biblScope unit="page" from="281" to="286" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">On the application of inductive machine learning tools to geographical analysis</title>
		<author>
			<persName><forename type="first">M</forename><surname>Gahegan</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Geographical Analysis</title>
		<imprint>
			<biblScope unit="volume">32</biblScope>
			<biblScope unit="page" from="113" to="139" />
			<date type="published" when="2000">2000</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Information System of Catering Selection by Using Clustering Analysis</title>
		<author>
			<persName><forename type="first">N</forename><surname>Boyko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Kh</forename><surname>Shakhovska</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Mochurad</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Campos</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 1st International Workshop on Digital Content &amp; Smart Multimedia (DCSMart 2019)</title>
				<meeting>the 1st International Workshop on Digital Content &amp; Smart Multimedia (DCSMart 2019)<address><addrLine>Lviv, Ukraine</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2019">2019</date>
			<biblScope unit="page" from="94" to="106" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Clustering Algorithms for Economic and Psychological Analysis of Human Behavior</title>
		<author>
			<persName><forename type="first">N</forename><surname>Boyko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Komarnytska</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Yu</forename><surname>Kryvenchuk</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Yu</forename><surname>Malynovskyy</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the International Workshop on Conflict Management in Global Information Networks (CMiGIN 2019)</title>
				<meeting>the International Workshop on Conflict Management in Global Information Networks (CMiGIN 2019)<address><addrLine>Lviv, Ukraine</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2019">2019</date>
			<biblScope unit="page" from="614" to="626" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Research of servers and protocols as means of accumulation, processing and operational transmission of measured information</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Kryvenchuk</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Vovk</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Chushak-Holoborodko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Khavalko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Danel</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Advances in Intelligent Systems and Computing</title>
		<imprint>
			<biblScope unit="volume">1080</biblScope>
			<biblScope unit="page" from="920" to="934" />
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Testing local spatial autocorrelation using</title>
		<author>
			<persName><forename type="first">С</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Murayama</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Intern. J. of Geogr. Inform. Science</title>
		<imprint>
			<biblScope unit="volume">14</biblScope>
			<biblScope unit="page" from="681" to="692" />
			<date type="published" when="2000">2000</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Amoeba: Hierarchical clustering based on spatial proximity using Delaunay diagram</title>
		<author>
			<persName><forename type="first">V</forename><surname>Estivill-Castro</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Lee</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">9th Intern. Symp. on spatial data handling</title>
				<meeting><address><addrLine>Beijing, China</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2000">2000</date>
			<biblScope unit="page" from="26" to="41" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Testing spacetime and more complex hyperspace geographical analysis tools</title>
		<author>
			<persName><forename type="first">I</forename><surname>Turton</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Openshaw</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Brunsdon</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Innovations in GIS</title>
		<imprint>
			<biblScope unit="volume">7</biblScope>
			<biblScope unit="page" from="87" to="100" />
			<date type="published" when="2000">2000</date>
			<publisher>Taylor &amp; Francis</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Efficiency of Using Utility for Username Verification in Online Community Management</title>
		<author>
			<persName><forename type="first">S</forename><surname>Fedushko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Yu</forename><surname>Syerov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Skybinskyi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Shakhovska</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Kunch</surname></persName>
		</author>
		<idno>CEUR-WS.org</idno>
		<ptr target="http://ceur-ws.org/Vol-2588/paper22.pdf" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the International Workshop on Conflict Management in Global Information Networks (CMiGIN 2019)</title>
				<meeting>the International Workshop on Conflict Management in Global Information Networks (CMiGIN 2019)<address><addrLine>Lviv, Ukraine</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2019">2020. November 29, 2019</date>
			<biblScope unit="volume">2588</biblScope>
			<biblScope unit="page" from="265" to="275" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">Spatial clustering in the presence of obstacles</title>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">K</forename><surname>Tung</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Hou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Han</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">The 17th Intern. conf. on data engineering (ICDE&apos;01)</title>
				<meeting><address><addrLine>Heidelberg</addrLine></address></meeting>
		<imprint>
			<biblScope unit="page" from="359" to="367" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">Elements of the formal model big date, The</title>
		<author>
			<persName><forename type="first">O</forename><surname>Veres</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Shakhovska</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">11th Intern. conf. Perspective Technologies and Methods in MEMS Design (MEMSTEH)</title>
				<meeting><address><addrLine>Polyana</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2015">2015</date>
			<biblScope unit="page" from="81" to="83" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">Automatic sub-space clustering of high dimensional data</title>
		<author>
			<persName><forename type="first">R</forename><surname>Agrawal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Gehrke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Gunopulos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Raghavan</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Data mining knowledge discovery</title>
		<imprint>
			<biblScope unit="volume">11</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page" from="5" to="33" />
			<date type="published" when="2005">2005</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">Big data processing technologies in distributed information systems</title>
		<author>
			<persName><forename type="first">N</forename><surname>Shakhovska</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Boyko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Zasoba</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Benova</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Procedia Computer Science, 10th International conference on emerging ubiquitous systems and pervasive networks (EUSPN-2019), 9th International conference on current and future trends of information and communication technologies in healthcare (ICTH-2019)</title>
				<meeting>edia Computer Science, 10th International conference on emerging ubiquitous systems and pervasive networks (EUSPN-2019), 9th International conference on current and future trends of information and communication technologies in healthcare (ICTH-2019)<address><addrLine>Lviv, Ukraine</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2019">2019</date>
			<biblScope unit="volume">160</biblScope>
			<biblScope unit="page" from="561" to="566" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">Distance based subspace clustering with flexible dimension partitioning</title>
		<author>
			<persName><forename type="first">L</forename><surname>Guimei</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Jinyan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Sim</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Limsoon</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. of the IEEE 23rd Intern. conf. on digital object identifier</title>
				<meeting>of the IEEE 23rd Intern. conf. on digital object identifier</meeting>
		<imprint>
			<date type="published" when="2007">2007</date>
			<biblScope unit="volume">15</biblScope>
			<biblScope unit="page" from="1250" to="1254" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<analytic>
		<title level="a" type="main">Finding generalized projected clusters in high dimensional spaces</title>
		<author>
			<persName><forename type="first">C</forename><surname>Aggarwal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Yu</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">ACM SIGMOD Intern. conf. on management of data</title>
				<imprint>
			<date type="published" when="2000">2000</date>
			<biblScope unit="page" from="70" to="81" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<analytic>
		<title level="a" type="main">ICEAGE: Interactive clustering and exploration of large and high-dimensional geodata</title>
		<author>
			<persName><forename type="first">D</forename><surname>Guo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">J</forename><surname>Peuquet</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Gahegan</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Geoinfor-matica</title>
		<imprint>
			<biblScope unit="volume">3</biblScope>
			<biblScope unit="issue">7</biblScope>
			<biblScope unit="page" from="229" to="253" />
			<date type="published" when="2003">2003</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<analytic>
		<title level="a" type="main">Comparison Of Machine Learning Libraries Performance Used For Machine Translation Based On Recurrent Neural Networks</title>
		<author>
			<persName><forename type="first">N</forename><surname>Boyko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Basystiuk</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">IEEE Ukraine Student, Young Professional and Women in Engineering Congress (UKRSYW)</title>
				<meeting><address><addrLine>Kyiv, Ukraine</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2018">2018. 2018</date>
			<biblScope unit="page" from="78" to="82" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b21">
	<analytic>
		<title level="a" type="main">A Monte Carlo algorithm for fast projective clustering</title>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">M</forename><surname>Procopiuc</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Jones</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><forename type="middle">K</forename><surname>Agarwal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">M</forename><surname>Murali</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">ACM SIGMOD Intern. conf. on management of data</title>
				<meeting><address><addrLine>Madison, Wisconsin, USA</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2002">2002</date>
			<biblScope unit="page" from="418" to="427" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b22">
	<analytic>
		<title level="a" type="main">Towards an effective cooperation of the user and the computer for classification</title>
		<author>
			<persName><forename type="first">M</forename><surname>Ankerst</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Ester</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H.-P</forename><surname>Kriegel</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. of the 6th ACM SIGKDD Intern. conf. on knowledge discovery and data mining</title>
				<meeting>of the 6th ACM SIGKDD Intern. conf. on knowledge discovery and data mining<address><addrLine>Boston, Massachusetts, USA</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2000">2000</date>
			<biblScope unit="page" from="179" to="188" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b23">
	<analytic>
		<title level="a" type="main">Automated document analysis for quick personal health record creation. 2nd International Workshop on Informatics and Data-Driven Medicine</title>
		<author>
			<persName><forename type="first">N</forename><surname>Boyko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Pylypiv</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Peleshchak</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Kryvenchuk</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Campos</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IDDM</title>
		<imprint>
			<biblScope unit="page" from="208" to="221" />
			<date type="published" when="2019">2019. 2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b24">
	<analytic>
		<title level="a" type="main">Analysis of the architecture of distributed systems for the reduction of loading high-load networks</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Kryvenchuk</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Mykalov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Novytskyi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Zakharchuk</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Malynovskyy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Řepka</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Advances in Intelligent Systems and Computing</title>
		<imprint>
			<biblScope unit="page" from="759" to="550" />
			<date type="published" when="1080">1080. 2020</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
