A method of iterative image normalization for tasks of visual navigation of UAVs M O Elantcev1, I O Arkhipov1 and R M Gafarov1 1 Kalashnikov Izhevsk State Technical University, Russian Federation, Izhevsk, Studencheskaya 7, 426069 e-mail: emo-name@hotmail.com, aio1024@mail.ru Abstract. The work deals with a method of eliminating the perspective distortion of an image acquired from an unmanned aerial vehicle (UAV) camera in order to transform it to match the parameters of the satellite image. The normalization is performed in one of the two ways. The first variant consists in the calculation of an image transformation matrix based on the camera position and orientation. The second variant is based on matching the current frame with the previous one. The matching results in the shift, rotation, and scale parameters that are used to obtain an initial set of pairs of corresponding keypoints. From this set four pairs are selected to calculate the perspective transformation matrix. This matrix is in turn used to obtain a new set of pairs of corresponding keypoints. The process is repeated while the number of the pairs in the new set exceeds the number in the current one. The accumulated transformation matrix is then multiplied by the transformation matrix obtained during the normalization of the previous frame. The final part presents the results of the method that show that the proposed method can improve the accuracy of the visual navigation system at low computational costs. 1. Introduction This paper is concerned with the development of a method for image normalization that is used in a visual navigation system of unmanned aerial vehicles (UAVs). The navigation system being developed [1] matches images acquired from an UAV onboard camera with the satellite image of the flying area. The matching yields the shift, rotation, and scale values that can be used to transform the input image to the corresponding part of satellite image. Knowing the coordinates of the satellite image and the shift, rotation, and scale parameters of the current input image, it is possible to determine the coordinates of the UAV. During the flight the tilt of the UAV is unstable, it changes because of weather conditions or when the UAV is making a pivot. It results in perspective distortions of images acquired from the onboard camera, even if a stabilized platform used. The distortions interfere with the correct matching of the images from the camera with the satellite image, which is usually represented in an orthogonal projection. To increase the likelihood of correct matching a method of image normalization for eliminating the perspective distortions is needed. The method should be computationally simple because it is supposed to run on low-power U AV processors. There are several approaches to solve this problem. The first approach is to calculate the matrix of perspective transformation based on the external parameters of the camera obtained from the UAV V International Conference on "Information Technology and Nanotechnology" (ITNT-2019) Image Processing and Earth Remote Sensing M O Elantcev, I O Arkhipov and R M Gafarov sensors, namely pitch, yaw, roll, and altitude. In [2], the distortion is compensated by determining the orientation of a virtual camera with its optical axis collinear to the normal vector of the shooting plane. This virtual camera provides us with an image without perspective distortions. Therefore, to eliminate the distortions of the input image we need to calculate the homography matrix between the images from the real and the virtual cameras. There are two difficulties with using this approach: the moments of shooting and of external parameters recording should be strictly synchronized, and the surface of the flying area should be flat. The second approach is to find and analyze the location of the known structural elements in the input image. This approach assumes that each frame from the camera is processed independently. The paper [3] focuses on the distortion elimination for the task of bar code recognition. Its idea is to find four straight lines bounding the quadrilateral barcode area. These lines provide four corner points that are used to calculate the transformation matrix. This approach is difficult to apply to the task of normalizing the UAV images because the shapes of the objects in the images are usually unknown. The third approach is to find the pairs of corresponding elements in the two images and use their positions to determine the transformation matrix of one image into another. Usually the methods search for the pairs of image points with characteristic local features known as keypoints. Depending on the task, different methods of keypoints extraction and matching can be used [4,5]. Under ideal conditions, four pairs of the keypoints are enough to determine the transformation matrix. However, in practice even the positions of the correctly matched points can be different because of noise. To obtain a more accurate solution a system of equations based on all the matched pairs is constructed. This system is usually solved by the least squares method so as to minimize the given error function (the DLT method [6]). In [7], it is proposed to perform this method iteratively. This method helps to increase the likelihood of image matching by means of eliminating the noise points. The set of the pairs of corresponding points can also contain mismatched pairs. In this case the DLT method can yield a wrong result and one of the robust methods should be used, such as RANSAC [8, 9]. In this method four pairs of corresponding points are randomly selected among all the matched pairs. Using them the homography matrix is calculated, and then the algorithm checks how many pairs fit this transform (inliers). The process is repeated, and other random pairs are selected from the set of inliers until a solution is found that satisfies the specified number of pairs or the maximum number of iterations is exceeded. Another similar robust method is LMS. It estimates four current selected pairs of points not by the number of inliers but by the median distance among all the pairs of points [6]. The described robust methods require a large number of iterations; therefore, they are not effective on low-power processors. The RHO method [10] aims to solve this problem. Instead of randomly electing four pairs of points at each iteration, the method preliminarily evaluates each pair, selects the most promising ones and then corrects the estimation. This allows to obtain the result in significantly fewer iterations. The fourth approach is to analyze the input image using a neural network, as it is demonstrated in [11]. It is noted that the advantage of this approach is the ability to work when there is a very small number of the keypoints or when the input images are blurred. However, this method is computationally complex and requires special hardware to work in real time, which makes it difficult to apply to the task of UAV visual navigation. Normalization based on the input image, Is the external parameters of the camera previous frame normalized successfully image previous frame matched? Normalization based on data keypoints matching Figure 1. Simplified scheme of the proposed method. V International Conference on "Information Technology and Nanotechnology" (ITNT-2019) 145 Image Processing and Earth Remote Sensing M O Elantcev, I O Arkhipov and R M Gafarov The proposed method of iterative image normalization for tasks of visual navigation of UAVs combines the first and the third approaches. A simplified scheme of the method is shown in Figure 1. Normalization based on the external parameters of the camera (pitch, yaw, roll, and flight altitude) is used either when processing the initial camera frame or in the case where the previous frame has not been successfully matched with the satellite image or the current frame cannot be matched with the previous one. In other cases the iterative method of homography matrix calculation based on matching the keypoints of the previous and the current frames is used. 2. Description of the peculiarities of the task considered The UAV flies along the given trajectory that contains both straight sections and turns. The considered flight altitude is 250-500m. There is a camera onboard the UAV that is directed strictly downwards. There is no stabilized platform on the UAV. The shooting speed of the camera is enough to provide at least 80% of overlap between the current and the previous frames. The parameters of focal length and the dimensions of the camera's sensor are known. The UAV is equipped with an accelerometer and a pressure sensor that provide information about the UAV orientation (roll, pitch, and yaw) and the flight altitude. However, according to experiments on real images, the moments of shooting and of parameters recording are not exactly the same. Therefore, strong wind gusts cause inaccuracies in the recorded external parameters of the camera. In addition, the pressure sensor estimates the flight altitude only above sea level. When calculating the flight altitude above ground level the height map of the flying area should be taken into account. The result of the normalization is the image transformed into the orthogonal projection so that later it can be matched with the satellite image. 3. Normalization based on the external parameters of the camera The first algorithm of the proposed normalization method is based on the information on the camera orientation and position in space, namely the roll ρ, pitch Ο„, yaw Ξ³, and flight altitude hf. In addition, the input information includes the terrain height ha at the shooting point and the reference altitude hr at which the shooting scale equals the satellite image scale. The output information is the coefficients of the transformation matrix that eliminates the distortion. The algorithm consists of two stages: the calculation of the transformation coefficients disregarding the scale and the calculation of the scaling factor. 3.1. Calculation of the transformation coefficients disregarding the scale The camera’s field of view can be represented as a pyramid that intersects the earth's surface plane. To simplify the calculation, instead of rotating the pyramid, according to the camera orientation the earth's surface plane is rotated and placed at one of the base vertices of the pyramid, as Figure 2a shows. a) b) Figure 2. The pyramid of vision: (a) a view at an angle, MNK is the section with the earth's surface plane, (b) the top view. The purpose of the first stage is to calculate the coordinates of the points M, N, K on the top view projection (Figure 2b). These coordinates can be used to calculate the resultant homography matrix Ta by solving an equation system constructed with the coordinates transformation equations of the pairs V International Conference on "Information Technology and Nanotechnology" (ITNT-2019) 146 Image Processing and Earth Remote Sensing M O Elantcev, I O Arkhipov and R M Gafarov of points: A-A, B-K, C-N, D-M. The transformation of each pair of points can be described by two equations [12]: 𝑑 π‘₯+𝑑 𝑦+𝑑 𝑑 π‘₯+𝑑 𝑦+𝑑 π‘₯ β€² = 𝑑11 π‘₯+𝑑12 𝑦+113 𝑦 β€² = 𝑑21 π‘₯+𝑑22 𝑦+123 , (1) 31 32 31 32 where x and y are the coordinates of the point before the transformation, x’ and y’ are the coordinates of the point after the transformation, t11, t12, t13, t21, t22, t23, t31 ΠΈ t32 are the coefficients of the transformation matrix Ta. The shape of the pyramid depends on the size of the photo sensor and the focal length. The viewing angles are calculated as follows: π‘š π‘š 𝛼x = 2 arctg w 𝛼y = 2 arctg h , (2) 2𝑓L 2𝑓L where mw and mh are the dimensions of the photo sensor, fL is the focal length. The vertex angles of the pyramid are calculated as follows: π‘š 𝛼 π‘š 𝛼 𝛽x = w sin 𝑦 𝛽y = h sin π‘₯ (3) π‘šh 2 π‘šw 2 The length of the edge of the pyramid le and the diagonal of the base of the pyramid ld are calculated as follows: π‘šh 2 + π‘š2 , 𝑙e = 𝑙d = οΏ½π‘šw h (4) 2sin (𝛽y /2) Depending on the orientation of the camera, the earth's surface plane is placed at one of the four base vertices, as shown in Figure 3. a) b) c) d) Figure 3. The location of the pyramid section with the earth's surface plane depending on the orientation of the camera: (a) Ο„ = 0, ρ <0, (b) Ο„ = 0, ρ> 0, (c) Ο„> 0, ρ = 0, (d) Ο„ <0, ρ = 0. The vertices are renamed according to the following rule: a vertex that coincides with a section corner is called A’, then in anticlockwise direction: B’, C ’, D’. The coordinates of the vertex K can be calculated with the following formula: sin 𝜌 𝑑k = π‘šπ‘€ /2βˆ’πœŒ) cos(𝛽x π‘š π‘šh K x = οΏ½Dβ€²x βˆ’ 𝑑k 2𝑙w οΏ½ K y = οΏ½Dβ€²y βˆ’ 𝑑k 2𝑙e οΏ½ (5) e The coordinates of the vertex M can be calculated with the following formula: sin 𝜏 𝑑m = π‘šβ„Ž /2βˆ’πœ) cos(𝛽y π‘š π‘šh Mx = οΏ½Bxβ€² βˆ’ 𝑑k 2𝑙w οΏ½ My = οΏ½Byβ€² βˆ’ 𝑑k 2𝑙e οΏ½ (6) e The coordinates of the vertex N can be calculated with the following formula: y πœ‹βˆ’π›½ y πœ‹βˆ’π›½ πœ‹βˆ’π›½y 𝑑nk = π‘šh βˆ’ 2𝑑k cos βˆ’ sin / sin(πœ‹ βˆ’ 𝜏 βˆ’ ) 2 2 2 𝑑�k = 𝑑e βˆ’ 𝑑k 𝑑nοΏ½b = βˆ’2𝑑�k cos 𝛽y (7) 2 2 2 𝑑nf = (�𝑑𝑛𝑓𝑏 βˆ’ 4�𝑑�k βˆ’ 𝑑nk οΏ½ βˆ’ 𝑑nοΏ½b )/2 Cβ€² βˆ’Aβ€² π‘š Cβ€² βˆ’Aβ€²y π‘š Nx = οΏ½ x 2 x + 𝑑nf 2𝑙w οΏ½ Ny = οΏ½ y + 𝑑nf 2𝑙hοΏ½ e 2 e To align the image with the northward direction, matrix multiplication can be used: cos 𝛾 βˆ’sin 𝛾 0 𝑇𝑛 = π‘‡π‘Ž οΏ½ sin 𝛾 cos 𝛾 0οΏ½ (8) 0 0 1 V International Conference on "Information Technology and Nanotechnology" (ITNT-2019) 147 Image Processing and Earth Remote Sensing M O Elantcev, I O Arkhipov and R M Gafarov 3.2. Calculation of the scale factor Since the earth's surface plane was placed at one of the base vertices of the pyramid, the obtained transformation matrix disregards the original scale of the image. To calculate the scaling factor the following formula can be used: 𝑑nd = οΏ½(Nx βˆ’ Aβ€²x )2 + (Ny βˆ’ Aβ€²y )2 𝑙 𝑑 𝑆𝑛 = 𝑑e nd 𝑙 (9) nf d To transform the image acquired from the UAV camera to the scale of the satellite image, the transformation matrix should be adjusted using the following scale factor: β„Ž βˆ’β„Ž π‘†π‘Ÿ = 𝑆𝑛 β„Žrβˆ’β„Ž a (10) f a 3.3. Obtaining the result of the normalization To obtain the final transformation matrix the following formula should be used: π‘†π‘Ÿ 0 0 π‘‡π‘Ÿ = 𝑇𝑛 οΏ½ 0 π‘†π‘Ÿ 0οΏ½ (11) 0 0 1 Figure 4 shows the result of the method of image normalization based on the external parameters of the camera. a) b) c) Figure 4. The result of image normalization based on the external parameters of the camera: (a) the original image, (b) the normalized image, (c) the corresponding part of the satellite image. 4. Normalization based on keypoints matching The first algorithm of the proposed method eliminates the distortions of the image coarsely, because, as it is mentioned above, the moment of recording the external parameters of the camera not necessarily coincides with the moment of shooting. Moreover, it is rather difficult to accurately take into account the elevation difference of the underlying surface. Therefore, it is used only when the previous frame from the camera cannot be matched with the satellite image. In other cases the second algorithm is used that is based on matching the current frame with the previous frame already normalized. 4.1. Matching the current frame with the previous one To reduce its complexity, the image matching algorithm analyzes not all the pixels but only the keypoints extracted with the method described in [13]. After the keypoint sets of the both images are found, they are used to determine the shift, rotation, and scale values between the two images using the statistical differentiation method [14]. These parameters form the following transformation matrix [15]: 𝑠 cos πœ™ βˆ’π‘  sin πœ™ 𝑑x 𝐻m = οΏ½ 𝑠 sin πœ™ 𝑠 cos πœ™ 𝑑y οΏ½, (12) 0 0 1 where tx and ty are the shift values along the X axis and the Y axis respectively, Ο† is the rotation value, s is the scale value. 4.2. Selecting pairs of corresponding points After the transformation matrix is found, it is used to determine the pairs of corresponding points. Some point A of the current frame corresponds to some point B of the previous frame when after V International Conference on "Information Technology and Nanotechnology" (ITNT-2019) 148 Image Processing and Earth Remote Sensing M O Elantcev, I O Arkhipov and R M Gafarov applying the transformation matrix to the coordinates of the point A the distance between A and B does not exceed the specified threshold (1-2 pixels). From the whole set of the pairs of corresponding points four pairs are selected that will be used to calculate the homography matrix (1). The choice is limited by the condition that no three points are collinear. The best case occurs when a quadrilateral of maximum area with as large angles as possible is found. However, the search for such a set of points is a computationally complex operation. The following approaches to selecting the points were tried: 1) the four points nearest to the corners of the image; 2) the four points most distant from the center of mass of the keypoints set and located in different quadrants relative to the center of mass; 3) the first point is the most distant from the center of mass, the second point is the most distant from point 1, the third is the point with the greatest sum of its distances from points 1 and 2, the fourth is the point with the greatest sum of its distances from points 1, 2 and 3; 4) a random selection of points. If this algorithm finds less than four corresponding pairs, the normalization based on the external parameters of the camera is applied. 4.3. Calculation of the transformation matrix After the four pairs of the corresponding keypoints are selected, the refined transformation matrix is calculated. This homography matrix matches the projection of the current frame with the projection of the previous frame already normalized. Then the search for the pairs of corresponding points is repeated using the current frame warped according to the refined transformation matrix. From the found pairs a new set of four pairs is selected and a new refined matrix is calculated. The process is repeated as long as the number of the pairs of corresponding points increases. Experiments show that in most cases two iterations are enough to obtain the solution. The final matrix of the transformation of the original image into the normalized one is determined by the formula: 𝑇r = 𝑇𝑝 βˆπ‘›π‘–=1 𝐻𝑖 (13) where Tp is the perspective transformation matrix for normalizing the previous frame, Hi are the transformation matrices obtained at each iteration, n is the number of iterations. Since the previous frame is used to normalize the current frame, the normalization error can accumulate. However, this method is used as a part of the visual navigation system that implies that the next step is to match the normalized image with the satellite image. This matching eliminates the accumulated error. 5. Experiments To test the proposed method, it has been integrated into the preprocessing stage of the UAV visual navigation system that is being developed. Then a series of experiments using real images of a UAV flight were performed. The trajectory of the UAV flight is shown in Figure 5 and consists of sections of two types: straight sections (270 frames) and turns (70 frames). Figure 5. The trajectory used in the experiments. The positions of the UAV at the straight sections are marked with green dots. The positions of the UAV at the turns are marked with yellow dots. The black lines show the boundaries between the sections. The white lines connect the coordinates of the UAV to the corresponding centers of photographing. V International Conference on "Information Technology and Nanotechnology" (ITNT-2019) 149 Image Processing and Earth Remote Sensing M O Elantcev, I O Arkhipov and R M Gafarov The first experiment aims to choose the best algorithm for image normalization based on keypoints matching. The proposed iterative algorithm was tested (with the four variants of selecting the four pairs of points described above), as well as the methods implemented in the OpenCV 3.4.5 library: DLT, RANSCAC, LMS and RHO. To determine the speed of the methods, each input frame was processed 100 times and the average execution time was calculated. Processing was performed on an AMD FX-8320E 3.2 GHz processor without using a graphics accelerator. Table 1. Comparison of the results of various methods of the homography matrix calculation. Proposed method Indicator Section variant 1 variant 2 variant 3 variant 4 DLT RAN LMS RHO of pair of pair of pair of pair SAC selection selection selection selection Average number of straight sections 156.8 pairs of turns 123.2 corresponding points before processing whole trajectory 150.0 Average number of straight sections 173.5 171.5 172.1 158.9 175.0 177.5 178.3 178.2 pairs of turns 136 7 133.6 135.0 125.6 138.7 141.3 142.6 142.0 corresponding points after processing whole trajectory 165.8 163.7 164.5 152 167.5 170 170.9 170.7 Average number of straight sections 1.9 1.8 1.8 1.2 - iterations of the turns 1.8 1.7 1.7 1.2 - proposed algorithm whole trajectory 1.9 1.8 1.8 1.2 - Average processing straight sections 0.06 0.05 0.07 0.03 0.33 0.52 2.72 0.313 time per one frame turns 0.05 0.04 0.06 0.03 0.31 0.57 2.70 0.24 (ms) whole trajectory 0.06 0.03 0.06 0.03 0.33 0.53 2.72 0.30 Table 1 presents the processing results. The methods were evaluated by the average number of the pairs of corresponding points (greater values are better) and the average processing per one frame (lower values are better). According to the first criterion, the methods implemented in the OpenCV library were the best: RANSAC, LMS, RHO. They increased the number of the pairs of corresponding points by ~ 20 pairs (~ 13% of the initial quantity). However, the speed of processing is critical for the implementation of a UAV navigation system. According to this criterion, the proposed iterative method with the variants 1-3 for the pair selection was the best. It takes ~0.03–0.06ms for processing one frame, which is ~5–10 times faster than the best of the OpenCV methods, with that it increases the number of the pairs of corresponding points by ~15 pairs (~10%). The high speed of the proposed method is due to the fact that in most cases two iterations are enough to obtain the resulting homography matrix. Thus, for the second experiment according to the both criteria the following methods were chosen: the proposed iterative method with the first variant of the pair selection and the RHO method. The second experiment aims to determine the effect of the integration of the normalization method into the UAV visual navigation system. There were three trials in the experiment: without normalization, with normalization using the proposed method, and using the RHO method. The first criterion was the number of the images from the UAV camera successfully matched with the satellite image. The second criterion was the average distance between the automatically calculated positions of the center of photographing and the manually marked ones. The results of the experiment are shown in table 2. After analyzing the results of the experiment, we can conclude that the preliminary normalization of the image significantly improves the output of the UAV visual navigation system being developed. It increases the number of the successfully matched frames as well as the accuracy of the calculated coordinates of the UAV. The methods work well both in the straight sections of the flight trajectory and in the sections with turns. The result of the proposed method is not worse than the one of the RHO V International Conference on "Information Technology and Nanotechnology" (ITNT-2019) 150 Image Processing and Earth Remote Sensing M O Elantcev, I O Arkhipov and R M Gafarov method from the OpenCV library, and at that the proposed method performs much faster, which is very important for integrating it into a UAV visual navigation system. Table 2. The effect of normalization methods on the output of the visual navigation system. Without Proposed method with Indicator Section RHO normalization variant 1 of pair selection straight sections 270 Number of frames turns 70 whole trajectory 340 straight sections 226 (83.7%) 246 (91.1%) 246 (91.1%) Number of successfully turns 59 (84.3%) 64 (91.4%) 62 (88%) matched frames whole trajectory 285 (83.8%) 310 (91.1%) 308 (90.6%) Average distance between straight sections 16.0 6.6 6.8 the found center of turns 7.5 6.4 6.7 photographing and the manually marked one whole trajectory 16.0 6.6 6.7 (pixels) 6. Conclusion The proposed method of iterative image normalization successfully solves the task of image preprocessing for developing a UAV visual navigation system. The integration of this method into the navigation system improves the accuracy of matching the images acquired from the onboard camera with the satellite image and reduces the probability of the UAV loss. Due to using the two normalization approaches the proposed method can eliminate perspective distortions resulting from both the camera orientation and the inclination of the underlying surface. At the same time, this method is computationally simple and can be integrated into the UAV navigation system, which should run on low-power processors. 7. References [1] Gafarov R M, Arkhipov I O, Korobeynikov A V and Elantsev M O 2017 Application of the method of statistical differentiation for the gridding of aerophoto image to a satellite image Intellektualnye sistemy v proizvodstve 15 109-112 DOI: 10.22213/2410-9304-2017-2-109-112 [2] Kholopov I S 2017 Projective distortion correction algorithm at low altitude photographing Computer Optics 41(2) 284-290 DOI: 10.18287/0134-2452-2017-41-2-284-290 [3] Boriskevich A A 2016 A robust method for normalization of barcode images scanned with a mobile device Izvestija Nacionalnoj akademii nauk Belarusi. Serija fiziko-tehnicheskih nauk 4 106-112 [4] Adel E, Elmogy M and Elbakry H 2014 Image stitching based on feature extraction techniques: a survey International Journal of Computer Applications 88 1-8 [5] Goncharenko M O 2015 Comparative analysis of key points detection methods as a part of video segmentation problem Bionics of Intelligense 85 90-94 [6] Hartley R and Zisserman A 2003 Multiple View Geometry in Computer Vision (Cambridge: Cambridge university press) [7] Efimov A I and Novikov A I 2016 An algorithm for multistage projective transformation adjustment for image superimposition Computer Optics 40(2) 258-265 DOI: 10.18287/2412- 6179-2016-40-2-258-265 [8] Agarwal A, Jawahar C V and Narayanan P J 2005 A survey of planar homography estimation techniques Centre for Visual Information Technology International Institute of Information Technology [9] Goshin Y V, Kotov A P and Fursov V A 2014 Two-stage formation of a spatial transformation for image matching Computer Optics 38(4) 886-891 DOI: 10.18287/0134-2452-2014-38-4-886- 891 [10] Bazargani H, Bilaniuk O and Laganiere R 2015 A fast and robust homography scheme for real- V International Conference on "Information Technology and Nanotechnology" (ITNT-2019) 151 Image Processing and Earth Remote Sensing M O Elantcev, I O Arkhipov and R M Gafarov time planar target detection Journal of Real-Time Image Processing 15 739-758 DOI: 10.1007/s11554-015-0508-4 [11] DeTone D, Malisiewicz T and Rabinovich A 2016 Deep image homography estimation arXiv (Preprint arXiv:1606.03798) [12] Glasbey C A and Mardia K V 1998 A review of image-warping methods Journal of applied statistics 25 155-171 [13] Arkhipov I O, Mishenkov M A, Shutov A A and Elantsev M O 2018 Algorithm of unmanned aerial vehicle’s positioning using sequence of video stream frames Intellektualnye sistemy v proizvodstve 16 66-69 DOI: 10.22213/2410-9304-2018-3-66-69 [14] Gafarov M R, Gafarov R M and Poltanova E S 2015 Determination of geometrical images parameters using statistical differentiation Intellektualnye sistemy v proizvodstve 25 41-44 [15] Chekhlov D O and Ablameyko S V 2004 Normalization of images relating to perspective transformation based on geometric options Informatika 3 67-76 Acknowledgments The reported study was funded by Kalashnikov ISTU according to the research project β„– 09.06.01/18 GMM. V International Conference on "Information Technology and Nanotechnology" (ITNT-2019) 152