Spline model of digital images based on moment characteristics Anhelina Zhultynska1,2,† and Pylyp Prystavka1,∗,† 1 National Aviation University, Liubomyra Huzara Ave. 1, Kyiv, 03058, Ukraine 2 Interregional Academy of Personnel Management, Frometivska Str., 2, Kyiv, 03039, Ukraine Abstract The paper proposes a spline model of a parametrized digital image based on two momentary texture features, namely, the locally averaged illumination intensity and the root mean square. The model is an estimate of the density function of the distribution of these characteristics and is an alternative to the model of a mixture of normal distributions. Keywords computer vision, image recognition, image preprocessing 1 1. Introduction and problem statement The problem of digital image segmentation is not new – there are numerous diverse and effective approaches to solving it, both in purely mathematical methods and through the use of hardware- software approaches. However, the segmentation of oversized realistic images or video data (e.g., from aerial or space observations) in real-time or near-real-time mode gains particular relevance. By its nature, the segmentation (clustering) problem relates to unsupervised pattern recognition, generally solved based on the model of feature distribution density defining certain characteristics of digital images. In general, the problem of pattern recognition can be broken down into the following main stages [1]: 1. Generating features of the recognized object. 2. Selecting features of the recognized object. 3. Building a classifier. 4. Assessing the quality of classification. Feature formation is one of the most critical stages in constructing any information processing system. The quality of the chosen feature system and algorithms for their computation depends on the specific requirements of the applied system for processing and analyzing digital images. Therefore, when prioritizing the requirement for segmentation speed, it is necessary to pose a demand for both the high adequacy of the density function model on one hand and the efficiency of obtaining and learning it on the other hand [2, 3]. Hence, the relevant research direction of this work will be to define a model of realistic digital images that would address the segmentation task in a mode close to real-time within modern computerized systems CH&CMiGIN’24: Third International Conference on Cyber Hygiene & Conflict Management in Global Information Networks, January 24–27, 2024, Kyiv, Ukraine ∗ Corresponding author. † These authors contributed equally. angelinaremark1@gmail.com (A. Zhultynska); chindakor37@gmail.com (P. Prystavka) 0000-0001-9178-897X (A. Zhultynska); 0000-0002-0360-2459 (P. Prystavka) © 2025 Copyright for this paper by its authors. Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0). CEUR ceur-ws.org Workshop ISSN 1613-0073 Proceedings 2. Literature review To further the analysis, let's introduce the concept of texture. Texture refers to fragments of an image that can be characterized by different properties. Texture can be homogeneous when the pattern or structure repeats with the same intensity in all directions and at all points. For example, the surface of sand on a beach or a field can have a homogeneous texture. On the other hand, texture can be heterogeneous when the pattern or structure varies at different points or directions. For instance, an image of a forest with different types of trees, shrubs, and other plants can have a heterogeneous texture. Let's consider methods for forming textural features related to the segmentation and classification of digital images, commonly encountered in tasks of processing aerial photography, aerial, and space monitoring data. The simplest and often-used segmentation method is based on dividing the image by a brightness threshold [4, 5]. For this, the intensities of each image element are used. If the intensity of a certain selected threshold value increases, the element is classified as an object component; otherwise, it's considered a background element. However, such methods may be ineffective for complex image structures like aerial photographs. The application of segmentation methods based on textural characteristics holds great potential in various image processing areas. Statistical methods, such as local dispersion analysis [6, 7], histogram-based analysis [8], and using local binary patterns for feature extraction and region merging [9], efficiently highlight textural areas of images by computing statistical characteristics. They can be effective in recognizing different types of textures but may require substantial computational resources for calculating a large number of characteristics. Structural methods, like the multi-layer adaptive graph cut method [10], excel in segmenting textures with complex structures by identifying primitive structures. However, they may be less effective for irregular textures or textures without known standards. Spectral methods, like deep learning-based methods and conditional random fields [11, 12], ensure high segmentation accuracy by processing deep features and modeling context using conditional random fields. Additionally, wavelet transform methods also prove effective for segmenting textural images. Wavelet transformation [13, 14] enables the extraction of textural details of various scales in an image using multi-scale analysis. The Gaussian Mixture Model (GMM) method for constructing digital image models is another prevalent approach to data segmentation [15]. This method involves creating a complex image model using a combination of several Gaussian functions with different parameters. The use of the Expectation-Maximization (EM) algorithm adapts GMM to a dataset by considering the multimodal data distribution as a combination of individual unimodal distributions. GMM assumes that data in a certain cluster is generated by a specific Gaussian distribution. The method allows approximating complex structures and details in an image, facilitating their further analysis. However, constructing a complex model requires significant computational resources, which can complicate processing a large amount of data. Therefore, the analysis of research and sources shows that a computationally simple approach to texture analysis is possible based on the construction of textural features using the distribution of statistical characteristics, given the simplicity of estimating (modeling) the corresponding probability density function. According to studies [16], two-dimensional splines based on B-splines meet similar conditions, closely resembling interpolations on average. This approach may have advantages in terms of computational efficiency and interpretability. However, its accuracy may be limited compared to some deep learning models capable of solving more complex texture classification tasks [17–21]. Therefore, the goal of this work is to formalize and justify the possibility of applying the mentioned splines to the task of constructing a digital image model based on moment characteristics for automated processing systems of digital images, including aerial observation data. 3. The study materials and methods = , , = 0, − 1; = 0, − 1 , Let's consider a certain digital image (DI), represented by the illumination intensity of each pixel (1) where is for instance, any of the color components in the RGB color space or a component representing DI in grayscale; , are the linear dimensions of DI. If we consider a realistic DI, such as an aerial image, then any element of intensity (1) is a random Let = , , , ,…, , ; = 1, variable, so when choosing a DI model, this fact should be taken into account. within a certain neighborhood determined by the size of a sliding window (2 + 1) × (2 + 1), be a set of textures that exhibit similar characteristics = 1,2, …, where − texture; ,! − characteristic of the -th texture; − number of textures; " − number of characteristics of the texture. In fact, each of the features of a separate texture ,! is defined as some random function of the pixel intensities in the local (2 + 1) × (2 + 1) area around any ( , )-th pixel of DI: ,! = #! $,% ,& = − , + ,' = − , + , ( = 1, ", = 0, − 1, = 0, − 1. (2) Therefore, we're discussing the parameterization of DI (2) based on texture characteristics. In such a case, a DI model can be proposed by estimating the density distribution of parameters in the texture characteristic space. Regarding each separate g-th group of "similar" textures, due to the randomness of characteristics, it's possible to assume the existence of some parameter distribution density * ( , , , , … , , ), function (3) . where the natural assumption is that the texture group distribution present in DI is a mixture *(Θ) = , - *( , , , ,…, , ), / . , - = 1. (4) / For instance, assuming a normal distribution of densities (3), the task of evaluating model (4) boils down to reproducing a mixture of normal distributions in the texture characteristic space. Solving this latter task would allow for image segmentation (clustering), for example, based on dividing the mixture followed by a naive Bayesian pixel classification (1). Despite all the advantages (low sensitivity to "noise" and anomalous values, independence of the clustering result from the data entry order, the ability to highlight clusters of different structures, etc.), the algorithms developed to solve such a task have significant drawbacks – the assumption of a normal density distribution (3) is not always valid, hence the limited applicability functional of the mixture model (4). Moreover, such algorithms are characterized by relative computational complexity, thus not recommended for processing large DIs or streaming video. In general, it can be assumed that the hypothesis of local continuity and smoothness of the density distribution of texture characteristics holds true – for most *( ) ∈ С2,…,2 , 3 ≥ 2 . realistic images, this is quite acceptable. Therefore, without reducing generality, let's consider that Further on, let's examine the case when the number of texture characteristics " = 2 is only two. For example, these could be parameters characterizing the average color (illumination) of a group of pixels and their variability. Publications [1, 4, 6] have shown that moment characteristics of DI are invariant to scale and rotation with respect to the size of the local window, so for the parameterization that defines the properties of textures, let's limit the choice to estimating the local 67 67 1 mean and standard deviation of pixel intensities in the area: 5 , = , , $,% , (2 + 1) (5) $/ 87 %/ 87 67 67 1 9, =: , , $,% − 5 , , (2 + 1) (6) $/ 87 %/ 87 = 3, − 3 − 1, = 3, − 3 − 1. dimensional sample (5, 9) in the texture feature space ; ,< = (5! , 9! ); ( = 1, = , where = is the Therefore, as a result of calculating the characteristics (5) and (6), we obtain a random two- worth noting that the order of forming the sample Ω ,< is not critical in this presentation: the number of sliding windows within which mean and standard deviation estimates are calculated. It's calculation can be done for a sliding window of size (2 + 1) × (2 + 1) with a shift of 1 pixel or any other shift of the central element of the area relative to the pixels of DI. Moreover, for the With the sample Ω ,< , we will proceed to construct a two-dimensional variational series divided following exposition, the magnitude of the positive integer is also not crucial. into classes. For this, we'll define the division ?@A,@B of the plane of realizations of characteristics (5) Algorithm for constructing the partition CDE ,DF . and (6) according to the following algorithm of actions. Step 1. The central point of the (G, H)-th partition element ?@A,@B is taken as the variant row (5I , 9I ) 5I = 5JIK + (G + 0,5)ℎJ , G = 0, NJ − 1, 9O = 9JIK + (H + 0,5)ℎP , H = 0, NP − 1, 5JWX − 5JIK 5JIK = minT5! U ; 5JWX = maxT5! U ; ℎJ = ; where ! ! NJ 9JWX − 9JIK 9JIK = minT9! U ; 9JWX = maxT9! U ; ℎP = ; ! ! NP where NJ , NP − the number of partition elements (classes) along the respective axes. Step 2. The frequency [I,O of the variant row is calculated as the number of sample points Ω ,< falling within the limits of the (G, H)-th partition element ?@A,@B , whereby the indices G, H determining to which of the partition elements ?@A,@B any point (5! , 9! ); ( = 1, = belongs, are 5! 9! G = \ ], H =\ ], found using the expressions ℎJ ℎP where ^∙` is the integer part. Step 3. The randomness aI,O of the variants (5I , 9I ) is determined by the relation bA 8 bB 8 [I,O aI,O = , , , aI,O = 1. = I/c O/c Let us consider the estimation of the feature density function of 5 and 9 as a parameterised image model and evaluate it on the basis of the sample ; ,< = (5! , 9! ); ( = 1, = and the two-dimensional 5I , 9O , aI,O ; G = 0, NJ − 1, H = 0, NP − 1 variation series built on its basis of this random variable. Let the density function be smooth. If (5I , 9I ) is the centre point of the (G, H)- th element of the partition ?@A,@B , then 1 Jh 6c,g@A Pf 6c,g@B ̅ = *I,O e e *(5, 9)'5'9 ℎJ ℎP Jh 8c,g@A Pf 8c,g@B is the average value of the probability density function of the features (5, 9) in the specified area and the following connection with the randomness of the option takes place ̅ ℎJ ℎP = aI,O = *I,O = i j: 5I − 0,5ℎJ ≤ m(j) < 5I + 0,5ℎJ , 9O − 0,5ℎP ≤ o(j) < 9O + 0,5ℎP Thus, up to the constant ℎJ ⋅ ℎP , the randomness is an estimate of the average value of the density function *(5, 9). Then, as a model for density approximation, we propose a model in the form of a two-dimensional polynomial spline based on q-splines that is close to the interpolation average. That is, the array of values a̅ = a̅I,O , G ∈ s, H ∈ s will be matched with the following local t ,c (a, 5, 9) = , , q ,@A (5 − GℎJ )q ,@B (9 − HℎP )aI,O polynomial spline [18]: I∈u O∈u (7) 0, 5 ∉ ^− 3ℎ⁄2 ; 3ℎ⁄2`, where (with the notation accuracy up to the split step) [12] ⎧ (3 + 25⁄ℎ) ⁄8 , 5 ∈ ^− 3ℎ⁄2 ; − ℎ⁄2`, q ,@ (5) = ⎨3⁄4 − (25⁄ℎ) ⁄4 , 5 ∈ ^− ℎ⁄2 ; ℎ⁄2`, ⎩ (3 − 25⁄ℎ) ⁄8 , 5 ∈ ^ℎ⁄2 ; 3ℎ⁄2`. 1 or in expanded form t ,c (a, 5, 9) = (1 − •) (1 − €) aI8 ,O8 + (1 − •) (6 − 2€ )aI8 ,O + 64 +(1 − •) (1 + €) aI8 ,O6 + (6 − 2• )(1 − €) aI,O8 + (6 − 2• )(6 − 2€ )aI,O + +(6 − 2• )(1 + €) aI,O6 + (1 + •) (1 − €) aI6 ,O8 + (1 + •) (6 − 2€ )aI6 ,O + (8) +(1 + •) (1 + €) aI6 ,O6 • = @ (5 − GℎJ ), |•| ≤ 1 ; € = @ (9 − HoP ), |€| ≤ 1. where A B precise estimation of the approximation error of the density distribution *(5, 9) indirectly through By means of the explicit analytical representation (8), the models (7) can obtain an asymptotically the estimation of the function a(5, 9), the values of which are the relative frequencies of the aI,O = a̅I,O + ‚I,O , respective two-dimensional variational series: 1 (Jh 6c,g) @A Pf 6c,g @B where āI,O = e e a(5, 9)'5'9 , ℎJ ℎP (Jh 8c,g) @A Pf 8c,g @B ‚I,O − an arbitrary error. In particular, for ℎJ → 0, ℎP → 0 for ∀a(5, 9) ∈ • , is true ^10`: ℎJ ” ℎP ’a(5, 9) − t ,c (a, 5, 9)’С ≤ ’aJ“ (5, 9)’С + ’aP”“ (5, 9)’С + 6 6 ℎJ ℎP (–) + •aJ“ P“ (5, 9)• + ‚‖a(5, 9)‖˜ + ™(ℎ– ) (9) 36 С ‚ = max ‚G.H , ℎ = maxTℎJ , ℎP U. where G,H 4. Results of the study a(5, 9), for instance, allows solving the segmentation problem. With this approach, individual Analyzing the local peculiarities of the constructed model of the two-dimensional normalized density segments correspond to clusters determined by local maxima (modes) of the estimate (7). As an application example of model (7), consider an image representing several textures: a grass lawn, a runway, airplanes (Figure 1). We'll illustrate how these textures look locally and what distribution they have. We'll consider the two-dimensional density functions depending on the mean and standard deviation for each of the textures. The lawn is a simple texture (Figure 2a). By “simple”, we mean relatively low variability of light (Figure 3) based on model (7) is unimodal and has a mean 44 ≤ 5 ≤ 84 and standard deviation 9 ≤ intensity within a local window. The estimation of the normalized density function of the “lawn” 14. Figure 1: Image with textures: grass lawn, runway, airplanes. a) b) c) Figure 2: Examples of textures: a) grass lawn; b) runway; c) airplane. intensity 134 ≤ 5 ≤ 170 and variability 9 ≤ 12 is also unimodal (Figure 4). Similarly, the runway is a simple texture (Figure 2b). The distribution density of the average a) b) window: а) = 4; b) = 8. Figure 3: Two-dimensional normalized density function of the lawn (Figure 2a) with a sliding a) b) window: а) = 4; b) = 8. Figure 4: Two-dimensional normalized density function of the runway (Figure 2b) with a sliding Complex textures are characterized by high variability in light intensity due to the presence within a local window of pixels from various 'simple' textures, for example, the 'airplane' texture (Figure 2c). The normalized density function (Figure 5) is multimodal, indicating the presence of pixels from the “grass lawn”, “runway” and airplane. a) b) window: а) = 4; b) = 8. Figure 5: Two-dimensional normalized density function of the airplane (Figure 2c) with a sliding parameterised by (5, 9) (Figure 1). There are three local groups of pixels defined by the proposed Let us consider (Figures 6 and 7) the density function of the entire original digital image model (7). According to a natural analysis, the first group of pixels belongs to the lawn, the second to the runway, and the third to the airplanes (Figures 6a and 7a). Visually, it's not difficult to identify local maxima and intuitively divide the model into separate segments (Figures 6b and 7b), which determine the corresponding segmentation of textures in the original image. = 4: а) spline model; b) 1 – a) b) Figure 6: Spline model of the image (Figure 1) with a sliding window local maxima defining separate image segments. a) = 8: а) spline model; b) 1 – b) Figure 7: Spline model of the image (Figure 1) with a sliding window local maxima defining separate image segments. 5. Conclusions In the study, a spline model (7) for parameterized digital imagery (DI) based on two momentary characteristics of textures was proposed, namely: the locally averaged illumination intensity (5) and the root mean square (6). This model represents an estimation of the probability density function of the mentioned characteristics and serves as an alternative to a mixture model of normal distributions. Thanks to the analytical presentation (8), the spline model demonstrates low computational complexity, substantiated acceptable approximation errors, and can be recommended for implementation in automated DI processing systems operating in near real-time regimes. Future research involves developing a DI segmentation method based on the introduced spline model and procedures for detecting low-probability textures that provide enhanced informativeness in the analysis of aerial observation data. Additionally, the results of this work are expected to be used in developing neural network models and exploring the hyperparameters of deep learning algorithms. Declaration on Generative AI The author(s) have not employed any Generative AI tools. References [1] R. M. Haralick, K. Shanmugam, I. Dinstein, Textural features for image classification, IEEE Transactions on Systems, Man and Cybernetics 3(6) (1973) 610–621. doi: 10.1109/tsmc.1973.4309314. [2] J. S. Al-Azzeh, M. Al Hadidi, R.S. Odarchenko, S. Gnatyuk, Z. Shevchuk, Z. Hu, Analysis of self- similar traffic models in computer networks, International Review on Modelling and Simulations 10(5) (2017) 328–336. doi: 10.15866/iremos.v10i5.12009. [3] O. Popov, A. Iatsyshyn, V. Kovach, V. Artemchuk, D. Taraduda, V. Sobyna, et al., Physical features of pollutants spread in the air during the emergency at NPPs, Nuclear and Radiation Safety 4 (84) (2019) 88–98. doi: 10.32918/NRS.2019.4(84).11. [4] R. D. da Silva, R. Minetto, W. R. Schwartz, H. Pedrini, Satellite Image Segmentation Using Wavelet Transforms Based on Color and Texture Features, ISVC 2 (2008) 113–122. doi: 10.1007/978-3-540-89646-3_12. [5] T. Tran et al., Enabling multicast and broadcast in the 5G core for converged fixed and mobile networks, IEEE Transactions on Broadcasting 66(2) (2020) 428–439. doi: 10.1109/TBC.2020.2991548. [6] P. Prystavka, K. Dukhnovska, O. Kovtun, O. Cholyshkina, A. Zhultynska, Devising information technology for determining the redundant information content of a digital image, Eastern- European Journal of Enterprise Technologies 6 (2021) 59–70. doi: 10.15587/1729- 4061.2021.248698. [7] A. Chyrkov, P. Prystavka, Method for suspicious object search in video from an aircraft camera based on histogram analysis, Science-based technologies 2(38) (2018) 210–219. doi: 10.18372/2310-5461.38.12827. [8] H. H. Permuter, J. M. Francos, I. Jermyn, A study of Gaussian mixture models of color and texture features for image classification and segmentation, Pattern Recognition 39 (2006) 695– 706. doi: 10.1016/j.patcog.2005.10.028. [9] G. Sun, Y. Hao, X. Chen, J. Ren, A. Zhang, B. Huang, Y. Zhang, X. Jia, Dynamic Post-Earthquake Image Segmentation with an Adaptive Spectral-Spatial Descriptor. Remote Sensing 9 (2017) 899. doi: 10.3390/rs9090899. [10] J. Liu, K. He, Multi-scale foreground extraction on graph cut, MATEC Web of Conferences 277 (2019). doi: 10.1051/matecconf/201927702031. [11] X. Song, L. Wu, G. Liu, Unsupervised color texture segmentation based on multi-scale region- level Markov random field models, Computer Optics 43(2) (2019) 264–269. doi: 10.18287/2412- 6179-2019-43-2-264-269. [12] S. Ghaffarian, A Kernel-Based Post-Process for Image Segmentation Using GVF Snake, Journal of Automation and Control Engineering 2(3) (2014) 277–281. doi: 10.12720/joace.2.3.277-281. [13] P. Ganesan, V. Rajini, R. G. Salai, Segmentation and Denoising of Noisy Satellite Images based on Modified Fuzzy C-Means Clustering and Discrete Wavelet Transform for Information Retrieval, Int. J Eng. Technol. 5 (2013). [14] C. Wang, W. Pedrycz, J. Yang, M. Zhou, Z. Li, Wavelet Frame-Based Fuzzy C-Means Clustering for Segmenting Images on Graphs, IEEE Trans. Cybern. 50 (2020) 3938–3949. doi: 10.1109/tcyb.2019.2921779. [15] L. P. Osco, et al., A review on deep learning in UAV remote sensing, International Journal of Applied Earth Observation and Geoinformation 102 (2021) 102456. doi: 10.1016/j.jag.2021.102456. [16] A. Ligun, A. Shumeiko, Asymptotic Methods of Curve Reconstruction, NASU Institute of Mathematics, 1997. [17] M. Unser, Splines: A Perfect Fit for Signal and Image Processing, IEEE Signal Processing Magazine 16(6) (1999) 22–38. doi: 10.1109/79.799930. [18] P. Prystavka, Polynomial Splines in Data Processing, Dnipropetrovsk, 2004. [19] L. Shapiro, G. Stockman, Computer Vision, Seattle Washington, 2001. [20] P. Prystavka, M. Ryabiy, A model of realistic images based on two-dimensional splines close to interpolation on average, Science-intensive technologies 3(5) (2012) 67–71. [21] P. Prystavka, A. Rogatyuk, Mathematical foundations of foreign object recognition in the video from unmanned aircraft, Proceedings of the National Aviation University 3(64) (2015) 133–139. doi: 10.18372/2306-1472.64.9048.