=Paper= {{Paper |id=Vol-3156/paper35 |storemode=property |title=On Image Complexity in Viewpoint of Image Processing Performance |pdfUrl=https://ceur-ws.org/Vol-3156/paper35.pdf |volume=Vol-3156 |authors=Vladimir Lukin,Sergey Krivenko,Fangfang Li,Sergey Abramov,Victor Makarichev |dblpUrl=https://dblp.org/rec/conf/intelitsis/LukinKLAM22 }} ==On Image Complexity in Viewpoint of Image Processing Performance== https://ceur-ws.org/Vol-3156/paper35.pdf
On Image Complexity in Viewpoint of Image Processing
Performance
Vladimir Lukinа, Sergey Krivenkoа, Fangfang Liа, Sergey Abramovа and Viktor Makarichevа
a
    National Aerospace University, 17 Chkalova Street, Kharkiv, 61070, Ukraine


                 Abstract
                 It is well known that image processing efficiency considerably depends on image properties.
                 Among operations of image processing, we mean quality assessment, noise characteristic
                 estimation, lossless and lossy compression, denoising, etc. In many papers, such terms as
                 “image complexity”, “rich image content”, “highly textural image” are used. Their meaning is
                 intuitively clear and described verbally but there is a limited amount of quantitative estimation
                 and analysis. In this paper, we show that there is a very high correlation of performance of
                 different operations of image processing like characteristics of lossless and lossy compression,
                 blind variance estimation and denoising and so on. We also recall what quantitative parameters
                 indirectly describe image complexity for different operations of image processing.

                 Keywords 1
                 Image complexity, performance, image processing operations, rank correlation.

1. Introduction
    Images have become an essential part of our life [1-3]. They are applied in numerous areas like
entertainment and advertising [4, 5], ecological monitoring, agriculture [6], forestry [7], etc. Images are
subject to many different operations at stages of their acquisition, transferring, pre-processing,
enhancement, and so on [1]. This can be quality estimation [8], blind evaluation of noise type and
characteristics [9], denoising and enhancement [4], lossless and lossy compression [10], object
detection and classification [11], etc.
    Certainly, not all these operations are employed in an image processing chain for a given image and
a given application. Some of them are not needed or can be skipped. Meanwhile, practically each
scientist/researcher dealing with image processing knows that there are images processing of which is
relatively easy and successful and there are images processing of which is problematic (difficult, does
not bring an obvious success). The latter usually happens for images that have a small percentage of
pixels belonging to homogeneous regions, contain a lot of small-sized objects, edges, textures. For
example, designers of nonlinear filters (denoising techniques) for grayscale and color images know [12]
that it is considerably more difficult to get good results for the test image Baboon than for the test
images Lena and Peppers. This is one of reasons why it is usually recommended to test performance of
an image processing method for a set of images containing images with different properties. This was
also the reason for creating image databases like LIVE (http://live.ece.utexas.edu/research /quality/),
TID2008         (http://www.ponomarenko.info/tid2008.htm)          and       TID2013          (http://www.
ponomarenko.info/tid2013.htm), Kodak (http://r0k.us/graphics/kodak/), etc. Nowadays, there is a
tendency of creating medical (for example, SICAS Medical Image Repository, https://www.smir.ch)
and remote sensing (for instance, ESA “Sentinel-2”, https://sentinel.esa.int/web/sentinel/missions/


IntelITSIS’2022: 3rd International Workshop on Intelligent Information Technologies and Systems of Information Security, March 23–25,
2022, Khmelnytskyi, Ukraine
EMAIL: lukin@ai.kharkov.com (V. Lukin); krivenkos@ieee.org (S. Krivenko); liff_niat@yahoo.com (Fangfang Li); s.abramov@khai.edu
(S. Abramov); victor.makarichev@gmail.com (V. Makarichev)
ORCID: 0000-0002-1443-9685 (V. Lukin); 0000-0001-6027-5442 (S. Krivenko); 0000-0002-7392-586X (Fangfang Li); 0000-0002-8295-
9439 (S. Abramov); 0000-0003-1481-9132 (V. Makarichev)
              ©️ 2022 Copyright for this paper by its authors.
              Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0).
              CEUR Workshop Proceedings (CEUR-WS.org)
sentinel-2) image databases for different purposes, in particular, for training modern neural networks
(deep learning) for which a huge volume of training and verification data is needed. There are also
texture image database USC-SIPI (https://sipi.usc.edu/database/), European Space Agency database of
color images (https://www.esa.int/ESA_Multimedia/Images) and many others. The existence of such
databases offers such opportunities as statistical analysis of indicators of image processing efficiency
[13], detection of outlying results, anomalies and “strange” images [14], thorough comparisons of
methods’ performance [15], scatter-plot obtaining and regression (curve fitting) [16, 17], etc.
    As the result of such analysis for a particular application, images are often classified to simple
structure and complex structure (textural) ones where there are no objective criteria for such a
classification (or the criteria are mostly intuitive or expressed verbally). Because of this, the goal of this
paper is to generalize the problem, to show some criteria characterizing image complexity, and to
demonstrate inter-connection between such criteria and image processing efficiency.

2. Lossless and lossy compression
    Probably, the most known characteristic (criteria) of image complexity is its entropy that directly
influences image compression ratio for lossless compression (coding) techniques [18, 19]. Let us give
one example. Consider 25 color test images used in the databasesTID2008 and TID2013. Twenty four
of them are natural scene images and one (#25) is artificial. Small copies of these images are given in
Fig. 1. As one can see, images are of different complexity where the test image #13 is the most complex
since it is practically impossible to find homogeneous (non-textural) fragments for it. Meanwhile, there
are also quite simple structure images like, e.g., the test images ## 3 or 20 with rather large
homogeneous regions.
    These properties appear themselves in entropy E calculated for all 25 test images. Its values are
given in Fig. 2,a. They are in the limits from about 6 for the test image # 2 till about 7.5 for the test
images ## 13, 14, and 19. Alongside, Fig. 2 shows the values of compression ratio (CR) attained by
ZIP. Comparison of data in Figures 2,a and 2,b clearly shows that a larger CR is usually obtained for
images having smaller entropy values and vice versa. The Spearman rank order correlation coefficient
(SROCC) is equal to -0.62 and its quite large absolute value shows that there is a rather strict rank
correlation between E and CR.
    Let us also present some data for grayscale images. In the paper [14], a set containing 61 images has
been analyzed. This set included standard test images as Baboon, Lena and others, a lot of texture
images from the USC-SIPI database, and some others. Entropy values have been determined and they
are shown by dots of different colors in Fig. 3. As one can see, E values are from 0.5 till 7.9, whist CR
for lossless compression CRlossless varies from 100.7 to 1. Correlation factors for these two indicators
(parameters) are the following: -0.74 for Pearson correlation, -0.88 for Spearman rank correlation, and
-0.76 for Kendall rank correlation. In other words, a larger CRlossless usually relates to a simpler structure
image with smaller entropy. In [14], it has been proposed to divide images into three groups: 1) complex
structure images having E>7; 2) middle complexity images for which 6≤E≤7; 3) simple structure images
with E<6. Fig. 4 shows examples for all three groups. The entropy values are equal to 5.82, 6.80, and
7.36, respectively.
    It is also interesting that among simple structure images (E<6) there are three images that have very
small E (about 3 and less). Analysis shows that these are artificially created images or scanned copies
of such images (see Fig. 5) that have large homogeneous regions (or a large percentage of pixels that
belong to such regions). Further analysis will show that performance characteristics of image processing
operations can differ a lot from performance characteristics for natural scene images. This indicates that
it is necessary to use such artificial images with care in analysis of different operations of image
processing (denoising, compression, classification) since conclusions and tendencies obtained based on
the corresponding studies can be wrong or, at least, be aside of tendencies and conclusions for most
typical images met in practice. In other words, if there is no possibility to carry out studies concerning
image processing except of using artificially created images, it is worth making such images as close to
natural as possible.
    In lossy image compression [10, 14, 16], there is a much wider variety of performance
characteristics. First, CR can be varied where this is done in different ways in different coders. CR can
be controlled and varied by quality factor, scaling factor, quantization step, bits per pixel (BPP) and so
on. Second, since distortions (losses) are introduced, they can be characterized using various criteria.
These can be traditional criteria as mean square error (MSE) or peak signal-to-noise ratio (PSNR)
strictly connected with MSE; meanwhile, visual quality metrics (see [14, 26] and the metric PSNR-
HVS-M, http://www.ponomarenko. info/psnrhvsm.htm) are also widely used since visual inspection is
often an ultimate goal of compressed images.




         1                    2                     3                    4                     5




         6                    7                     8                    9                    10




        11                    12                    13                   14                   15




        16                    17                    18                   19                   20




        21                    22                    23                   24                   25

Figure 1: Small copies of color images in TID2008




                         a                                                    b
Figure 2: Entropy (a) and CR (b) values for 25 test color images in TID2008




Figure 3: Entropy values for 61 test grayscale images




                 a                               b                                 c
Figure 4: Examples of simple (a), middle (b) and complex (c) structure grayscale images

    Visual quality metrics take into account some peculiarities of human vision system (HVS). For
example, the metric PSNR-HVS-M (see PSNR-HVS-M official website http://www.ponomarenko.
info/psnrhvsm.htm) takes into account two properties: a larger sensitivity to distortions in low spatial
frequencies and a masking effect of texture. PSNR-HVS-M is expressed in dB and it has smaller values
for more degraded images. Due to aforementioned properties, PSNR-HVS-M is usually larger than
PSNR. This takes place until distortions due to lossy compression become too large and spatially
correlated, then masking effect disappears and PSNR and PSNR-HVS-M become practically equal or
even PSNR-HVS-M occurs to be smaller than PSNR. Such an example is shown in Fig. 5,a for the
coder SPIHT (http://www.spiht.com/). Recall that since compression parameters can be varied we have
to analyze the so-called rate/distortion curves that can be presented in different ways. For SPIHT, this
is dependence of a metric on BPP whilst, e.g., for the coder BPG (better portable graphics,
https://bellard.org/bpg/) these are dependences of a metric on parameter Q that controls compression
(Q falls into integer values from 1 to 51 and a larger Q results in worse quality). Examples of
rate/distortion curves are given in Fig. 5,b. Note that usually the rate/distortion curves are smooth
monotonous functions (Fig. 5,b). Meanwhile, they can be “not very smooth” like in Fig. 5,a and even
not monotonous as this happens for “strange images” – see examples in Fig. 5,c and 5,d.
    Behavior of the rate/distortion curves for particular images has common features but is individual.
The examples in Fig. 6 clearly show this for two coders: SPIHT and AGU. As one can see, there are
sufficient differences in PSNR values for the same BPP for SPIHT and QS for AGU.
                            a                                               b




                         c                                            d
Figure 5: PSNR and PSNR-HVS-M on BPP for SPIHT, the test image Goldhill (a), PSNR and PSNR-HVS-M
on Q for BPG, the test image Goldhill (b), PSNR on quantization step (QS) for the coder AGU
(http://www.ponomarenko.info/agu.htm) for “strange images” (c), PSNR-HVS-M on QS for AGU for
“strange images” (d)




                        a                                                    b
Figure 6: Examples of rate/distortion curves for particular images for different coders: PSNR on BPP
for SPIHT (a) and PSNR on QS for AGU (b)

   Although some curves intersect, it is quite probable that if a metric value for an image1 is larger
than for an image2 for BPP1, it will be also larger for BPP2. Similar holds for rate/distortion curves for
the coder AGU – the curves for more complex structure images go “below” the curves for simpler
structure images.
    It is easy to show that entropy that characterizes image complexity is highly correlated with
performance of lossy image compression. But it is also possible to show that other performance
parameters are also highly correlated. We have analyzed CR (compression ratio for ZIP highly
correlated with entropy) and parameters PSNR40, PSNR35, PSNR30, PSNRHVSM40,
PSNRHVSM35, PSNRHVSM30 where, e.g., PSNR40 means PSNR provided by AGU coder for
QS=40. It occurs that there is correlation close to unity (both Pearson and Spearman) for such pairs of
parameters as PSNR40 and PSNR35, PSNR30 and PSNR40, PSNR35 and PSNR30, PSNRHVSM40
and PSNRHVSM35, PSNRHVSM30 and PSNRHVSM40, PSNRHVSM35 and PSNRHVSM30. There
is also a rather high correlation (about 0.6) for such pairs of parameters as PSNR30 and PSNRHVSM30,
PSNR35 and PSNRHVSM35. This means that such parameters as, e.g., PSNR30 or PSNR35 are able
to characterize image complexity. Larger values of these parameters relate to simpler structure images.
Correlation of these parameters with CR is not high – SROCC is of the order 0.13.
    Note that there is no necessity to carry out compression and decompression of an image compressed
by AGU with QS=30 to determine PSNR30 and compare it to a threshold to classify this image. There
is a very fast method that allows predicting PSNR for a given QS by estimating MSE in a limited
number (e.g., 500) of 8x8 blocks for which direct DCT is carried out.
    All data and conclusions that are given above relate to images that are practically noise-free. If
images to be compressed are noisy, lossless compression is practically useless since it produces very
small values of CR whilst lossy compression has specific features that will be considered later.
    Two things are worth noting here. First, analysis of entropy or CR for ZIP can be useful in other
applications. For example, analysis of CR for electrocardiograms compressed by ZIP has allowed to
assess ECG quality. Analysis of entropy can indicate that images have been pre-compressed. For
example, Sentinel sites offer images that have been slightly compressed by JPEG2000. This leads to
entropy values that are slightly smaller than for uncompressed images. For example, for visualized
three-channel images in Fig. 7 the entropy values are equal to 4.15, 5.67, 4.46, and 5.72, respectively.
As one can see, entropy values are larger for the images SS2 and SS4 which are obviously more textural.
These images are compressed worse (CR for them is smaller than for the images SS1 and SS3 for the
same quality characterized by PSNR of PSNR-HVS-M) and they are classified worse [18] irrespectively
are the compression techniques based on DCT used or based on atomic functions employed [18].

3. Noise characteristic estimation, image filtering and lossy compression of
   noisy images
    In many practical situations, noise type and/or characteristics are unknown a priori and this makes
problems for image processing since many techniques exploit this information for providing high
efficiency (good examples are well known local statistic Lee and sigma filters that have versions for
additive and multiplicative noise). Thus, noise type and characteristics have to be estimated once for a
certain type of images with stable noise characteristics or each time if these characteristics can vary. If
the number of such images or its components (like in hyperspectral data) is large, it is difficult to carry
out such estimation “manually” (in interactive way) using special tools of image analysis. Then, one
has to apply automatic or blind methods for estimation of noise characteristics [9].
    Let us quickly recall what characteristics of the noise can be estimated. If the noise is white, these
are statistical characteristics like noise variance for additive noise, relative variance if noise is pure
multiplicative, and two or several parameters if noise is signal-dependent. If noise is not white, then
spatial spectral or correlation characteristics have to be estimated. The most complicated situation is,
probably, for the noise with non-stationary characteristics where variations of spatial and spectral
properties are possible.
    The main problem in blind estimation of noise characteristics (BENC) is to separate noise from
signal/noise (image/noise) mixture. Pre-filtering and further analysis of difference data is often
inefficient since filtering without exploiting noise characteristics is far away from perfect and, thus,
difference signal, in addition to the noise, contains errors due imperfect pre-filtering. Most existing
approaches try to detect (extract) image regions (blocks) that are, most probably, homogeneous and for
which it is possible to obtain local estimates of noise characteristics that are quite close to the true ones.
For this purpose, different methods and mathematical tools are used. Fractal theory can be employed
to discriminate texture informative and noise informative regions. Robust estimation and robust curve
fitting techniques are used to minimize the influence of abnormal (highly erroneous) local estimates of
noise characteristics obtained in heterogeneous (locally active) blocks falling into texture regions and/or
edge/detail neighborhoods [21].




                          a                                                      b




                        c                                                    d
Figure 7: Three-channel Sentinel-2 images SS1 (a), SS2 (b), SS3 (c), and SS4 (d)

   Here we do not want to go into theory and practice of BENC deeply. Our desire is to demonstrate
that image complexity and noise characteristic estimates are highly correlated. For this purpose, let us
consider the most typical case of zero mean additive white Gaussian noise (AWGN). Some data are
presented for the so-called minimal inter-quantile method of noise variance estimation in grayscale
images in [21]. It is shown that estimation bias and variance are sufficiently larger for the highly textural
image Baboon compared to the middle complexity images Barbara and Goldhill.
   A more complex case has been analyzed by us where it has been supposed that noise is independent
in color (RGB) components and has equal variances. A modification of the method [21] has been
applied. Instead of obtaining local estimates in blocks in spatial domain, estimates of noise standard
deviation have been obtained in 8x8 blocks using robust processing of discrete cosine transform
coefficients. An example of the obtained estimates of the noise standard deviation is presented in Fig.
8. Note that it is usually desired to get the standard deviation estimates that differ from the true values
by no more than 10%. Because of this, two horizontal lines show the desired margins of estimates.




Figure 8: Estimates of the noise standard deviation for R, G, and B color components for noise variance
equal to 65 for images of TID2008 (Fig. 1)

    As one can see, there are some estimates below the lower margin, e.g., for red and green components
of the test image # 6. These estimates are smaller than desired due to the noise saturation effects
observed for very intensive (almost white) and very dark (almost black) regions of images.
    To give a better explanation of basic principle of operation for many BENC methods, consider the
histograms of local estimate distribution for three cases. Fig. 9,a shows a typical distribution obtained
for the red component of the test image # 3 from TID2008 (Fig. 1) for the noise variance equal to 65.
The local estimates of the noise variance have been obtained in 5x5 blocks in spatial domain. As one
can see, there is a mode of local estimates (approximately equal to the true value) that correspond to
normal local estimates and a right hand heavy tail that relates to abnormal local estimates. Thus, the
task is to find the distribution mode quite accurately. However, there are cases when there are two
modes where one mode corresponds to erroneous local estimates (Fig. 9,b) that result from clipping. It
is also possible that there are very many abnormal local estimates that form their own mode that has no
relation to the true value of the noise variance.
    One observation is that the estimates of noise standard deviation are the largest for textural images
## 5, 8, 13, 18. This means that despite of the efforts to minimize the image content influence on noise
characteristic estimation it still exists.
     Fig. 10 shows the noise variance estimates for inter-quantile method operating in spatial domain.
Again most estimates are within the required limits (20% with respect to the true value). However, there
are two estimates (for the image # 20, green and red components) that are smaller than required due to
clipping effects and a few estimates that are larger than required due to high complexity of image
content.
    We have calculated SROCC between variance estimates and E (calculated for noise-free
components of color images). The SROCC values are equal to 0.53, 0.28, and 0.18 for red, green, and
blue components, respectively. Therefore, the estimates have a certain correlation with entropy.
    One problem is that the estimated SROCC is for variance estimates and entropy for noise-free
images where the latter ones are absent. Hence, we have also calculated E for noisy images and
determined SROCC for E and the estimates in Fig. 10. They are equal to 0.31, 0.07, and 0.03,
respectively. Thus, for noisy images their entropy cannot serve as a feature characterizing their
complexity and some other parameters (features) should be employed. (note that E values for noisy
images are from 6.25 for the image # 2 till 7.62 for the image # 25).
                                   200

                                   150

                                   100

                                    50

                                     0
                                         0   50       100     150      200     250      300
                                                              a
                                   300


                                   200


                                   100


                                    0
                                         0   50       100     150      200     250      300

                                                              b
                                    60


                                    40


                                    20


                                     0
                                         0    50      100     150      200     250     300

                                                  c
Figure 9: Examples of histograms of local estimates in 5x5 blocks for noise variance equal to 65 for the
standard case (red component, image # 3); image with clipping effect (green component, image # 20),
and highly textural image (green component, image # 13)

                          300
                                                                              Blue component
                                                                              Green component
                          250                                                 Red component



                          200
                  ^n 2




                          150



                          100



                          50



                           0
                                1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25
                                                    Test image index

Figure 10: Blind estimates of AWGN variance for interquantile method [21] operating in spatial domain
for components of color images from TID2008 for noise variance equal to 65

    Meanwhile, as an auxiliary information, the interquantile method [21] is able to produce an estimate
of percentage p of homogeneous image blocks. Such estimates for images in TID2008 corrupted by
AWGN with the same variance values in all components are represented in Fig. 11. As one can see, this
percentage is the largest for simple structure images ## 3, 15, and 23 that really contain quite large
quasi-homogeneous regions. On the contrary, the percentage estimates are the smallest for highly
textural images ## 5, 8, 18. For the image # 13 which is textural as well, the percentage estimate is
larger since the local estimate histogram has a very specific appearance (Fig. 10,c).




Figure 11: Estimates of percentage of homogeneous blocks in components of color images in TID2008
for noise variance equal to 65 in all components

     Let us now analyze how image complexity influences visual quality of noisy images and efficiency
of their denoising. It has been mentioned earlier that noise can be fully or partially masked by textures
and then image quality is perceived better than when noise is visually seen. Because of masking effects,
PSNR-HVS-M is larger than PSNR for a given image if the masking effect takes place. Fig. 12 shows
PSNR-HVS-M values (thick lines of the corresponding color) for component noisy images of TID2008
for the case of AWGN with variance equal to 65 (PSNR is about 30 dB). As one can see, all PSNR-
HVS-M values are larger than 30 dB where the largest values take place for the most textural (complex
structure) images ## 1, 5, 6, 13, 14, 25, 18. Meanwhile, the smallest values are observed for the simple
structure images ## 3, 15, 16, 20, and 23. The SROCC for PSNR-HVS-M and E calculated for noisy
images is equal to 0.55, 0.59, and 0.57, respectively. This shows that the correlation is quite high but
not large enough to accurately predict performance characteristics.

                          PSNR − HVS − M k (n), dB
                          41

                          40

                          39

                          38

                          37

                          36

                          35

                          34

                          33

                          32 0       5        10       15       20        25
                                                                               n
Figure 12: Values of PSNR-HVS-M for original (noisy) and filtered component images (TID2008), noise
variance is equal to 65
      Concerning image denoising, there are numerous existing methods. Our goal is not to study and
compare them, but to analyze general tendencies of denoising efficiency. In this sense, it is worth
considering potential and attained (practical) efficiency of image filtering. In the sense of potential
efficiency, very interesting results have been obtained by P. Milanfar et al in [22-25]. We have
determined potential efficiency for grayscale images corrupted by AWGN processed by non-local
filters. Some data have been got for component images of TID2008 (see data in Table 1 for AWGN
variance equal to 65). Potential (lower bound) efficiency is characterized by MSElb, the smaller the
better. As one can see, MSElb for different color components are practically the same. This is explained
by the fact that color components of the same image are highly correlated and, thus, are practically of
the same complexity. The smallest MSElb are observed for simpler structure images, e.g., images ## 10
and 23. In turn, the largest values are observed for the most complex structure images ## 13, 1, and 5.
It is interesting that MSElb is only 1.5 times smaller than noise variance, i.e. even potentially the noise
removal efficiency is very low.

Table 1
Lower bound MSE and output MSE for the DCT-based filter for components of color images in TID2008
 Image index        R component                G component                   B component
  in TID2008     MSElb      MSEDCT          MSElb            MSEDCT        MSElb        MSEDCT
        1        28.9         36.8          29.9              36.5         28.9          36.8
        5        27.5         30.4          28.0              30.9         26.4          30.8
        6        23.7         31.2          23.5              31.2         22.4          30.8
        8        23.2         32.5          23.3              32.1         22.6          32.0
       10         7.1         17.6           7.5              18.1          7.3          17.8
       11        16.4         26.4          16.2              26.3         15.3          25.8
       13        41.0         46.6          42.6              46.2         37.2          46.3
       14        20.7         31.0          20.0              30.6         19.5          30.7
       18        21.4         28.7          20.4              28.1         17.4          28.3
       23         4.6         12.9           4.7              12.7          4.8          12.6
       25        14.7         19.4          14.0              18.2         13.5          18.3

     Among possible methods of image denoising, consider the DCT-based filter. It is one of the best
transform-based filters which is only a little worse compared to the BM3D filter [15]. The MSE values
obtained for the DCT-based filter (denoted as MSEDCT) are given in Table 1. As one can see, MSEDCT
are almost the same as MSElb for complex structure images. Meanwhile, there are sufficient differences
between MSEDCT and MSElb (they differ by several times) for simple structure images.
     However, it does not mean that efficiency of denoising for simple structure images is low.
Comparison of PSNR-HVS-M before and after filtering (Fig. 12) shows that the largest differences in
visual quality take place just for simple structure images whilst filtering can be practically useless for
the complex structure images (see data for the image # 13). Thus, image complexity plays the key role
in image filtering efficiency. Additional materials concerning noise variance estimation and filtering
efficiency can be found in [26].
     Keeping this in mind, we have started to develop the direction of studies dealing with prediction of
image denoising efficiency and decision undertaking on expedience of filtering use for a given image.
It contains a thorough analysis of filtering efficiency for quite many test images, a set of modern filters
and several quality metrics. The outcomes of this study are very interesting. First, it is shown that
filtering efficiency can be predicted for many modern filters including transform-based, nonlocal and
other ones, i.e. filters based on different principles. Second, it is demonstrated that filtering efficiency
for many modern filters is of the same order and close to potential efficiency for texture images. Third,
improvement of many metrics due to filtering can be quite accurately predicted even if one or two input
parameters are used.
     Here it is worth recalling the main principle and assumptions put into basis of filtering efficiency
prediction. The following has been assumed: 1) there is a parameter that adequately characterizes
filtering efficiency (e.g., improvement of PSNR or PSNR-HVS-M due to filtering); 2) there exist one
or several parameters that adequately characterize image and noise properties; 3) there is a rather strict
dependence between input parameter(s) and a considered parameter characterizing filtering efficiency
where this dependence is established in advance (before filtering) and it is quite simple; 4) input
parameter(s) can be calculated easily and quickly (faster than image denoising) to make possible
decision on is it worth filtering a given image or it is reasonable to skip denoising and save time and
resources.
    It has been shown that all these assumptions are valid and efficient prediction is possible. Many
variants have been proposed. One of the main input parameters is Pkσ (or its variants for signal-
dependent noise) where k is a factor and σ is AWGN standard deviation supposed to be known in
advance. The parameter Pkσ is determined as a mean of local estimates Pkσ(n), n=1,…,N in N 8x8 pixel
blocks. A local estimate Pkσ(n) is calculated as Nkσ(n)/63 for the n-th 8x8 pixel block where Nkσ(n) is
the number of DCT coefficients that have absolute values smaller than kσ (the DCT coefficient that
corresponds to the block mean is excluded from analysis, k is usually set equal to 2 although other
variants are possible, e.g., P0.5σ can be analyzed).
    The parameter Pkσ jointly characterizes image complexity and noise intensity. Its small values (for
example, P0.5σ about 0.15) correspond to complex structure images and/or low intensity noise. On the
contrary, large values (for example, P0.5σ about 0.35) relate to simple structure images and/or high
intensity noise for which the filtering efficiency is usually high, i.e. large improvements of PSNR and
PSNR-HVS-M are observed. The difference in these improvements for images of different complexity
can be up to 10 dB.
    The use of more than one input parameter improves prediction making it more accurate. This can
be illustrated by the plot in Fig. 13 where points show the estimated parameters for images used in
analysis and the regression plain is presented (fitted). Note that it is not a problem to calculate noisy
image PSNR if noise variance is known a priori or pre-estimated with high accuracy whist σ has to be
known for calculation of P0.5σ as well. Also note that more complicated but more accurate prediction
approaches have been proposed recently based on trained neural networks [26]. They are applicable to
images corrupted by signal-dependent spatially correlated noise (speckle) typical for synthetic aperture
radar images produced by Sentinel-1. An advantage of this approach is that, based on prediction carried
out for the Lee filter with different scanning window sizes, the optimal size is recommended.
    One idea that follows from the results obtained in [26] is that image quality and complexity can be
assessed using neural network training approach.




Figure 13: Illustration of prediction of improvement of PSNR have two input parameters, P0.5σ and
PSNR for noisy image
     The last application considered here is lossy compression of noisy images. The known peculiarity
is the noise filtering effect [27]. One more peculiarity is that performance can be analyzed with respect
to noisy image (subject to compression) and noise-free (ideal) image. Dependences of traditional type,
e.g., when a metric is determined for the compressed image and image subject to compression, behave
as they are thought conventionally, i.e. the compressed image quality decreases if CR increases. Here
it is worth mentioning the recently obtained dependences for the BPG coder, an example is given in
Fig. 14,a. As one can see, the dependences are almost linear in the limits from Q≈15 till Q≈36 where
PSNR is approximately equal to input PSNR (about 26.7 dB for the considered case of AWGN with
σ=12) where Q is the parameter that controls compression in the BPG coder.
     Meanwhile, the dependences of PSNRtc calculated for the compressed noisy and noise-free (true)
images behave in a specific manner (Fig. 14,b). There can be maxima of these curves observed for such
Q that PSNRnc≈PSNRinp=10lg(2552/σ2). Then, if σ2 is known, it is easy to determine Q.




                                                   a




                                               b
Figure 14: Examples of dependences PSNRnc(Q) (a) and PSNRtc(Q) (b) for three test grayscale images
corrupted by AWGN with σ2=144
    The argument for which maximum takes place is the so-called optimal operation point (OOP). If it
exists (and this happens for simple structure images corrupted by rather intensive noise), it is worth
compressing a noisy image in OOP or its neighborhood (we say neighborhood because noise standard
deviation can be known or pre-estimated not accurately). An example of image compression in OOP is
given in Fig. 15. Noise filtering effect is seen in homogeneous image regions.
    In turn, if OOP does not exist, it is worth using a slightly smaller Q than for OOP (e.g., Q=33 for
the case considered in Fig. 14 for the test image Diego which is highly textural). Thus, again the image
processing parameters should be adapted to image complexity.




Figure 15: Original (noisy) image (left) with σ2=100 and this image compressed in OOP (right)

     The procedure of OOP existence for the BPG coder has not been developed and tested yet.
However, it has been developed and tested for the DCT-based coder AGU [25]. Not surprisingly, the
parameter P2σ again allows prediction (see the scatter-plot in Fig. 16). This time it allows predicting the
difference of a metric for potential OOP and for noisy image. The polynomial of the fifth order is fitted
very well and it shows that for P2σ>0.81 the OOP exists with high probability. Recall that this happens
if an image has a quite simple structure and/or noise is intensive.




Figure 16: The scatter-plot of difference PSNR vs P2σ for the AGU coder for images corrupted by AWGN

    We hope that the methodology for prediction OOP existence for the BPG coder will be developed
soon. Meanwhile, not only OOP existence can be predicted (it is supposed that OOP according to a
given metric exists if the predicted metric improvement is positive). The metric value itself can be
predicted as well.
    The presented materials show that for the case of noisy images not entropy but other parameters
relate to image complexity. An image can be considered simple if efficiency of its processing is high
(metric improvement due to filtering is large, OOP exists for lossy compression). Obviously, noise
properties play the role too.

4. Conclusions
    In this paper, we have tried to provide a systematic analysis of image complexity and its influence
on image processing performance. It has been shown that, for noise-free images, the entropy can be a
good characteristic of image complexity that strictly affects lossless compression and has a sufficient
impact on characteristics of lossy compression.
    If noise is present, then entropy becomes almost useless to characterize image complexity. Other
statistical parameters that jointly describe image properties and noise intensity can be used as indicators
of image complexity. The image complexity sufficiently influences potential and practical efficiency
of noise suppression in terms of conventional and visual quality metrics. It also has impact on existence
of optimal operation point in lossy compression of images corrupted by different types of noise for
various compression techniques.
    Meanwhile, the task of designing a unified metric for characterizing the image complexity is not
solved yet. Maybe, the use of neural networks able to unite different indicators in nonlinear manner can
help.

5. References
[1] E. Provenzi (Ed.), Special issue: Color Image Processing, 2018. doi: 10.3390/books978-3-03842-
     958-6.
[2] D. K. Pillai, New Computational Models for Image Remote Sensing and Big Data, in: P.
     Swarnalatha, P. Sevugan, (Eds.), Big Data Analytics for Satellite Image Processing and Remote
     Sensing, IGI Global, 2018, pp. 1–21.
[3] D. Shen, G. Wu, and H.-I. Suk, Deep learning in medical image analysis, Annual review of
     biomedical engineering 19 (2017), 221-248. doi: 10.1146/annurev-bioeng-071516-044442
[4] A. Pizurica, L. Platisa, T. Ruzic, B. Cornelis, A. Dooms, M. Martens, H. Dubois, B. Devolder, M.
     De Mey, I. Ingrid, Digital Image Processing of The Ghent Altarpiece: Supporting the painting's
     study and conservation treatment, IEEE Signal Processing Magazine 32 (2015) 112-122. doi:
     10.1109/MSP.2015.2411753.
[5] E.V. Bataeva, Y.A. Polyakova, Motive-analysis of TV-advertising visual content,
     Sotsiologicheskiy Zhurnal, 24 (2018) 66–89.
[6] N. Kussul, M. Lavreniuk, A. Shelestov, S. Skakun, Crop inventory at regional scale in Ukraine:
     Developing in season and end of season crop maps with multi-temporal optical and SAR satellite
     imagery, European Journal of Remote Sensing 51 (2018) 627-636. doi: 10.1080/22797254.
     2018.1454265.
[7] A.L. Mitchell, A. Rosenqvist, B. Mora, Current remote sensing approaches to monitoring forest
     degradation in support of countries measurement, reporting and verification (MRV) systems for
     REDD+, Carbon Balance Manage 12 (2017). doi: 10.1186/s13021-017-0078-9.
[8] K. Okarma, Current Trends and Advances in Image Quality Assessment, Elektronika Ir
     Elektrotechnika, 25 (2019) 77-84. doi: 10.5755/j01.eie.25.3.23681.
[9] Y. Zhang, G. Wang, J. Xu, Parameter estimation of signal-dependent random noise in CMOS/CCD
     image sensor based on numerical characteristic of mixed Poisson noise samples. Sensors 18
     (2018). doi: 10.3390/s18072276.
[10] I. Blanes, E. Magli, J. Serra-Sagrista, A Tutorial on Image Compression for Optical Space Imaging
     Systems, IEEE Geoscience and Remote Sensing Magazine 2 (2014) 8-26. doi:
     10.1109/MGRS.2014.2352465.
[11] M. Radosavljevic, B. Brkljac, P. Lugonja, V. Crnojevic, Z. Trpovski, Z. Xiong, D. Vukobratovic,
     Lossy Compression of Multispectral Satellite Images with Application to Crop Thematic Mapping:
     A HEVC Comparative Study, Remote Sens 12 (2020) 1590. doi: 10.3390 /rs12101590.
[12] I. Irum, M. Sharif, M. Raza, S. Mohsin, A Nonlinear Hybrid Filter for Salt & Pepper Noise
     Removal from Color Images, Journal of Applied Research and Technology 13 (2015) 79-85. doi:
     10.1016/S1665-6423(15)30015-8.
[13] M. S. Al-Ani, Khattab, M. A. Alheeti, Precision Statistical Analysis of Images Based on Brightness
     Distribution, Advances in Science, Technology and Engineering Systems Journal 2 (2017), 99-
     104. doi: 10.25046/aj020413.
[14] F. Li, V. Lukin, K. Okarma, Y. Fu, J. Duan, Intelligent lossy compression method of providing a
     desired visual quality for images of different complexity, in: Proceedings of 2021 International
     Conference on Applied Mathematics, Modeling and Computer Simulation, AMMCS 2021,
     Wuhan, China, 2021 500-505. doi: 10.3233/ATDE220050.
[15] Image and video denoising by sparse 3D transform-domain collaborative filtering. Block-matching
     and     3D     filtering     (BM3D)      algorithm    and      its    extensions,     2020.     URL:
     https://webpages.tuni.fi/foi/GCF-BM3D/.
[16] V. Makarichev, I. Vasilyeva, V. Lukin, B. Vozel, A. Shelestov, N. Kussul, Discrete Atomic
     Transform-Based Lossy Compression of Three-Channel Remote Sensing Images with Quality
     Control, Remote Sens. 14 (2022) 125. doi: 10.3390/rs14010125.
[17] S. Jiao, Y. Gao, J. Feng, T. Lei, X. Yuan, Does deep learning always outperform simple linear
     regression      in       optical    imaging?       2020.      URL:        https://arxiv.org/ftp/arxiv/
     papers/1911/1911.00353.pdf.
[18] Z. Chen, S. Gu, G. Lu, D. Xu, Exploiting Intra-slice and Inter-slice Redundancy for Learning-
     based Lossless Volumetric Image Compression, IEEE Transactions on Image Processing. 31
     (2022). doi: 10.1109/TIP.2022.3140608.
[19] K. Sayood, Introduction to data compression, 4th ed.; Morgan Kaufman, 2017.
[20] M. Behrisch et al., Quality metrics for information visualization, in Computer Graphics Forum, 37
     (2018) 625-662: Wiley Online Library.
[21] V. Abramova, A Blind Method for Additive Noise Variance Evaluation Based on Homogeneous
     Region Detection Using the Fourth Central Moment Analysis, Telecommunications and Radio
     engineering, 74 (2015) 1651-1669. doi: 10.1615/TelecomRadEng.v74.i18.50.
[22] P. Chatterjee, P. Milanfar, Is Denoising Dead?, IEEE Transactions on Image Processing 19 (2010)
     895-911. doi: 10.1109/TIP.2009.2037087.
[23] P Milanfar, A Tour of Modern Image Filtering: New Insights and Methods, Both Practical and
     Theoretical,     IEEE      Signal    Processing    Magazine        30    (2013)     106-128.      doi:
     10.1109/MSP.2011.2179329.
[24] Y. Romano, M. Elad, P. Milanfar, The little engine that could: Regularization by denoising, SIAM
     Journal on Imaging Sciences, 10 (2017) 1804-1844. doi: 10.1137/16M1102884.
[25] H. Talebi, X. Zhu, P. Milanfar, How to SAIF-ly Boost Denoising Performance, IEEE Transactions
     on Image Processing, 22 (2013) 1470-1485. doi: 10.1109/TIP.2012.2231691.
[26] O. Rubel, V. Lukin, A. Rubel, K. Egiazarian, Selection of Lee Filter Window Size Based on
     Despeckling Efficiency Prediction for Sentinel SAR Images, Remote Sens. 13 (2021) 1887. doi:
     10.3390/rs13101887.
[27] A. N. Zemliachenko, S. K. Abramov, V. V. Lukin, B. Vozel, K. Chehdi, Prediction of optimal
     operation point existence and parameters in lossy compression of noisy images, in: Proceedings
     SPIE 9244, Image and Signal Processing for Remote Sensing XX, 2014, 92440H. doi:
     10.1117/12.2065947.