<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Research of Methods for Image Sharpness Evaluation in Photos of People</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Victoria</forename><surname>Vysotska</surname></persName>
							<email>victoria.a.vysotska@lpnu.ua</email>
							<affiliation key="aff0">
								<orgName type="institution">Lviv Polytechnic National University</orgName>
								<address>
									<addrLine>Stepan Bandera</addrLine>
									<postCode>79013</postCode>
									<settlement>Lviv</settlement>
									<country key="UA">Ukraine</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Nataliia</forename><surname>Sharonova</surname></persName>
							<email>nvsharonova@ukr.net</email>
							<affiliation key="aff1">
								<orgName type="institution">National Technical University &quot;KhPI&quot;</orgName>
								<address>
									<addrLine>Kyrpychova str. 2</addrLine>
									<postCode>61002</postCode>
									<settlement>Kharkiv</settlement>
									<country key="UA">Ukraine</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Mariya</forename><surname>Shirokopetleva</surname></persName>
							<email>marija.shirokopetleva@nure.ua</email>
							<affiliation key="aff2">
								<orgName type="institution">Kharkiv National University of Radio Electronics</orgName>
								<address>
									<addrLine>Nauky Ave. 14</addrLine>
									<postCode>61166</postCode>
									<settlement>Kharkiv</settlement>
									<country key="UA">Ukraine</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Oleksandr</forename><surname>Dolhanenko</surname></persName>
							<email>oleksandr.dolhanenko@nure.ua</email>
							<affiliation key="aff2">
								<orgName type="institution">Kharkiv National University of Radio Electronics</orgName>
								<address>
									<addrLine>Nauky Ave. 14</addrLine>
									<postCode>61166</postCode>
									<settlement>Kharkiv</settlement>
									<country key="UA">Ukraine</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Anastasiya</forename><surname>Chupryna</surname></persName>
							<email>anastasiya.chupryna@nure.ua</email>
							<affiliation key="aff2">
								<orgName type="institution">Kharkiv National University of Radio Electronics</orgName>
								<address>
									<addrLine>Nauky Ave. 14</addrLine>
									<postCode>61166</postCode>
									<settlement>Kharkiv</settlement>
									<country key="UA">Ukraine</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Serhii</forename><surname>Smelyakov</surname></persName>
							<email>serhii.smeliakov@nure.ua</email>
							<affiliation key="aff2">
								<orgName type="institution">Kharkiv National University of Radio Electronics</orgName>
								<address>
									<addrLine>Nauky Ave. 14</addrLine>
									<postCode>61166</postCode>
									<settlement>Kharkiv</settlement>
									<country key="UA">Ukraine</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Research of Methods for Image Sharpness Evaluation in Photos of People</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">6E103C93206BC9F3058FB6DA83F32D9D</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T19:23+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Sharpness, image, aperture, depth of field, focus, FFT, variance of the Laplacian 1 0000-0001-6417-3689 (V. Vysotska) 0000-0002-8161-552X (N. Sharonova) 0000-0002-7472-6045 (M. Shirokopetleva)</term>
					<term>0000-0002-3996-6940 (O. Dolganenko)</term>
					<term>0000-0003-0394-9900 (A. Chupryna)</term>
					<term>0000-0002-5791-2479 (S. Smelyakov)</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>The subject matter of the article is image sharpness evaluation in photos of people. The goal of the work is to analyse the existing methods of image sharpness evaluation, compare their performance and quality of results, suggest improvements for the use case of sharpness classification of photos of people, where large quantities of background blur is present due to the aperture effect. In this article the methods for image sharpness evaluation were described and tested on a set of selected images. The images contained different subject sizes and types of blur. The following methods were used: Fast Fourier Transform (FFT), Variance of the Laplacian, Appearance-based face detection algorithms, metadata analysis, linear trendline analysis. As a result, the problem of naturally blurred background was demonstrated and conclusions were made. An alternative method of sharpness evaluation was de-scribed, which solves the mentioned problem. The suggested improved algorithm was tested to determine if it satisfies expectations and solves the identified problems. To implement the Fast Fourier Transform and Variance of the Laplacian methods, the OpenCV library was used. The following results were ob-tained -when using the default implementation, the FFT and Variance of the Laplacian methods are not reliable for evaluating sharpness for images containing large and unstable quantities of naturally blurred background (due to the open aperture) and when using different settings of the lens and camera. The following conclusions were made: steps need to be taken to eliminate the factors of unstable quantities of naturally blurred background and camera preferences and in this way improve accuracy, reliability of sharpness evaluation. This means evaluating the sharpness of parts of the images and the images as a whole. These steps include but are not limited to face position detection, identifying faces that were supposed to be in focus when the photo was taken, sharpness evaluation of only areas that were in-tended to be in focus. Plans were set for further research and improvements of the suggested algorithm.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>Photography is very subjective. There are many ways to take a photo, since all of us see the world in very different ways. There are, however, some basic principles for achieving a generally well composed photograph. You may or may not follow these principles and still get a great photo that tells a story, but statistically speaking the best looking and most objectively recognised as professional and pleasing photos are the ones which have the subject in critical sharp focus.</p><p>Achieving focus on a photo is a very complicated process, which has multiple factors influencing the final result. A general smartphone user takes this process for granted, as the phone does all decision making and photo processing for us in real time, most of the time achieving pretty good results for general social media use. The algorithms used there are fine-tuned, subject oriented. Most of the time our phones use machine learning to detect the scene and tune the settings to achieve better results.</p><p>Our smartphones are getting better and better every year in terms of photography, there is even the ability to imitate expensive lenses by adding a fake depth-of-field to the photos <ref type="bibr" target="#b0">[1]</ref>. However, any well-established professional photographer will say how important it is to be in control of the manual settings of the camera, being the decision maker and scene establisher, getting the exact result the creator wants, and not the tool on its own. Moreover, they use professional hardware, which may contain some "smart" features, but generally is very exposed to manual overriding, giving the operator more flexibility. However, where humans are involved mistakes are present.</p><p>Having photographed an event, say, a wedding, the photographer usually spends hours and sometimes days looking at more than 2000 photos, filtering out the ones that are to be deleted and highlight the best ones for further editing. This is a very long and complicated process of comparing, what seem to be, identical photographs at first glance (but actually different in slight ways), which can be a big factor for the final result. The most important thing the photographer looks at is the sharpness of the subjects. It is a general rule, that, when photographing humans or animals, the eyes are the ones that need to be in critical sharp focus. Everything else in the photo can be changed -the lighting can be increased or decreased, some elements can be added or removed, but sharpening a blurry subject is a very destructive process which is highly discouraged.</p><p>The goal of this work is to analyse and compare different algorithms and program solutions for identifying the sharpness of a photo (containing humans as subjects) automatically and proposing improvements of the existing algorithms to solve the problems which will be described further on.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Related Works</head><p>When it comes down to solving a scientific problem, it is required to operate with objective terms and calculations. That is why it is important to define what a sharp photo is. The common way of measuring the sharpness is by the "rise distance" of edges within the image. That way, the sharpness is determined by the distance of a pixel level between 10% and 90%. Therefore, it can be stated that sharpness is measured by analysing the intensity of edge gradients <ref type="bibr" target="#b1">[2]</ref>. However, the threshold level of intensity to classify a photo being sharp or not is very circumstantial and cannot be standardised. During the following experiments, a set of reference photos will be used to determine the minimum and maximum values of sharpness and thus, providing the threshold. Apart from that, the problem of image quality assessment <ref type="bibr" target="#b2">[3]</ref> is well-known and was attempted do be generally solved by many.</p><p>In order to start detecting and classifying the sharpness of photos it is first needed to clarify which camera and lens settings influence it. The camera in combination with the lens has 3 main settings: aperture, shutter speed and ISO.</p><p>Aperture is the main setting of the lens. It is generally an opening that can be bigger or smaller and thus let in more or less light. It is usually preferred to let in as much light as possible, therefore "opening up", or "making the aperture wide open", but this has a side effect that can directly influence the sharpness of the subject. The more open the aperture -the more shallow the depth of field is. For cameras that can only focus on one object distance at a time, depth of field is the distance between the nearest and the furthest objects that are in acceptably sharp focus.</p><p>By knowing the DOF (Depth Of Field) we can understand what depth of the image had to be in focus. To give a more clear understanding of why this is important an example image is provided, shot with a very shallow DOF (see Figure <ref type="figure" target="#fig_0">1</ref>).</p><p>Figure <ref type="figure" target="#fig_0">1</ref> (photo by Dolhanenko O.) demonstrates a shot with a very open aperture of 2.8 (the lower the number the more open the aperture is). The subject that lays within the shallow depth of field is in critical sharp focus, however the secondary subjects which are before and after the field boundaries are not in focus at all (in this case only before the subject on the right).</p><p>By means of traditional photo sharpness detection algorithms, this is not a sharp photo, as the area in critical sharp focus is very small compared to the blurry part (this will be experimentally tested further on). However, if the subject is correctly identified among the two and is exclusively checked, then the photo is in fact in critical sharp focus and is acceptable for further editing. Another setting that influences the sharpness of a photo is the shutter speed. The rule is "the faster -the better" for general photography. The quicker the curtain collapses before the sensorthe less blur there will be on the photo, as each movement of subjects in the photo during the shot will cause their sharpness to decrease. Especially when shooting in low light conditions, when fast shutter speeds are not available (otherwise the photo will be too dark) this motion blur is quite noticeable due to minor movements.</p><p>The pursuit of accurate image sharpness assessment has led to the exploration of various approaches, encompassing spatial domain-based methods, spectral domain-based methods, learning-based methods <ref type="bibr" target="#b3">[4]</ref>, and a combination of these techniques, each presenting unique advantages and challenges, including utilization of methods such as Local Phase Coherence <ref type="bibr" target="#b4">[5,</ref><ref type="bibr" target="#b5">6]</ref>, Edge Information analysis approach <ref type="bibr" target="#b6">[7]</ref>, Normal-Gradient-Based approach <ref type="bibr" target="#b7">[8]</ref>, Gradient Neighbourhood-Weighted approach <ref type="bibr" target="#b8">[9]</ref> and others.</p><p>Zhu et al. (2023) <ref type="bibr" target="#b9">[10]</ref> conducted a comprehensive review, offering insight into the current trends and performance comparisons of notable algorithms, revealing a landscape of ongoing innovation aimed at overcoming the shortcomings of existing methods.</p><p>Bielievtsov et al. (2018) <ref type="bibr" target="#b10">[11]</ref> investigated network technology for the transmission of visual information, highlighting the importance of maintaining image quality in the context of digital communication and storage. This work is foundational, setting the stage for further research into image quality assessment methods that are critical for various applications, including but not limited to, facial recognition, social media, and professional photography.</p><p>In the domain of image search and retrieval, Smelyakov et al. (2020) <ref type="bibr" target="#b11">[12]</ref> introduced an innovative approach to image engine search for big data warehouses, emphasizing the necessity of high-quality image processing for efficient and accurate image retrieval. This development is particularly relevant to our research as it underscores the significance of image sharpness in enhancing the performance of search engines, which often rely on visual content analysis to function effectively.</p><p>Furthermore, the effectiveness of preprocessing algorithms for natural language processing applications was explored by Smelyakov et al. (2020) <ref type="bibr" target="#b12">[13]</ref>, illustrating the broad applications of image and signal processing techniques across various fields of computer science. Although focused on natural language processing, the principles of preprocessing and quality enhancement are applicable to the domain of image processing, providing insights into methods that could potentially improve image sharpness evaluation algorithms.</p><p>The development of no-reference (NR) <ref type="bibr" target="#b13">[14]</ref> sharpness metrics has been particularly noteworthy, with <ref type="bibr" target="#b14">Duan et al. (2021)</ref>  <ref type="bibr" target="#b14">[15]</ref> introducing an efficient NR objective sharpness assessment metric designed for images with shallow depth of field, a common characteristic in portraits and photos emphasizing human subjects. This metric, which calculates sharpness based on bidirectional pixel intensity differences, addresses the limitations of traditional sharpness assessment tools when applied to such images.</p><p>Research by <ref type="bibr" target="#b15">Her and Yang (2019)</ref>  <ref type="bibr" target="#b15">[16]</ref> on image sharpness assessment algorithms for autofocus systems further exemplifies the field's evolution. Their work evaluates the performance of several spatial domain functions, highlighting the scene adaptability and anti-jamming capabilities of the Benner algorithm and the sensitivity of the Laplace algorithm, among others. This research underscores the critical role of sharpness evaluation functions in enhancing the quality of images captured by various imaging systems.</p><p>The development of advanced artificial intelligence systems, as explored by Kyrychenko, Tereshchenko, Proniuk, and Geseleva (2023) <ref type="bibr" target="#b16">[17]</ref>, through the use of predicate clustering methods, presents potential avenues for refining image sharpness evaluation techniques Moreover, advancements in image quality assessment for zoom photos, as investigated by <ref type="bibr" target="#b17">Han et al. (2023)</ref>  <ref type="bibr" target="#b17">[18]</ref>, reveal the challenges posed by small sensor sizes and fixed focal lengths in smartphones. Their novel no-reference zoom quality metric incorporates traditional sharpness estimation with image naturalness concepts, demonstrating significant improvements in assessing image quality over traditional metrics.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Methods</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1.">Existing methods</head><p>The Fast Fourier Transform is a convenient mathematical algorithm for computing the Discrete Fourier Transform (DFT). It is used for converting a signal from one domain into another <ref type="bibr" target="#b18">[19]</ref>.</p><p>The FFT is used in different areas, such as mathematics, music, engineering, etc. This method is widely used, as sometimes calculations are much easier performed when the time-series signals are converted into the frequency domain. This method can also be used to convert the frequency domain back to the original format. When talking about FFT in image processing and computer vision, it is important to note that the image is represented in the Fourier and Spatial domains. So, the image is represented in both imaginary and real components.</p><p>The obtained values can be analysed to perform blurring or blur detection, edge detection, analysis of textures, etc.</p><p>There is a sampled Fourier Transform, which is called DFT. It contains only the set of image samples which are enough to fully represent the spatial domain image <ref type="bibr" target="#b19">[20]</ref> (which is often used for further quality metrics extraction).</p><p>Given an image with size N*N, the resulting DFT matrix can be defined as follows:</p><formula xml:id="formula_0">𝐸𝐹(𝑘, 𝑙) = ∑ ∑ 𝑓(𝑖, 𝑗)𝑒 !"#$% !" # &amp; $" # ' (!) *+, (!) -+,<label>(1)</label></formula><p>where f(a,b) is the image in the spatial domain and the exponential term is the basis function corresponding to each; point F(k,l) in the Fourier space.</p><p>The equation can be interpreted as: the value of each point F(k,l) is obtained by multiplying the spatial image with the corresponding base function and summing the result. At the base level functions, operations are represented as sine and cosine waves which have increasing frequencies. For example, F(0,0) represents the DC-component of the image (average image brightness) and F(N-1,N-1) is the highest frequency of the image.</p><p>The ordinary one-dimensional DFT has N 2 complexity. If the Fast Fourier Transform (FFT) is used, the complexity can be reduced to Nlog2N. For computing large images this improvement is crucial. However, some forms of the FFT may restrict the maximum size of the input to N=2 n .</p><p>The result is an output image represented with complex numbers . This image can be displayed in two states: either with the real and imaginary part or with magnitude and phase (see Fig. <ref type="figure">2</ref>).</p><p>When solving problems in the area of image processing, commonly only the magnitude of the Fourier Transform is displayed. The example image of the result of such transformation is illustrated above.</p><p>For clear and reliable results, a contrast detection threshold must be calculated beforehand <ref type="bibr" target="#b20">[21]</ref> in order to understand if the image has enough contrast for further analysis. In case of contrast availability and FFT algorithm completion, a floating point value of the mean of the magnitude indicates the relative sharpness of the whole image. Of course, since this value is relative, conclusions cannot be made without a reference sharpness value.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Figure 2: The result of calculating and displaying the Fourier Transform magnitude of an input image</head><p>Another option is to use the variance of the Laplacian. The Laplacian of a function f at a point p is (up to a factor) the rate at which the average value of f over spheres cantered at p deviates from f(p) as the radius of the sphere shrinks towards 0.</p><p>The Laplacian operator is defined as the divergence of the gradient of function f, as shown in formula <ref type="bibr" target="#b1">(2)</ref>.</p><p>∆𝑓(𝑥, 𝑦) = 𝑑𝑖𝑣3𝑔𝑟𝑎𝑑(𝑓)7</p><p>(2) In this definition, the gradient is the slope of steepest accent and it gives information about the point and direction of the highest accent in local maxima and, likewise, the local minima.</p><p>In case of image sharpness detection, divergence is the vector field associated with blurriness from subject motion or natural background blurriness. The calculated matrix of Laplacian is demonstrated on Figure <ref type="figure" target="#fig_1">3</ref>. Using the process of convulsion with the source image, it is being transformed based on the Laplacian kernel. This is used to find areas of rapid source changes (but this works if there is no noise in the image <ref type="bibr" target="#b21">[22]</ref>).</p><p>Using the OpenCV library this method will be tested on sample images which contain both blurry areas and sharp subjects. This partial blurriness was caused by a very open aperture, which made a very distinct background separation.</p><p>var imageMat = Highgui.imread(image.absolutePath, Highgui.CV_LOAD_IMAGE_GRAYSCALE) val destination = Mat() Imgproc.Laplacian(imageMat, destination) val median = MatOfDouble() val std = MatOfDouble() Core.meanStdDev(destination, median, std) val variance = Math.pow(std.get(0, 0).get(0), 2.0)</p><p>To get a single floating point number representing the overall sharpness of a photo it is first needed to load the source image in grayscale, after which the basic Laplacian function should be called.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.">The alternative algorithm</head><p>Looking at the results of the two popular algorithms it can be stated that neither one is optimised enough for the problem stated at the beginning. During photo sessions most portrait photos will be taken with an open aperture which will result in the background having many areas with low frequencies <ref type="bibr" target="#b22">[23]</ref>. If the area with naturally low frequencies is greater than the area with high frequencies it will result in the image being labelled as "not sharp", which is not necessarily true. The question arises -how to make judgements on the photo subject sharpness if analysing the whole image is not optimal?</p><p>As a top level solution, the sharpness detection algorithm should be modified in such a way, so that only areas that are supposed to be in focus are evaluated and the background/foreground is ignored. This is the general principle of evaluating visible errors in specific areas of the photo <ref type="bibr" target="#b23">[24]</ref>. An example of such approach can be found in the framework for measuring sharpness in natural images <ref type="bibr" target="#b24">[25]</ref>. Also, there is research about image quality assessment based on regions of interest, which are identified as features which are highly spatially nonstationary <ref type="bibr" target="#b25">[26]</ref>.</p><p>The solution will be developed based on the limitation that the subjects that need to be in focus are human faces. This limitation, however, can be eliminated by modifying the algorithm and providing support for more subjects.</p><p>The following list describes the steps of the algorithm:</p><p>1. Find the coordinates and boxing boundaries of every face in the frame 2. Extract the focus distance from the photo 3. Calculate the distance to every face in the frame 4. Calculate the ideal depth of field 5. Select the faces that are within the intended focus plane (focus distance +-ideal depth of field) 6. Apply the sharpness detector only for the boxes containing the selected faces 7. Calculate the average sharpness score based on the individual sharpness scores Next, the steps of the algorithm will be described in more detail with the propositions and variants for implementation.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.1.">Finding the boundaries and coordinates of faces in the frame</head><p>There are quite a few methods and options when it comes to face detection in images. Given an arbitrary image, the goal of face detection is to determine whether or not there are any faces in the image and, if present, return the image location and extent of each face. Some methods for detecting faces include:</p><p>• Knowledge-Based Top-Down Methods;</p><p>• Feature invariant approaches;</p><p>• Template matching methods;</p><p>• Appearance-based methods. Knowledge-based methods are developed with the scientific knowledge about human faces as the primary source of information. The problem with this approach is the difficulty in translating human knowledge into well-defined rules. It is very hard to find a perfect balance of strictness in these rules. If the rules are defined too strictly -the method may fail to detect faces that do not pass all rules at once. On the other hand, if the rules are too general -the method may result in false positives.</p><p>The feature-based methods are based on the fact that humans have the natural ability to recognize a face in different lighting conditions, under different angles and circumstances, meaning we are trained to detect certain features. This method builds on this fact and many variants were proposed, when first the features are extracted and analysed. One problem with these feature-based algorithms is that the image features can be severely corrupted due to illumination, noise, and occlusion.</p><p>In template matching methods a face pattern is manually predefined. When analyzing an input image, the correlation is computed for separate parts of the presumable face: the contour, the nose, eyes, mouth. The result is positive, if the mean correlation value of these components is above certain threshold. This method, however, is not ideal due to the lack of flexibility when it comes to different face shapes, poses and scale.</p><p>Lastly, the appearance-based methods are more close-to-life, than the previous ones. The templates are not generated by experts (like in template matching), rather taken from samples of actual image databases. This method relies heavily on statistical analysis and machine learning.</p><p>Appearance-based methods have many different implementations, starting like distributionbased, support vector machines, hidden Markov model, cascade classifiers, cascaded convolutional networks <ref type="bibr" target="#b26">[27]</ref>, etc.</p><p>Taking into account the advantages and disadvantages of the abovementioned face detection methods, it was decided that appearance-based methods are well suited for the task. When implementing the improved sharpness detection algorithm, cascade classifiers can be used. OpenCV contains pre-trained open classifiers that can be freely downloaded. To retrieve face coordinates using the OpenCV library one can use the CascadeClassifier.detectMultiScale function. The arguments for this function are: the input image (in gray scale), scaleFactor and minNeighbours. The scaleFactor specifies how much the image size is reduced with each scale. minNeighbours specifies how many neighbors each candidate rectangle should have to retain it. These parameters will be fine-tuned during the development process.</p><p>Of course, everything that involves image processing and subject recognition will always take up some valuable processing time, so there is another way of extracting the faces in the frame. The other method relies heavily on the camera's integrated ability to detect faces in real time and write this information to the metadata. Of course, the algorithm should not be strictly reliant on using this optional metadata, but definitely should utilize it if available -that would dramatically improve the performance of the method.</p><p>As a result of image scanning with classifiers or face extraction using metadata, the coordinates and boundaries of all faces in the frame will be retrieved for further analysis.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.2.">Extracting the focus distance and other useful parameters from the photo</head><p>Focusing plane is the image sensor of the camera. Focusing distance is the distance from the focusing plane to the subject.</p><p>When using automatic lenses (lenses with autofocus, that do not require manual input to put the subject in critical sharp focus) there are complex algorithms running in order to produce optimal results in terms of focus. The camera constantly monitors the image, understands the distance to the subject in real time and adjusts the focus motors accordingly. To achieve the autofocus, the camera lens moves to a position where the clearest image is obtained. The maximum clarity is measured from the histograms of the images on which a filter with the role of highlighting the edges was initially applied. Knowing the focal length of the lens, the distances can be found from the lens equation <ref type="bibr" target="#b27">[28]</ref>. So, we can see that this is clearly possible and cameras perform these calculations all the time. Moreover, cameras can utilize different focusing algorithms <ref type="bibr" target="#b28">[29]</ref> for accuracy. The distance that the focus motor travelled in the lens before the photo was taken corresponds to the estimated focusing distance which is needed for the modified sharpness detection algorithm.</p><p>It is possible to retrieve the calculated focusing distance by looking at the photo metadata. EXIF (Exchangeable Image File Format) is a standard that allows adding information (metadata) to photos and videos. This format is quite flexible, meaning that all users can modify it by adding new data entries with original names. This means that not all camera bodies from different manufacturers produce the same metadata. Having that said, many cameras contain valuable information that can be used for the purpose of this research. By viewing the metadata using an EXIF reader we can find the following relevant information (Table <ref type="table" target="#tab_0">1</ref>). The parameter "Focus Distance 2" implies the calculated distance (in meters) to the focused subject. This calculation was performed by the camera by analysing the rotation angle of the focus motor when the edges of the image were in focus. This parameter can be used as a ready solution and will be very useful for further calculations. There is no other reliable way of retrieving/calculating the focus distance of a photo, especially when the photo is not guaranteed to be in critical focus initially. This is why the method heavily relies on this meta parameter. Some other useful parameters include the number of faces detected and the detected faces positions. This is a "luxury" to have these parameters in the metadata and only modern cameras provide such information. Heavily relying on the presence of these parameters would limit the method to work only for modern high-tech camera bodies. So, the method will utilise these shortcuts in metadata if they are present, but still have the ability to detect faces' boundaries described in the section above.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.3.">Calculating the distance to every face in the frame</head><p>It is required to find distances to all faces in the frame because the method needs to determine which faces in particular need to be in focus. The "need to be in focus" criteria is described in the next point.</p><p>At this step the focus distance, as well as all face boundaries (and thus, sizes) are already known. To find the distance to the human face on the photo it is necessary to first estimate the height of the faces in the real life.</p><p>To find the real-world height of each face in the frame, two methods can be used:</p><p>• the average statistical face height value (21.8cm -23.9 cm);</p><p>• calculated from a proportionality formula based on the distance between the eyes and the mouth. The second option is very complicated, involves extra operations like edge detection to find the location of eyes and mouth, depends on the pose (will not work for side portraits) and does not guarantee accuracy greatly more than the first option. This is why the first option of assuming that the height of the face is somewhere between 21.8cm and 23.9cm will be used during the implementation.</p><p>When the real world height of all faces in the frame is known, the real world distance to the faces can be determined in two ways:</p><p>• predefined proportions at 1meter distance method (see Figure <ref type="figure">4</ref>, photo by Dolhanenko O.); • using an alleged "focused face" from the focus point coordinates as a reference. The first method can be described as follows: practically or mathematically find the percentage height (from the frame height) of a 23.9 cm high object at a distance of 1 meter at a focal length of 50mm (which gives 1.0 times magnification) on a 35mm full frame sensor.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Figure 4:</head><p>The result of measuring a 23.9 cm high object at a distance of 1 meter with specific settings As seen from the photo, the height of the face on the image is 1857px (46.4%) on a horizontal photo with 4000*6000 resolution. On a vertical photo of the same resolution, the facial height is 30.95% of the full height. This gives a point of reference which can be proportionally scaled based on the given (non-full frame) sensor size or/and focal length.</p><p>To convert the reference measurement to a different focal length on a different format camera sensor, the following formula can be used (see formula 3):</p><formula xml:id="formula_1">ℎ = 𝐻 * 𝑓 50 ,<label>(3)</label></formula><p>where h is the new reference height of the face, f is the 35mm equivalent focal length of the selected lens, H is the initial reference height of the face (taken via 50mm full frame lens from 1 meter).</p><p>When the reference height of the 23.9 cm face is calculated taking into account the sensor format and the focal length, the distance D to a particular face in the frame can be calculated by comparing the height of the face to the reference height and determining proportionally the distance to the sensor (see formula 4).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>𝐷 = ℎ 𝑔 (4)</head><p>where e D is the distance (in meters) to the face in the frame, g is the height of the face of interest in pixels, h is the reference height of the face.</p><p>The method described above is simplified. In future work the accuracy can be improved by adding support for lens distortions and other factors that may influence the measurements.</p><p>The second method heavily relies on metadata. The coordinates of the focus point in the meta data indicates the precise area of focus interest on the photo. Since the coordinates and boxes of all faces are available at this point, the face within the area of interest can be selected as the reference. Since the distance to this face is already available, once again, from the metadata, we can calculate the proportional distance to other faces in the frame by using the (4) formula.</p><p>The second method can be reliable if the image does not have stacked, slightly shifted faces, but since this is not a guarantee, the first method can be preferred. However, if the conditions are ideal, the second method will produce more precise results, since it uses more non-estimated measurements. There are other methods of object distance extraction by using reference targets <ref type="bibr" target="#b29">[30]</ref>, however, such methods are not suitable for general photography.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.4.">Calculating the ideal depth of field</head><p>Depth of field (DOF) is the distance between the nearest and the farthest objects that are in acceptably sharp focus in an image.</p><p>To understand whether or not some subject at a specific distance is supposed to be sharp in the photo (based on the camera and lens settings) it is required to first calculate the depth of field using the captured camera and lens settings. The formulas for calculating the distance to the front plane of the focus area and the distance to the back plane of the focus area are as shown in formula <ref type="bibr" target="#b4">(5)</ref>:</p><p>𝑅 * 𝑓 # 𝑓 # − 𝐾 * 𝑓 * 𝑧 + 𝐾 * 𝑅 * 𝑧 ;</p><formula xml:id="formula_2">𝑅 # = 𝑅 * 𝑓 # 𝑓 # + 𝐾 * 𝑓 * 𝑧 − 𝐾 * 𝑅 * 𝑧 ,,<label>(5)</label></formula><p>where R1 is distance to the front edge of the critical focus plane, R is distance of focus (can be retrieved from metadata), R2 is distance to the back edge of the critical focus plane, f is the focal length of the lens in meters (can be retrieved from metadata), K is the f-stop of the lens (can be retrieved from metadata), Z is the Circle Of Confusion (can be retrieved from metadata).</p><p>By subtracting R1 from R2 the depth of field can be found. However, it will be easier to use the raw R1 and R2 values for further calculations.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.5.">Applying sharpness detection only for faces required to be in focus</head><p>To understand which faces to check for sharpness it is first required to understand the criteria of expected portrait framing depth (EPFD).</p><p>The EPFD is a generalized assumption of the maximum intended distance between the subjects when taking photos of multiple rows of people at once. For example, a photographer is taking a group portrait photo with 10 people, which are placed in two rows. In this case, the expected portrait framing depth is within 0.8 meters, since there are two rows and each one is about 0.3 meters wide +-a margin in between. For such portrait framing depth, the photographer needs to set a particular aperture (f-stop), so that the depth of field is equal to or more than the framing depth.</p><p>The EPFD is a very subjective parameter and cannot be calculated from analyzing the photo. That is because it is infinitely hard for a computer to determine if the photo of a group of people is intended to be shot with an open aperture (to have only front subjects in focus), or it was intended to have both rows of people in focus and the low aperture value was chosen by mistake.</p><p>The subjective nature of this parameter leaves this part of the method to be fine-tuned by the end user, selecting one of two options on a collection of photos before the sharpness detection: single/couple styled photos or intended multi-row group photos.</p><p>For the second option, where a group photo (composed in two or more rows) is selected, the faces can be classified as "intended to be sharp" by the following sequence of formulas:</p><p>Given the measured light value (EV) from the metadata, it is required to calculate the maximum aperture, using which the EV value would be the same. The bigger the aperture value, the greater the depth of field is, meaning more faces need to be in critical sharp focus. The photographer could make the mistake of setting a low aperture, which results in only one row of people being in focus. This is why the maximum aperture value needs to be determined and the photo needs to be analysed as if these "ideal" settings were set in camera.</p><formula xml:id="formula_3">𝐸𝑉 = log # I 100 * 𝐾 # 𝐼 * 𝑆 M ,,<label>(6)</label></formula><p>where EV is the exposure value; K is the f-stop value, I is the ISO. Given the maximum acceptable ISO value by the photographer (the greater the ISO the more noise there is in the photo, thus it is down to personal photographer preference), the maximum value of the aperture to achieve the same light value can be determined as follows:</p><formula xml:id="formula_4">𝐾𝑚𝑎𝑥 = O 2 ./ * 𝐼𝑚𝑎𝑥 * 𝑆𝑚𝑎𝑥 100 (<label>7</label></formula><formula xml:id="formula_5">)</formula><p>where Kmax is the maximum f-stop number for the pre-defined EV and ISO, Imax is the maximum ISO value, acceptable by the photographer, Smax is the maximum duration of the shutter speed.</p><p>When operating a camera, the photographer may or may not use automatic settings. This is why in the formula above the shutter speed is not taken from the metadata, rather calculated as well. The maximum acceptable shutter speed can be calculated as follows:</p><formula xml:id="formula_6">𝑆𝑚𝑎𝑥 = 1 𝐹𝑙<label>(8)</label></formula><p>where Fl is the focal length used for the photo, taken from the metadata. However, there is an unformal rule, stating, that when shooting portraits, one should not select shutter speeds below 0.008 of a second. This may also be an external fine-tuning setting available for the end user.</p><p>Next, the maximum depth of field needs to be calculated. To achieve this, the formulas (5) can be used, inserting the maximum aperture f-stop value (Kmax). As a result, R1 (distance to the front edge of the critical focus plane) and R2 (distance to the back edge of the critical focus plane) with the "ideal" camera settings can be found.</p><p>Given the calculated distance to the subjects face, the criteria of in focus intention can be formulated as follows: the face of a subject was intended to be in critical sharp focus, if the distance D satisfies the condition:</p><formula xml:id="formula_7">𝐷 ≤ 𝑅2<label>(9)</label></formula></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.6.">Sharpness detection only for regions containing the selected faces</head><p>At this stage of the algorithm, the most valuable information is already achieved -the understanding of which faces in the frame were most likely intended to be in focus. Having this information, the sharpness detection algorithm of choice can be applied these regions exclusively. This will result in very accurate results, as the background and other subjects will not be evaluated.</p><p>It is worth mentioning, that having this information, a whole window of possibilities for photo categorizing opens up. Not only the sharpness can be evaluated, but also the open state of the eyes (open/half-open/closed), preferred facial expressions of focused subjects, even the poses can be classified as appealing and not.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Experiment</head><p>The purpose of the following experiment is to demonstrate relations of the Laplacian and FFT sharpness evaluation algorithms to different factors of the photo. Revealing these dependencies on factors will lead to conclusions about the steps needed to be taken in order to improve the effectiveness of the algorithms.</p><p>In this experiment the following set of images will be used: 1.</p><p>A reference representation of a completely blurry photo 2.</p><p>A reference representation of a very sharp photo in all parts (no background blurriness) 3.</p><p>A photos which contain sharp and blurred parts due to the aperture effect The first and second images are used to determine the dramatic maximum and minimum sharpness results both algorithms provide.</p><p>All other 7 test photos contain human subjects and should be the same resolution and have identical lighting conditions for optimal evaluation results. There should not be any colour or sharpness corrections done before the experiment. All photos from the experiment are similar to figure <ref type="figure" target="#fig_0">1</ref> and contain same subjects and same scenes but in different configurations.</p><p>The following Table <ref type="table" target="#tab_1">2</ref> verbally describes the photos that were used for the experiment.</p><p>The first photo was made with a very closed aperture (f22) and has extremely "busy" foreground, which leads to extreme sharpness results by the Laplacian method. The second photo is fully out-of-focus and results in very low value results by both algorithms. No conclusions can be made from these results yet. The following Table <ref type="table" target="#tab_2">3</ref> represents the results of the main 7 photos analysis (provided in the Laplacian, FFT columns), alongside additional calculated parameters that will be needed for dependencies analysis. It should be stated, that the Laplacian and FFT methods produce values that do not relate to each other and thus should not be compared directly. The "background quantity" is the percentage value of how much of the photo is taken up by the background. The "total subject size" from photo is the opposite value, describing how much area of a photo is taken up by the subjects.</p><p>The next phase of the experiment includes utilizing the parts of the improved algorithm that was described in the previous section. Within this experiment the dependencies which lead to unstable results were determined. One of the main parts of the described algorithm is face detection. The following test was conducted on a set of images that were previously cut to contain only faces. Images that contained two subjects are represented by two individual cut images.</p><p>The experiment conditions and results are displayed in table <ref type="table" target="#tab_2">3</ref>. The test images are extracted faces from the frame (see Figure <ref type="figure" target="#fig_2">5</ref>, photos by O. Dolhanenko) and the test results are stated in Table <ref type="table" target="#tab_3">4</ref>. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Results</head><p>By plotting the sharpness result values obtained during the experiment against the background quantity, the following chart can be achieved (see Figure <ref type="figure" target="#fig_3">6</ref>) As seen from the chart, a clear linear trendline describes the relation of the quantity of the background to the calculated sharpness (both in Laplacian and FFT methods). This means that the more background there is visible on a photo (even if the subject is in critical sharp focus), the less likely the photo will be classified as sharp.</p><p>By plotting the sharpness result values obtained during the experiment against the f-stop values (see Figure <ref type="figure" target="#fig_4">7</ref>), the following chart can be achieved. The relation on this chart is not as obvious as the previous one, since the data set is rather small, but even here a visible linear trendline for both Laplacian and DFT methods is noticeable. This trend implies that the higher the f-stop value, the more likely the photo will be classified as sharp.</p><p>Having analyzed both relations it can be stated, that the result of DFT and Laplacian methods are dependent on the scene and camera settings, which means that they are not reliable for generic automatic photo filtering. Furthermore, comparing the average Laplacian sharpness value for subject photos (45.12) to the average sharpness value for the 2 reference photos (325.54) reveals a major difference. This difference demonstrates the source of errors that can appear by selecting an incorrect reference value. The "Image(s) name" is the name of the analyzed photo. If the photo contained multiple subjects, it was split into separate images and visually merged in the table.</p><p>The "Expectation" is a non-bias subjective rating of image quality and usability given by the photographer, where 0 is unusable (blurry) and 1 -usable (sharp). The "Laplacian", "FFT" are actual results of sharpness evaluation. The "Laplacian normalized" and "FFT normalized" are conversions of actual results to a 0-1 scale, where 1 is the maximum value taken from the algorithm output. This way the normalized values of different algorithms are relatable to the "Expectation" values. The "time" columns represent the time taken to process the image in milliseconds.</p><p>After the data was collected, the normalized columns were analyzed for visual trends. To fulfil the first objective of the experiment, the sorted input values trend (the expectation) must meet the trend of the sorted normalized output values. By viewing the plotted results (see Figure <ref type="figure" target="#fig_5">8</ref>) it can be stated, that both algorithms generally fulfil this requirement.</p><p>Only one data point is either unnaturally classified as sharp (by both algorithms), or the subjective usability parameter was defined not accurately.</p><p>The second objective of the experiment was to determine which algorithm is more stable and should be used in the actual implementation. Comparing two charts (see Figure <ref type="figure" target="#fig_5">8</ref>) a more linear trend line is received from the FFT algorithm, rather than the Laplacian. Moreover, the first method produces very unbalanced results. The third and final objective was to determine which algorithm works faster. By comparing the average speed of both algorithms -17.7ms for Laplacian and 241.25ms for the FFT, it can be concluded that the FFT method is 14 times slower.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.">Discussions</head><p>Two methods were analyzed and compared -FFT (Fast Fourier Transform) and the Variance of Laplacian for image sharpness evaluation. During the first experiment it was determined that both algorithms produce results that are highly dependent on the photo scene, composition and camera settings -photos with large quantities of background (produced by a low f-stop value) were classified as blurry.</p><p>An improved algorithm was proposed which is based on subject detection <ref type="bibr" target="#b30">[31,</ref><ref type="bibr" target="#b31">32]</ref> and further sharpness evaluation exclusively of the subject box. This way all dependencies were eliminated and more natural evaluation results were obtained.</p><p>The improved algorithm uses metadata of the photo to simplify calculations, if the relevant data "shortcuts" are available. In some cases, the metadata may contain the detected faces coordinates, which greatly optimize the performance of the algorithm. The method is built on the principle of data analysis and does not require much user input to function properly. The only parameter that cannot be achieved automatically is the intended photography style -whether the user was shooting groups of people to get all in focus, or it was intended to focus on only one-two in the frame, while isolating the rest of the background. Another parameter that the user may input is the maximum acceptable ISO value. This parameter has a default value but is very subjective and so depends on the user preference.</p><p>The goals of the experiment were to identify whether the improved algorithm results meet the expectation and to determine which of the sharpness evaluation methods work best for the task. As a result it was found that the improved algorithm meets the expectations and works best with the FFT sharpness evaluation method as it produces more linear trending results. The Laplacian method, however, is 14 times faster and may be preferred for very large datasets, where accuracy is not as important as time efficiency.</p><p>The results of the research were used to implement a software solution prototype for automated image files sorting by subject sharpness (see Figure <ref type="figure" target="#fig_6">9</ref>).</p><p>The resulting software was configured to operate with JPG files, built using primarily Kotlin and OpenCV, deployed and run on a MacOS environment. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="7.">Conclusions</head><p>The purpose of this research was to analyze the existing methods of image sharpness evaluation and suggest improvements for the use case of sharpness evaluation of photos of people. The goal end result is an automated photo sorting solution, which classifies images from sharp and usable to blurry and unusable.</p><p>The comparison of FFT and Variance of Laplacian methods for image sharpness evaluation showed that their accuracy is affected by the photo's scene, composition, and camera settings, leading to misclassification of detailed background photos as blurry.</p><p>An enhanced algorithm leveraging subject detection for focused sharpness evaluation showed significant improvements by removing biases and yielding more accurate results. It utilizes photo metadata to streamline processes, especially when such metadata includes coordinates of detected faces, enhancing performance efficiency. The method minimizes user input, requiring only the photography style and an optional maximum acceptable ISO value to adjust for personal preference. Final testing confirmed the algorithm's effectiveness, particularly in conjunction with the FFT method for its linear results, though the faster Laplacian method may be chosen for large datasets where speed trumps precision.</p><p>The results of the research and experiments have been used to implement a prototype of an automation system for sorting photos by subject sharpness, detecting unwanted motion within the image, subject blurriness.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: Example of a photo with a shallow DOF</figDesc><graphic coords="3,218.73,149.38,159.02,101.00" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figure 3 :</head><label>3</label><figDesc>Figure 3: The Laplacian kernel for convulsion with the source image</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>Figure 5 :</head><label>5</label><figDesc>Figure 5: Some examples of the test images used for the experiment</figDesc><graphic coords="13,156.45,72.00,297.62,63.10" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_3"><head>Figure 6 :</head><label>6</label><figDesc>Figure 6: Results of both algorithms plotted against the background quantity</figDesc><graphic coords="13,175.25,268.07,245.75,191.45" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_4"><head>Figure 7 :</head><label>7</label><figDesc>Figure 7: Results of both algorithms plotted against the background quantity</figDesc><graphic coords="13,185.52,565.82,225.44,189.65" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_5"><head>Figure 8 :</head><label>8</label><figDesc>Figure 8: The result of the second experiment with sharpness expectations</figDesc><graphic coords="15,62.97,72.00,470.55,199.00" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_6"><head>Figure 9 :</head><label>9</label><figDesc>Figure 9: A software solution for automatic sorting based on subject sharpness</figDesc><graphic coords="16,86.20,72.00,452.50,156.30" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1 Partial, most relevant information extracted from the test photo metadata</head><label>1</label><figDesc></figDesc><table><row><cell>Parameter Name</cell><cell>Value</cell></row><row><cell>Lens Spec</cell><cell>FE 70-200mm F4 G OSS</cell></row><row><cell>Min Focal Length</cell><cell>70.0 mm</cell></row><row><cell>Max Focal Length</cell><cell>200.0 mm</cell></row><row><cell>Focal Length</cell><cell>159.0 mm (35 mm equivalent: 159.0 mm)</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 2 Description of the test images</head><label>2</label><figDesc></figDesc><table><row><cell>Image</cell><cell>Verbal description</cell></row><row><cell>name</cell><cell></cell></row><row><cell>1.JPG</cell><cell>A sharp photo containing one in-focus subject which takes up 24.7% of the frame</cell></row><row><cell>2.JPG</cell><cell>A sharp photo containing one in-focus subject which takes up 37.5% of the frame</cell></row><row><cell>3.JPG</cell><cell>A sharp photo containing two in-focus subjects, high aperture (background blurriness is</cell></row><row><cell></cell><cell>high), collectively taking up 38.38% of the frame.</cell></row><row><cell>4.JPG</cell><cell>A sharp photo containing one in-focus and one out-of-focus subject, collectively taking</cell></row><row><cell></cell><cell>up 36.09% of the frame</cell></row><row><cell>5.JPG</cell><cell>A sharp photo containing two in-focus subjects, low aperture (background blurriness is</cell></row><row><cell></cell><cell>low), taking up 33.13% of the frame</cell></row><row><cell>6.JPG</cell><cell>A sharp photo containing one in-focus and one in-motion subject, collectively taking up</cell></row><row><cell></cell><cell>37.87% of the frame</cell></row><row><cell>7.JPG</cell><cell>A sharp photo containing one in-focus and one out-of-focus subject, collectively taking</cell></row><row><cell></cell><cell>up 57.24% of the frame</cell></row><row><cell>8.JPG</cell><cell>A very "busy" and sharp photo, shot with a closed aperture (f22)</cell></row><row><cell>9.JPG</cell><cell>A completely blurry, out-of-focus photo of the same scene as in 8.JPG</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>Table 3 Parameters and results of the experiment with the main photos</head><label>3</label><figDesc></figDesc><table><row><cell cols="3">Image Name Laplacian FFT</cell><cell>Background</cell><cell>Total subjects size</cell><cell>ff-stop value</cell></row><row><cell></cell><cell></cell><cell></cell><cell>quantity (%)</cell><cell>from photo (%)</cell><cell></cell></row><row><cell>1.JPG</cell><cell>33.84</cell><cell>20.89</cell><cell>75.22</cell><cell>24.78</cell><cell>4</cell></row><row><cell>2.JPG</cell><cell>77.03</cell><cell>13.39</cell><cell>62.475</cell><cell>37.525</cell><cell>4</cell></row><row><cell>3.JPG</cell><cell>60.63</cell><cell>40.95</cell><cell>61.62</cell><cell>38.38</cell><cell>20</cell></row><row><cell>4.JPG</cell><cell>23.97</cell><cell>25.53</cell><cell>63.91</cell><cell>36.09</cell><cell>4</cell></row><row><cell>5.JPG</cell><cell>36.76</cell><cell>35.02</cell><cell>66.87</cell><cell>33.13</cell><cell>4</cell></row><row><cell>6.JPG</cell><cell>44.06</cell><cell>32.12</cell><cell>62.129</cell><cell>37.871</cell><cell>4</cell></row><row><cell>7.JPG</cell><cell>39.57</cell><cell>21.24</cell><cell>42.76</cell><cell>57.24</cell><cell>2.8</cell></row><row><cell>8.JPG</cell><cell>647.4</cell><cell>64.066</cell><cell>7.1</cell><cell>92.9</cell><cell>22</cell></row><row><cell>9.JPG</cell><cell>3.71</cell><cell>-6.80</cell><cell>100</cell><cell>0</cell><cell>2.8</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_3"><head>Table 4 Results of the experiment for face-only evaluation</head><label>4</label><figDesc></figDesc><table><row><cell>Image(s) name</cell><cell cols="2">Expectatio n (1-sharp, 0 -blurry) Laplacian</cell><cell>Laplacian avg</cell><cell>Laplacian normalize-ed</cell><cell>Laplaci an time (ms)</cell><cell>FFT</cell><cell>FFT avg</cell><cell>FFT normalized</cell><cell>FFT time (ms)</cell></row><row><cell cols="2">10-2.JPG 0 10.JPG</cell><cell>7.1056 39.4608</cell><cell cols="2">23.2832 0.0898</cell><cell>7 10</cell><cell cols="2">-4.2154 7.1930 18.6013</cell><cell>0.2517</cell><cell>132 155</cell></row><row><cell>4-2.JPG 4.JPG</cell><cell>0.2</cell><cell>7.0557 149.0316</cell><cell cols="2">78.0437 0.3010</cell><cell>12 10</cell><cell>2.1674 22.1968</cell><cell cols="2">12.1821 0.4263</cell><cell>233 96</cell></row><row><cell>3-2.JPG 3.JPG</cell><cell>0.3</cell><cell cols="3">45.4098 92.2698 0.3559 139.1298</cell><cell>5 13</cell><cell cols="3">13.8346 18.2867 0.6400 22.7388</cell><cell>75 201</cell></row><row><cell>7-2.JPG 7.JPG</cell><cell>0.4</cell><cell>6.5819 137.9015</cell><cell cols="2">72.2417 0.2786</cell><cell>18 40</cell><cell>9.8695 26.3808</cell><cell cols="2">18.1252 0.6343</cell><cell>353 589</cell></row><row><cell>6-2.JPG 6.JPG</cell><cell>0.4</cell><cell cols="3">373.5180 203.0986 0.7834 32.6791</cell><cell>11 7</cell><cell cols="3">26.3782 19.9107 0.6968 13.4432</cell><cell>104 102</cell></row><row><cell>5-2.JPG 5.JPG</cell><cell>0.7</cell><cell cols="3">80.4655 213.9522 0.8252 347.4389</cell><cell>6 8</cell><cell cols="3">17.0262 21.8001 0.7630 26.5740</cell><cell>50 133</cell></row><row><cell>1.JPG</cell><cell>0.8</cell><cell cols="3">221.8970 221.8970 0.8559</cell><cell>52</cell><cell cols="3">26.2262 26.2262 0.9179</cell><cell>277</cell></row><row><cell>2.JPG</cell><cell>1</cell><cell cols="3">259.2589 259.2589 1.0000</cell><cell>49</cell><cell cols="3">28.5732 28.5732 1.0000</cell><cell>879</cell></row></table></figure>
		</body>
		<back>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Synthetic depth-of-field with a single-camera mobile phone</title>
		<author>
			<persName><forename type="first">N</forename><surname>Wadhwa</surname></persName>
		</author>
		<idno type="DOI">10.1145/3197517.3201329</idno>
	</analytic>
	<monogr>
		<title level="j">ACM Transactions on Graphics</title>
		<imprint>
			<biblScope unit="volume">37</biblScope>
			<biblScope unit="issue">4</biblScope>
			<biblScope unit="page">64</biblScope>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">No-reference image blur assessment using multiscale gradient</title>
		<author>
			<persName><forename type="first">M.-J</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">C</forename><surname>Bovik</surname></persName>
		</author>
		<idno type="DOI">10.1186/1687-5281-2011-3</idno>
	</analytic>
	<monogr>
		<title level="j">J Image Video Proc</title>
		<imprint>
			<biblScope unit="volume">3</biblScope>
			<date type="published" when="2011">2011. 2011</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Image information and visual quality</title>
		<author>
			<persName><forename type="first">H</forename><forename type="middle">R</forename><surname>Sheikh</surname></persName>
		</author>
		<idno type="DOI">10.1109/tip.2005.859378</idno>
	</analytic>
	<monogr>
		<title level="j">IEEE Transactions on Image Processing</title>
		<imprint>
			<biblScope unit="volume">15</biblScope>
			<biblScope unit="issue">2</biblScope>
			<biblScope unit="page" from="430" to="444" />
			<date type="published" when="2006">2006</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Blind Image Quality Assessment Using a Deep Bilinear Convolutional Neural Network</title>
		<author>
			<persName><forename type="first">W</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Ma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Yan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Deng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Wang</surname></persName>
		</author>
		<idno type="DOI">10.1109/TCSVT.2018.2886771</idno>
	</analytic>
	<monogr>
		<title level="j">IEEE Transactions on Circuits and Systems for Video Technology</title>
		<imprint>
			<biblScope unit="volume">30</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page" from="36" to="47" />
			<date type="published" when="2020">2020. 2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Image Sharpness Assessment Based On local Phase Coherence</title>
		<author>
			<persName><forename type="first">S</forename><surname>Chattopadhyay</surname></persName>
		</author>
		<idno>doi: 0.48047/nq.2022.20.5.nq22800</idno>
	</analytic>
	<monogr>
		<title level="j">NeuroQuantology</title>
		<imprint>
			<biblScope unit="volume">5</biblScope>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Image Sharpness Assessment Based on Local Phase Coherence</title>
		<author>
			<persName><forename type="first">R</forename><surname>Hassen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">M A</forename><surname>Salama</surname></persName>
		</author>
		<idno type="DOI">10.1109/TIP.2013.2251643</idno>
	</analytic>
	<monogr>
		<title level="j">IEEE Transactions on Image Processing</title>
		<imprint>
			<biblScope unit="volume">22</biblScope>
			<biblScope unit="issue">7</biblScope>
			<biblScope unit="page" from="2798" to="2810" />
			<date type="published" when="2013">2013</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">An Improved Method for Evaluating Image Sharpness Based on Edge Information</title>
		<author>
			<persName><forename type="first">Z</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Hong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Gan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Chen</surname></persName>
		</author>
		<idno type="DOI">10.3390/app12136712</idno>
	</analytic>
	<monogr>
		<title level="j">Applied Sciences</title>
		<imprint>
			<biblScope unit="volume">12</biblScope>
			<biblScope unit="page">6712</biblScope>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Image sharpness evaluation method based on normal gradient feature</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Zhang</surname></persName>
		</author>
		<idno type="DOI">10.1109/ISRIMT53730.2021.9596808</idno>
	</analytic>
	<monogr>
		<title level="m">Proceeding of the 3rd International Symposium on Robotics &amp; Intelligent Manufacturing Technology (ISRIMT)</title>
				<meeting>eeding of the 3rd International Symposium on Robotics &amp; Intelligent Manufacturing Technology (ISRIMT)<address><addrLine>Changzhou, China</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2021">2021</date>
			<biblScope unit="page" from="308" to="314" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Multidirectional gradient neighborhood-weighted image sharpness evaluation algorithm</title>
		<author>
			<persName><forename type="first">X</forename><forename type="middle">Y</forename><surname>Yan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Lei</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Zhao</surname></persName>
		</author>
		<idno type="DOI">10.1155/2020/7864024</idno>
	</analytic>
	<monogr>
		<title level="j">Math. Problems Engin</title>
		<imprint>
			<biblScope unit="volume">7864024</biblScope>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Review: A Survey on Objective Evaluation of Image Sharpness</title>
		<author>
			<persName><forename type="first">M</forename><surname>Zhu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Yu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Ke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Zhi</surname></persName>
		</author>
		<idno type="DOI">10.3390/app13042652</idno>
	</analytic>
	<monogr>
		<title level="j">Applied Sciences</title>
		<imprint>
			<biblScope unit="volume">13</biblScope>
			<biblScope unit="issue">4</biblScope>
			<biblScope unit="page">2652</biblScope>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Network technology for transmission of visual information</title>
		<author>
			<persName><forename type="first">S</forename><surname>Bielievtsov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Ruban</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Smelyakov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Sumtsov</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Selected Papers of the XVIII International Scientific and Practical Conference &quot;Information Technologies and Security&quot; (ITS 2018)</title>
				<meeting><address><addrLine>Kyiv, Ukraine</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2018-11-27">November 27, 2018. 2018</date>
			<biblScope unit="volume">2318</biblScope>
			<biblScope unit="page" from="160" to="175" />
		</imprint>
	</monogr>
	<note>CEUR Workshop Proceedings</note>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Search by Image Engine for Big Data Warehouse</title>
		<author>
			<persName><forename type="first">K</forename><surname>Smelyakov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Chupryna</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Sandrkin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Kolisnyk</surname></persName>
		</author>
		<idno type="DOI">10.1109/eStream50540.2020.9108782</idno>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2020 IEEE Open Conference of Electrical, Electronic and Information Sciences (eStream)</title>
				<meeting>the 2020 IEEE Open Conference of Electrical, Electronic and Information Sciences (eStream)<address><addrLine>Vilnius, Lithuania</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="1" to="4" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Effectiveness of Preprocessing Algorithms for Natural Language Processing Applications</title>
		<author>
			<persName><forename type="first">K</forename><surname>Smelyakov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Karachevtsev</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Kulemza</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Samoilenko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Patlan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Chupryna</surname></persName>
		</author>
		<idno type="DOI">10.1109/PICST51311.2020.9467919</idno>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2020 IEEE International Conference on Problems of Infocommunications. Science and Technology (PIC S&amp;T)</title>
				<meeting>the 2020 IEEE International Conference on Problems of Infocommunications. Science and Technology (PIC S&amp;T)<address><addrLine>Kharkiv, Ukraine</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="187" to="191" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">No-reference image sharpness assessment via difference quotients</title>
		<author>
			<persName><forename type="first">Jiye</forename><surname>Qian</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Hengjun</forename><surname>Zhao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Jin</forename><surname>Fu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Wei</forename><surname>Song</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><surname>Jide Qian</surname></persName>
		</author>
		<author>
			<persName><surname>Xiao</surname></persName>
		</author>
		<idno type="DOI">10.1117/1.JEI.28.1.013032</idno>
	</analytic>
	<monogr>
		<title level="j">Journal of Electronic Imaging</title>
		<imprint>
			<biblScope unit="volume">28</biblScope>
			<biblScope unit="issue">1</biblScope>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">An Effective Sharpness Assessment Method For Shallow Depth-Of-Field Images</title>
		<author>
			<persName><forename type="first">Z</forename><surname>Duan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Fan</surname></persName>
		</author>
		<idno type="DOI">10.1109/ICIP42928.2021.9506498</idno>
	</analytic>
	<monogr>
		<title level="m">Proceeding of the IEEE International Conference on Image Processing (ICIP)&apos;21</title>
				<meeting>eeding of the IEEE International Conference on Image essing (ICIP)&apos;21<address><addrLine>Anchorage, AK, USA</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2021">2021</date>
			<biblScope unit="page" from="1449" to="1453" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">Research of Image Sharpness Assessment Algorithm for Autofocus</title>
		<author>
			<persName><forename type="first">L</forename><surname>Her</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Yang</surname></persName>
		</author>
		<idno type="DOI">10.1109/ICIVC47709.2019.8980980</idno>
	</analytic>
	<monogr>
		<title level="m">Proceeding of the 2019 IEEE 4th International Conference on Image, Vision and Computing (ICIVC)</title>
				<meeting>eeding of the 2019 IEEE 4th International Conference on Image, Vision and Computing (ICIVC)<address><addrLine>Xiamen, China</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2019">2019</date>
			<biblScope unit="page" from="93" to="98" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">Predicate Clustering Method and its Application in the System of Artificial Intelligence</title>
		<author>
			<persName><forename type="first">I</forename><surname>Kyrychenko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Tereshchenko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Proniuk</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Geseleva</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 7th International Conference on Computational Linguistics and Intelligent Systems</title>
				<meeting>the 7th International Conference on Computational Linguistics and Intelligent Systems<address><addrLine>COLINS-</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2023">2023. 2023</date>
			<biblScope unit="page" from="395" to="406" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">Image Quality Assessment for Realistic Zoom Photos</title>
		<author>
			<persName><forename type="first">Z</forename><surname>Han</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Xie</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Zhai</surname></persName>
		</author>
		<idno type="DOI">10.3390/s23104724</idno>
	</analytic>
	<monogr>
		<title level="j">Sensors</title>
		<imprint>
			<biblScope unit="volume">2023</biblScope>
			<biblScope unit="page">4724</biblScope>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<analytic>
		<title level="a" type="main">Automatic detection of blurred images in UAV image sets</title>
		<author>
			<persName><forename type="first">T</forename><surname>Sieberth</surname></persName>
		</author>
		<idno type="DOI">10.1016/j.isprsjprs.2016.09.010</idno>
	</analytic>
	<monogr>
		<title level="j">ISPRS Journal of Photogrammetry and Remote Sensing</title>
		<imprint>
			<biblScope unit="volume">122</biblScope>
			<date type="published" when="2016">2016</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<analytic>
		<title level="a" type="main">Bridging the Gap Between Imaging Performance and Image Quality Measures</title>
		<author>
			<persName><forename type="first">E</forename><surname>Fry</surname></persName>
		</author>
		<idno type="DOI">10.2352/issn.2470-1173.2018.12.iqsp-231</idno>
	</analytic>
	<monogr>
		<title level="j">Electronic Imaging</title>
		<imprint>
			<biblScope unit="volume">12</biblScope>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<analytic>
		<title level="a" type="main">Contrast sensitivity in images of natural</title>
		<author>
			<persName><forename type="first">S</forename><surname>Triantaphillidou</surname></persName>
		</author>
		<idno type="DOI">10.1016/j.image.2019.03.002</idno>
	</analytic>
	<monogr>
		<title level="j">Signal Processing: Image Communication</title>
		<imprint>
			<biblScope unit="volume">75</biblScope>
			<biblScope unit="page">67</biblScope>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b21">
	<analytic>
		<title level="a" type="main">Edge Detection Techniques for Quantifying Spatial Imaging System Performance and Image Quality</title>
		<author>
			<persName><forename type="first">O</forename><surname>Van Zwanenberg</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Triantaphillidou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Jenkin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Psarrou</surname></persName>
		</author>
		<idno type="DOI">10.1109/cvprw.2019.00238</idno>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)</title>
				<meeting>the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)<address><addrLine>Long Beach, CA, USA</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2019">2019</date>
			<biblScope unit="page" from="1871" to="1879" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b22">
	<analytic>
		<title level="a" type="main">Perceptual blur and ringing metrics: application to JPEG2000</title>
		<author>
			<persName><forename type="first">P</forename><surname>Marziliano</surname></persName>
		</author>
		<idno type="DOI">10.1016/j.image.2003.08.003</idno>
	</analytic>
	<monogr>
		<title level="j">Signal Processing: Image Communication</title>
		<imprint>
			<biblScope unit="volume">19</biblScope>
			<biblScope unit="issue">2</biblScope>
			<biblScope unit="page">167</biblScope>
			<date type="published" when="2004">2004</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b23">
	<analytic>
		<title level="a" type="main">Image Quality Assessment: From Error Visibility to Structural</title>
		<author>
			<persName><forename type="first">Z</forename><surname>Wang</surname></persName>
		</author>
		<idno type="DOI">10.1109/tip.2003.819861</idno>
	</analytic>
	<monogr>
		<title level="j">IEEE Transactions on Image Processing</title>
		<imprint>
			<biblScope unit="volume">13</biblScope>
			<biblScope unit="issue">4</biblScope>
			<biblScope unit="page">601</biblScope>
			<date type="published" when="2004">2004</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b24">
	<analytic>
		<title level="a" type="main">A framework for measuring sharpness in natural images captured by digital cameras based on reference image and local areas</title>
		<author>
			<persName><forename type="first">M</forename><surname>Nuutinen</surname></persName>
		</author>
		<idno type="DOI">10.1186/1687-5281-2012-8</idno>
	</analytic>
	<monogr>
		<title level="j">EURASIP Journal on Image and Video Processing</title>
		<imprint>
			<date type="published" when="2012">2012</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b25">
	<analytic>
		<title level="a" type="main">Method of image quality assessment based on region of interest</title>
		<author>
			<persName><forename type="first">W</forename><surname>Yang</surname></persName>
		</author>
		<idno type="DOI">10.3724/sp.j.1087.2008.01310</idno>
	</analytic>
	<monogr>
		<title level="j">Journal of Computer Applications</title>
		<imprint>
			<biblScope unit="volume">28</biblScope>
			<biblScope unit="issue">5</biblScope>
			<biblScope unit="page" from="1310" to="1312" />
			<date type="published" when="2008">2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b26">
	<analytic>
		<title level="a" type="main">Face Detection Method Based on Cascaded Convolutional Networks</title>
		<author>
			<persName><forename type="first">R</forename><surname>Qi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R. -S</forename><surname>Jia</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q. -C</forename><surname>Mao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H. -M</forename><surname>Sun</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L. -Q</forename><surname>Zuo</surname></persName>
		</author>
		<idno type="DOI">10.1109/access.2019.2934563</idno>
	</analytic>
	<monogr>
		<title level="j">IEEE Access</title>
		<imprint>
			<biblScope unit="volume">7</biblScope>
			<biblScope unit="page" from="110740" to="110748" />
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b27">
	<analytic>
		<title level="a" type="main">Video distance measurement based on focus</title>
		<author>
			<persName><forename type="first">H</forename><surname>Eugen</surname></persName>
		</author>
		<idno type="DOI">10.1109/epe50722.2020.9305593</idno>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2020 International Conference and Exposition on Electrical And Power Engineering (EPE)</title>
				<meeting>the 2020 International Conference and Exposition on Electrical And Power Engineering (EPE)<address><addrLine>Iasi, Romania</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page">65</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b28">
	<analytic>
		<title level="a" type="main">Analysis of focus measure operators for shape-from-focus</title>
		<author>
			<persName><forename type="first">S</forename><surname>Pertuz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Puig</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">A</forename><surname>Garcia</surname></persName>
		</author>
		<idno type="DOI">10.1016/j.patcog.2012.11.011</idno>
	</analytic>
	<monogr>
		<title level="j">Pattern Recognition</title>
		<imprint>
			<biblScope unit="volume">46</biblScope>
			<biblScope unit="issue">5</biblScope>
			<biblScope unit="page" from="1415" to="1432" />
			<date type="published" when="2013">2013</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b29">
	<analytic>
		<title level="a" type="main">A Method for Distance Measurement of Moving Objects in a Monocular Image</title>
		<author>
			<persName><forename type="first">Z</forename><surname>Xu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Wang</surname></persName>
		</author>
		<idno type="DOI">10.1109/SIPROCESS.2018.8600495</idno>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 3rd International Conference on Signal and Image Processing (ICSIP)</title>
				<meeting>the 3rd International Conference on Signal and Image Processing (ICSIP)<address><addrLine>Shenzhen, China</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2018">2018</date>
			<biblScope unit="page" from="245" to="249" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b30">
	<analytic>
		<title level="a" type="main">Local Feature Extraction in Images</title>
		<author>
			<persName><forename type="first">A</forename><surname>Sergiyenko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Serhiienko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Orlova</surname></persName>
		</author>
		<idno type="DOI">10.20535/2708-4930.2.2021.244191</idno>
	</analytic>
	<monogr>
		<title level="j">Information, Computing and Intelligent systems</title>
		<imprint>
			<biblScope unit="volume">2</biblScope>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b31">
	<analytic>
		<title level="a" type="main">Improved covariant local feature detector</title>
		<author>
			<persName><forename type="first">Z</forename><surname>Huo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Zhang</surname></persName>
		</author>
		<idno type="DOI">10.1016/j.patrec.2020.03.027</idno>
	</analytic>
	<monogr>
		<title level="j">Pattern Recognition Letters</title>
		<imprint>
			<biblScope unit="volume">135</biblScope>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
