=Paper=
{{Paper
|id=Vol-3892/paper19
|storemode=property
|title=Artificial Intelligence-Based Method for Face Skin Diagnostic
|pdfUrl=https://ceur-ws.org/Vol-3892/paper19.pdf
|volume=Vol-3892
|authors=Olga Pavlova,Vitalii Alekseiko,Vladyslav Karabaieva,Andrii Kuzmin
|dblpUrl=https://dblp.org/rec/conf/iddm/PavlovaAKK24
}}
==Artificial Intelligence-Based Method for Face Skin Diagnostic==
Artificial Intelligence-based method for face skin
diagnostic⋆
Olga Pavlova1,∗,†, Vitalii Alekseiko1,∗,†, Vladyslav Karabaiev1,2,† and Andrii Kuzmin1,†
1
Khmelnytskyi National University, Institutska str., 11, Khmelnytskyi, 29016, Ukraine
2
Healthy Face Clinic, Stepana Bandery str., 5a, Khmelnytskyi, 29000, Ukraine
Abstract
The skin of the face is a complex organ, the condition of which can vary due to various factors such as
genetics, lifestyle, environmental conditions and age. An accurate assessment of the condition of the skin
is critical for the correct selection of care and treatment methods, which stimulates the development of
technologies for its analysis. Recent advances in artificial intelligence (AI) provide new opportunities for
automated skin analysis, which improves the accuracy of diagnosis and the efficiency of procedures.
This work focuses on the review and analysis of the application of facial skin analysis systems integrated
with artificial intelligence algorithms. It is important to understand the working principles of such
systems and their potential in practical application both in cosmetology and in other medical fields. The
purpose of this work is to study the technical aspects of creating such skin analyzers, in particular their
software part, and to evaluate their practical effectiveness based on modern machine learning algorithms.
Keywords 1
Artificial Intelligence (AI), facial diagnostic, decision support, IT solutions for medicine
1. Introduction
Skin health is a critical aspect of overall well-being, and early detection of skin conditions plays a
vital role in effective treatment and prevention. Facial skin, in particular, is highly susceptible to
various dermatological issues, such as acne, hyperpigmentation, dryness, and signs of aging.
Traditional skin diagnostic methods often require clinical expertise, specialized equipment, and
time-consuming procedures. In recent years, advancements in artificial intelligence (AI) have
offered new opportunities to enhance skin diagnostic processes, providing efficient and accurate
solutions. Artificial intelligence, in particular machine learning, is opening new horizons in skin
diagnostics, allowing the creation of systems capable of analyzing skin images and making
recommendations for care and treatment based on identified problems.
At this stage of technology development, there are already devices that allow you to assess the
condition of the skin, both companies and science offer new approaches to facial skin diagnostics.
This paper proposes an AI-based method for face skin diagnostics, utilizing data obtained from a
smart skin analyser. The smart skin analyser collects comprehensive facial skin data, including
moisture levels, pigmentation patterns, pore size, and other relevant features. By integrating this
data with AI algorithms, the proposed method aims to automate and optimize the diagnostic
process, delivering reliable and consistent results.
Our approach leverages neural network application and image processing algorithms to analyse
the facial skin's characteristics. The proposed method identifies various skin conditions and
provides personalized recommendations based on the analysed data and offers a non-invasive,
efficient, and accessible solution for users seeking professional-level skin analysis and care in real-
time.
IDDM’2024: 7th International Conference on Informatics & Data-Driven Medicine, November 14-16, 2024, Birmingham, UK.
∗ Corresponding author.
† These authors contributed equally.
pavlovao@khmnu.edu.ua (O. Pavlova); vitalii.alekseiko@gmail.com (V.Alekseiko); vladkarabaev@gmail.com (V.
Karabaiev); andriy1731@gmail.com(A.Kuzmin).
0000-0003-2905-0215 (O. Pavlova); 0000-0003-1562-9154 (V.Alekseiko); 0009-0005-6489-225X (A.Kuzmin).
© 2024 Copyright for this paper by its authors. Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0).
CEUR
ceur-ws.org
Workshop ISSN 1613-0073
Proceedings
2. Related works
In the course of the study, an analysis of the latest scientific publications in the field of skin
diagnostic was carried out.
In [1] AI-based facial skin diagnosis system (Dr. AMORE®) uses facial images of Korean women
to analyse wrinkles, pigmentation, skin pores, and other skin red spots. The system is trained using
clinical expert evaluations and deep learning.
The aim of [2] is to evaluate the current state of AI-based techniques used in combination with
non-invasive diagnostic imaging modalities including reflectance confocal microscopy (RCM),
optical coherence tomography (OCT), and dermoscopy. It also aimed to determine whether the
application of AI-based techniques can lead to improved diagnostic accuracy of melanoma.
The objective of [3] is to design a system that combines metaheuristic optimizers with various
AI based classifiers to detect and diagnose skin diseases. In order to accomplish this objective,
numerical and image datasets have been taken, pre-processed, and visually analysed in order to
comprehend their patterns.
In [4] the propensity of skin cancer to metastasize highlights the importance of early detection
for successful treatment. This narrative review explores the evolving role of artificial intelligence
(AI) in diagnosing head and neck skin cancers from both radiological and pathological perspectives.
The proposed in [5] model has the potential to aid qualified healthcare professionals in the
diagnosis of melanoma. Furthermore, the authors propose a mobile application to facilitate
melanoma detection in home environments, providing added convenience and accessibility.
The paper [6] delves into unimodal models’ methodologies, applications, and shortcomings
while exploring how multimodal models can enhance accuracy and reliability.
The study [7] presents an automated skin lesion detection and classification technique utilizing
optimized stacked sparse autoencoder (OSSAE) based feature extractor with backpropagation
neural network (BPNN), named the OSSAE-BPNN technique.
The insights in [8] demonstrate the bias towards deep learning methods and the shortage of
studies on rare and precancerous skin lesions.
In paper [9] five different algorithms of artificial intelligence have been selected and used to
skin disease dataset.
The purpose of the study [10] was to assess the diagnostic accuracy of the teledermoscopy
method using the FotoFinder device as well as the Moleanalyzer Pro artificial intelligence (AI)
Assistant and to compare them with the face-to-face clinical examination for the diagnosis of
melanoma confirmed with histopathology.
In [11] we propose a methodology for consideration of civil-legal grounds in medical decision-
making process.
The paper [12] proposes a health recommender system for smart cities. The methodology
proposes the smart distribution of healthcare institutions which are located the closest to the
patient.
It is impossible not to notice that along with scientific developments, many new devices are
appearing on the market that allow to measure skin parameters and even predict the result after
surgical correction in plastic surgery.
For example, the VECTRA H2 imaging system [13] is a portable hardware skin diagnosis system
with volumetric body imaging for use in cosmetology, aesthetic medicine, and dermatology. The
features of VECTRA H2 from Canfield Scientific:
- Automatic merging: three face or body shots are automatically merged into one 3D image
by VECTRA software.
- Accurate assessment of contours: the gray visualization mode allows you to evaluate the
contours of the face and body without being distracted by color when planning and
studying the result of corrective procedures.
- Face and body measurement in automatic mode: volumetric visualization (3D mode) and
digital data help your patients understand the underlying problems.
- Breast Sculptor software application: technology for creating three-dimensional breast
models, based on selected implants, taking into account gravity, shape and location.
- Visual comparison: visualization of several breast augmentation surgery scenarios by
parameters, sizes and style of implants.
- Visualization of expectations: illustrative display of benefits after breast augmentation
surgery.
- Mastopexy program. A software application for simulating lifting operations taking into
account areas of skin excision.
- Volumetric measurements of the body: automatic measurement of the circumference and
volume of body contours.
- Quantification of the subcutaneous structures of the face: Canfield's patented technology
shares the unique color shades of red and brown facial skin. This allows you to get a
complete picture of the skin condition and improve the quality of imaging.
- Measurement of volume change: volume data is automatically measured with one click of
the mouse in grayscale mode with parallel color display of changes in facial contours.
- Markerless tracking: a dynamic assessment of changes in the surface of the facial skin is
carried out: alignment, direction and final result.
- Full picture of changes: the program creates a holistic picture of change, reflecting all the
hopes and expectations of your patient.
3D LifeViz® Mini [14] is the most compact 3D system for skin analysis and modelling, a
convenient solution for cosmetologists, dermatologists, cosmetic and plastic surgeons. Analyses the
condition of the patient's skin according to 6 parameters and reproduces the image of the face on
the screen in 3D format. The patient can see what his face could look like after contour plastic
surgery or surgery. The system is based on a special type of stereophotogrammetry, where 2D
images are automatically combined into a three-dimensional representation.
LifeViz® technology allows you to quantitatively change the volume and determine the small
details of the skin surface with extreme accuracy. An example of an image created by 3D LifeViz®
is presented in Figure 1.
Figure 1: Example of the image created by 3D LifeViz® Mini system for skin analysis and
modelling [14]
Taking into account the relevance of the problem of application modern information technologies
in medical area, namely, analysis of the results taken from the Smart skin analyzer [15] for facial
skin defects detection it was decided to develop a methodology of neural network application for
solving this issue.
Therefore, the purpose of the research is:
1) to consider convolutional neural network architectures for medical image analysis;
2) to evaluate the effectiveness of models for the task of detecting and classifying skin defects;
3) to consider possible ways of improving the performance of models by changing the input
parameters.
3. Methodology
The data for the experiment were obtained using the skin analyzer AISIA [15], which is presented
in Figure 2. It provides the analysis of the skin according to the following parameters:
- Pore size (spectral visualization of RGB pores).
- The presence of blackheads and postacne: an analysis of all spots of a round shape with a
color darker than usual.
- Age changes: the degree of wrinkles and the depth of creases.
- Skin texture: imaging changes in relief, texture of the dermis, as well as predicting the
degree of future changes.
- The level of secretion of sebum and the localization of black spots.
- Pigmentation zones: an image not only of the actual state, but also of forecasting future
formations.
- Hydration.
- Areas of sensitivity: determination of zonal sensitivity, its susceptibility to allergic reactions.
- Brown zones: an image of metabolic processes in cells, areas of current recovery.
- Injury by UV rays: an image of pigmentation at different levels of the epidermis. Fixation of
the size and depth of such spots.
- Diagnosis of age-related changes: a picture of the future aging of the dermis, wrinkles, if the
client does not change his care.
Figure 2: Smart portable skin analyzer AISIA [15]
During the research, a wide range of methods was used, including general scientific methods:
theoretical (modeling, analysis, synthesis), empirical (observation, comparison, experiment).
Medical diagnostic methods were also used to form the dataset, and artificial intelligence tools
were used, in particular, machine learning models for image analysis.
As machine learning technologies demonstrate their effectiveness in the analysis of medical
images [16,17], we consider it appropriate to consider basic convolutional neural network (CNN)
models for facial skin defect detection.
3.1. Residual Networks
The main idea of ResNet is to learn the residual mapping [22]:
F(x) = H(x) − x, (1)
where:
F(x) – residual mapping,
H(x) – desired mapping,
x – input.
Thus, H(x) from formula (1) can be represented as:
H(x) = F(x) + x, (2)
A typical residual block consists of two or more convolutional layers with batch normalization
(BN) and ReLU activation:
y = F(x; {Wi}) + x, (3)
where:
Wi – weights of the convolutional layers;
y – output of the block.
Key Properties of ResNet are Identity Mapping and Training Deep Networks. The skip
connections enable identity mapping, facilitating gradient flow and addressing the vanishing
gradient problem. ResNets can have hundreds or thousands of layers while remaining easier to
train compared to traditional deep networks.
3.2. Dense Convolutional Network
In DenseNet, each layer receives inputs from all preceding layers. The output of layer can be
computed as:
�� = �� (�0 , �1 , . . . , ��−1 ), (4)
where:
Hl – operations performed by the layer;
xi – feature maps from all preceding layers.
Instead of adding the inputs as in ResNet, DenseNet concatenates feature maps:
�� = ��−1 , ��−2 , . . . , �0 , (5)
Key Properties of DenseNet are Feature Reuse and Gradient Flow [18]. DenseNet emphasizes
feature reuse, which reduces the number of parameters while still maintaining high accuracy. The
dense connectivity pattern allows gradients to flow through many paths during backpropagation,
enhancing learning.
3.3. EfficientNet
EfficientNet uses a compound scaling method, which balances the scaling of depth d, width w, and
resolution r. The scaling can be defined as:
� = �� , � = �� , � = �� , (6)
where:
k – constant;
α, β, γ – scaling coefficients.
EfficientNet starts from a baseline model and scales it. For instance, the total number of
parameters P in the model can be expressed as:
� = � ∙ �2 ∙ �2 , (7)
where c – constant that defines the efficiency of the architecture.
Key Properties of EfficientNet are Optimized Architecture and Efficiency. The architecture is
optimized through neural architecture search, allowing for a balance between model size and
performance. EfficientNet achieves state-of-the-art accuracy with fewer parameters compared to
previous models [19].
3.4. MobileNet
MobileNet introduces depthwise separable convolutions, which factor the standard convolution
into two separate layers: Depthwise Convolution and Pointwise Convolution [20].
Depthwise Convolution can be represented as a single filter, which is applied to each input
channel.
�� = � ⊙ �� , (8)
Pointwise Convolution can be represented as a 1×1 convolution that combines the output from
the depthwise layer.
� = ���������(��) = �� ∙ �� , (9)
Key Properties of MobileNet are Efficiency and Width Multiplier. The reduction in the number
of parameters and computations compared to standard convolutions makes MobileNet highly
suitable for mobile and edge devices. MobileNet allows for a width multiplier α to reduce the
number of channels in each layer, further optimizing the model size.
3.5. Dataset Structure
The Skin Disease Classification Dataset from Kaggle [21] was used to test the performance of the
models. This dataset contains a collection of photographs of human faces divided into three distinct
classes: acne, bags under the eyes, and facial redness. To ensure a comprehensive presentation and
accurate classification, it is advisable to consider three photos for each person. These images
include a front view along with side profiles on both the right and left. This multi-angle approach
not only increases the variability of the data, but also helps the models learn to identify and
differentiate the subtle nuances of skin diseases that may be missed in a single photo. The dataset's
structured format and comprehensive documentation make it an invaluable resource for
developing, training, and fine-tuning machine learning algorithms aimed at classifying skin
diseases, ultimately contributing to advances in dermatology diagnostics and personalized skin
care solutions. All data are presented in a generalized form and used exclusively within the scope
of scientific research. The work used data from open resources, supplemented with medical data
from clinical practice. Before using medical data, permission for their use was obtained from each
of the patients. The research provides the principles of responsible artificial intelligence.
Confidentiality of information is guaranteed.
4. Models’ comparison
For detecting skin defects such as acne, redness, etc., the choice of CNN architecture will depend
on factors such as dataset size, problem complexity, and available computing resources. The most
commonly used CNN architectures include:
– Residual Networks (ResNet);
– Dense Convolutional Network (DenseNet);
– EfficientNet;
– MobileNet.
Table 1 shows a comparison of the capabilities of the presented architectures.
Using ResNet enables deep learning, thanks to the use of residual connections, which allows
you to train deeper networks without facing the gradient vanishing problem. This can help the
model capture more complex patterns in skin texture and blemishes. The model shows high
performance in the analysis of medical images.
Pre-trained ResNet models (such as ResNet-50 or ResNet-101) can be fine-tuned on datasets to
improve performance, especially when using limited datasets.
This model is recommended for complex tasks where high accuracy is required.
DenseNet connects each layer to all other layers, which promotes feature reuse and results in
stronger gradients. This is useful when detecting subtle skin imperfections such as acne or
discoloration.
Compared with ResNet, DenseNet achieves high accuracy using fewer parameters, which can
reduce training time while maintaining high performance.
DenseNet is widely used in medical image classification tasks, so the model can be effective for
dermatological image analysis.
The model is recommended for medium and large data sets with a focus on achieving high
accuracy, especially in conditions of limited computing resources.
EfficientNet has high performance while using less computation. The model systematically
scales width, depth, and resolution, making it extremely efficient in terms of both accuracy and
computational resources.
The EfficientNet pre-trained models are very effective, when fine-tuned, for medical imaging
tasks, including skin defect detection. EfficientNet has been proven to outperform models such as
ResNet and DenseNet in various medical image classification tasks and, at the same time, is more
resource efficient.
The use of the model is recommended for projects that require a balance between high accuracy
and efficiency, especially when working with large-scale images or in cases of limited computing
power.
Table 1
Comparison of the CNN architectures capabilities
CNN Key Features Advantages Disadvantages Common
Architecture Applications
Residual Depth increases Avoids vanishing Large parameter Image
Networks with stacked gradients. size for deeper classification,
residual blocks. Deeper networks versions. object
Skip connections improve Computationally detection, face
allow gradient performance. intensive on larger recognition
flow. Robust and widely ResNet variants.
used.
Dense Dense connections Reduces number of High Image
Convolutional throughout the parameters. computational cost classification,
Network network. Efficient feature in memory due to medical image
Fewer parameters reuse and learning. dense connections. analysis
due to feature Good for small More prone to
reuse. datasets. overfitting on small
Uses growth rate data.
to control new
features at each
layer.
EfficientNet Optimized for both Good accuracy with Complex design. Mobile vision,
efficiency and less computation. Heavier versions image
accuracy. Flexible still require classification,
Compound scaling architecture for substantial object detection
(balance between different computational
depth, width, and constraints. power.
resolution). Scalable.
MobileNet Depthwise Highly efficient on – Lower accuracy Mobile vision,
separable resource- compared to larger real-time image
convolutions to constrained networks. processing
reduce devices. – Limited model
computation. Lightweight and capacity for
Optimized for fast. complex tasks.
mobile and edge Suitable for mobile
devices. applications.
MobileNet is a lightweight and fast model optimized for efficiency and ideal for real-time detection
of skin defects on mobile or embedded devices.
The model has low computational requirements because it uses depth-separated convolutions,
greatly reducing the number of parameters. Thus, the model is ideal for applications where
computing resources are limited.
Based on the above, it can be concluded that the model is suitable for deployment on mobile
platforms and is an excellent solution for applications that detect skin defects using a smartphone
camera.
5. Experiments & Results
Testing of the proposed models revealed that the ResNet and EfficientNet models demonstrate low
results on the test sample, while DenseNet and MobileNet perform the task of recognizing skin
defects with high accuracy (Figure 3).
Figure 3: Model Training Accuracy Comparison
The loss rate quantifies the error the model makes on the test set. Usually reflects how well the
predicted probabilities agree with the actual values. This indicator was the lowest for MobileNet,
and DenseNet was somewhat inferior (Figure 4). The loss indicators for ResNet and EfficientNet
turned out to be too high, so it was concluded that it is not appropriate to use these models for the
task of classifying skin defects.
Although the models show high accuracy and low loss on the training data set, on the
validation set the accuracy is slightly lower and the loss is higher, indicating overfitting of the
model. Thus, we propose to change the approach to the formation of the dataset, by including more
parameters, in particular, adding images made using the ultraviolet spectrum, as well as converting
images into heat maps. This will allow for a more qualitative assessment of the image and
contribute to a more accurate determination of the nature of the defect. In addition, it is advisable
to use sensors to determine skin moisture. The main parameters of the proposed dataset are listed
in Table 2.
Figure 4: Model Training Loss Comparison
Table 2
The main parameters of the proposed dataset
Parameter Unit Data type
ID None int
Moisture
Sensitivity
Pigment
UV spots
Texture
Blackheads
Pores
Stains % float
Color
UV acne
Sebum
Sebumt
Acne
Wrinkle
Porphyrin
Front photo
Left-side photo
Right-side photo
Front UV photo
Left-side UV photo None jpg image
Right-side UV photo
Front heatmap photo
Left-side heatmap photo
Right-side heatmap photo
6. Conclusions & Future work
Skin health is an essential component of overall well-being, and the early identification of skin
conditions is crucial for effective treatment and prevention. The facial skin, in particular, is
especially prone to various dermatological concerns, including acne, hyperpigmentation, dryness,
and signs of aging.
In the course of this study the main models of convolutional neural networks are considered,
the peculiarities of their application in the tasks of medical image analysis, in particular for the
detection of facial skin defects, are analyzed. The advantages and disadvantages of the most widely
used architectures, as well as the features of their application, are described. The models were
tested on a dataset including medical images. The effectiveness of the models was evaluated
according to the accuracy and loss metrics. On the basis of the conducted research, the structure of
the dataset is proposed, based on a larger number of parameters, which will significantly improve
the accuracy of the models by selecting the most relevant features and analyzing them.
For further research, there are plans to systematically identify and extract the key features that
are critical for effectively classifying skin defects through the application of Convolutional Neural
Networks (CNNs). This involves conducting an in-depth analysis of the various characteristics
present in the dataset, such as texture, color variations, and the specific patterns associated with
each skin condition. By leveraging techniques like feature extraction and selection, we aim to
isolate the most informative attributes that contribute to the accurate identification of skin defects.
This process will not only enhance the performance of the CNN models but also improve their
interpretability, allowing for better understanding and insight into how these models make
classification decisions. Furthermore, this feature extraction phase will enable us to refine the
model architectures and optimize hyperparameters, leading to more robust and generalizable
outcomes. Ultimately, this research will contribute to the development of more precise and reliable
diagnostic tools in dermatology, potentially transforming the way skin conditions are assessed and
treated in clinical practice.
7. Acknowledgements
The authors would like to thank Healthy Face Clinic and Dr. Vladyslav Karabaiev for providing the
equipment for data gathering that made the experiments and this work possible.
8. Declaration on Generative AI
It was used ChatGPT and Microsoft Copilot to rephrase sentences to improve style. The text was
carefully checked by the authors after using these software. The authors take full responsibility for
the publication.
9. References
[1] Park, H., Park, S. R., Lee, S., Hwang, J., Lee, M., Jang, S. I.& Kim, E. (2024). Development and
application of artificial intelligence‐based facial skin image diagnosis system: Changes in facial
skin characteristics with ageing in Korean women. International Journal of Cosmetic Science,
46(2), 199-208.
[2] Patel, R. H., Foltz, E. A., Witkowski, A., & Ludzik, J. (2023). Analysis of artificial intelligence-
based approaches applied to non-invasive imaging for early detection of melanoma: a
systematic review. Cancers, 15(19), 4694.
[3] Singh, J., Sandhu, J. K., & Kumar, Y. (2024). An analysis of detection and diagnosis of different
classes of skin diseases using artificial intelligence-based learning approaches with hyper
parameters. Archives of Computational Methods in Engineering, 31(2), 1051-1078.
[4] Semerci, Z. M., Toru, H. S., Çobankent Aytekin, E., Tercanlı, H., Chiorean, D. M., Albayrak, Y.,
& Cotoi, O. S. (2024). The Role of Artificial Intelligence in Early Diagnosis and Molecular
Classification of Head and Neck Skin Cancers: A Multidisciplinary Approach. Diagnostics,
14(14), 1477.
[5] Orhan, H., & Yavşan, E. (2023). Artificial intelligence-assisted detection model for melanoma
diagnosis using deep learning techniques. Mathematical Modelling and Numerical Simulation
with Applications, 3(2), 159-169.
[6] Strzelecki, M., Kociołek, M., Strąkowska, M., Kozłowski, M., Grzybowski, A., & Szczypiński, P.
M. (2024). Artificial Intelligence in the detection of skin cancer: state of the art. Clinics in
Dermatology.
[7] Ogudo, K. A., Surendran, R., & Khalaf, O. I. (2023). Optimal Artificial Intelligence Based
Automated Skin Lesion Detection and Classification Model. Computer Systems Science &
Engineering, 44(1).
[8] Rezk, E., Haggag, M., Eltorki, M., & El-Dakhakhni, W. (2023). A comprehensive review of
artificial intelligence methods and applications in skin cancer diagnosis and treatment:
Emerging trends and challenges. Healthcare Analytics, 100259.
[9] Pattnayak, P., Patnaik, S., Gourisaria, M. K., Singh, S., Barik, L., & Patra, S. S. (2024, August).
Analysis and Detection of Skin Disorders using Artificial Intelligence-based learning. In 2024
Second International Conference on Networks, Multimedia and Information Technology
(NMITCON) (pp. 1-5). IEEE.
[10] Yazdanparast, T., Shamsipour, M., Ayatollahi, A., Delavar, S., Ahmadi, M., Samadi, A., & Firooz,
A. (2024). Comparison of the Diagnostic Accuracy of Teledermoscopy, Face-to-Face
Examinations and Artificial Intelligence in the Diagnosis of Melanoma. Indian Journal of
Dermatology, 69(4), 296-300.
[11] Hnatchuk Y., Hovorushchenko T., Pavlova O. Methodology for the development and
application of clinical decisions support information technologies with consideration of civil-
legal grounds. Radioelectronic and Computer Systems. pp. 33-44
[12] Bouhissi, H.E., Tagzirt, D., Bouredjioua, F., Pavlova O. Health Recommender System for Smart
Cities. CEUR Workshop Proceedings, 2023, 3426, pp. 334–343
[13] Vectra H2 official website. URL: https://beautix.com.ua/equipment/diagnostika_skin/vectra_h2
(Last accessed September 26, 2024).
[14] Lascos Aesthetic Medicine: Lifeviz Pro Mini. URL: https://www.lascos.com.ua/apparaty/3d-
photo-cameri/lifeviz-pro-mini-1030785657 (Last accessed September 26, 2024).
[15] Aisia 3D Smart Face Skin Analyzer. URL:
https://medunion.com/product/rFaTtwEYXdWC/China-Aisia-3D-Smart-Face-Skin-Analyzer-
for-Salon-Hot-Skin-Scanner-Facial-Analyzer.html (Last accessed September 26, 2024).
[16] D. R. Sarvamangala, & R. V. Kulkarni. Convolutional neural networks in medical image
understanding: a survey. Evolutionary Intelligence, 15(1), 1–22. 2021.
https://doi.org/10.1007/s12065-020-00540-3
[17] Q. Zhou, Z. Huang, M. Ding, & X. Zhang. Medical Image Classification Using Light-Weight
CNN With Spiking Cortical Model Based Attention Module, in IEEE Journal of Biomedical and
Health Informatics, vol. 27, no. 4, pp. 1991-2002, April 2023.
https://doi.org/10.1109/JBHI.2023.3241439
[18] H. A. Ahmed, H. M. Hama, S. I. Jalal, & M. H. Ahmed. Deep learning in grapevine leaves
varieties classification based on dense convolutional network. Journal of Image and Graphics,
11(1), 98–103. 2023. https://doi.org/10.18178/joig.11.1.98-103
[19] M. A. Talukder, M. A. Layek, M. Kazi, M. A. Uddin, & S. Aryal. Empowering COVID-19
detection: Optimizing performance through fine-tuned EfficientNet deep learning architecture.
Computers in Biology and Medicine, 168, 2023. 107789.
https://doi.org/10.1016/j.compbiomed.2023.107789
[20] M. S. Al Reshan, K. S. Gill, V. Anand, S. Gupta, H. Alshahrani, A. Sulaiman, & A. Shaikh.
Detection of Pneumonia from Chest X-ray Images Utilizing MobileNet Model. Healthcare,
11(11), 2023. 1561. https://doi.org/10.3390/healthcare11111561
[21] Skin Disease Classification Dataset. Kaggle. 2023, November 16. URL:
https://www.kaggle.com/datasets/trainingdatapro/skin-defects-acne-redness-and-bags-under-
the-eyes (Last accessed September 26, 2024).
[22] L. Abdelrahman, M. Al Ghamdi, F. Collado-Mesa, & M. Abdel-Mottaleb. Convolutional neural
networks for breast cancer detection in mammography: A survey. Computers in Biology and
Medicine, 131, 2021. 104248. https://doi.org/10.1016/j.compbiomed.2021.104248