<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Vision transformers transfer learning for smoking detection in public spaces with transparent AI decisions</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Olexander Mazurets</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Maryna Molchanova</string-name>
          <email>m.o.molchanova@gmail.com</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Olha Zalutska</string-name>
          <email>zalutsk.olha@gmail.com</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Ihor Kok</string-name>
          <email>kokigorek03@gmail.com</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Vitaly Levashenko</string-name>
          <email>vitaly.levashenko@fri.uniza.sk</email>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Abdel-Badeeh M. Salem</string-name>
          <email>abmsalem@yahoo.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Ain Shams University</institution>
          ,
          <addr-line>El-Khalyfa El-Mamoun Street Abbasya, Cairo</addr-line>
          ,
          <country country="EG">Egypt</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Khmelnytskyi National University</institution>
          ,
          <addr-line>11, Instytuts'ka str., Khmelnytskyi, 29016</addr-line>
          ,
          <country country="UA">Ukraine</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Zilina University</institution>
          ,
          <addr-line>Univerzitná 8215, 010 26 Žilina</addr-line>
          ,
          <country country="SK">Slovakia</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2025</year>
      </pub-date>
      <abstract>
        <p>The automated detection of smoking in public spaces is increasingly critical for ensuring compliance with public health regulations and mitigating the economic burden of smoking-related diseases. However, existing computer vision approaches often struggle with limited labeled data and lack the interpretability required for trustworthy automated surveillance. In this work, we propose a transfer learning method using the Vision Transformer (ViT) architecture combined with transparent AI decision-making mechanisms. Our approach employs partial parameter freezing to adapt to small datasets while preventing overfitting and utilizes Grad-CAM heat maps to provide visual explanations of the model's focus. Experimental results demonstrate that the proposed method achieves an accuracy of 0.98, surpassing the InceptionResNetV2 baseline by 0.011 and outperforming the YOLOv5small model by 0.045 in Precision and 0.089 in Recall. The study concludes that the integration of explainable AI not only enhances detection performance but also fosters trust by verifying that the model attends to relevant semantic features, thereby contributing directly to Sustainable Development Goals #3 and #11.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;ViT</kwd>
        <kwd>smoking detection</kwd>
        <kwd>transparent AI decisions</kwd>
        <kwd>transfer learning</kwd>
        <kwd>machine learning</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Tobacco smoking is one of the main causes of various diseases [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], such as cancer, cardiovascular diseases
and respiratory diseases, which significantly reduce the quality of life and increase the economic burden
on the health care system [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. Given restrictions on smoking in public places in many countries [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ],
automation of detecting violations process of such prohibitions becomes an important component for
efective control over compliance with the law [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ].
      </p>
      <p>
        Despite the existence of legal restrictions on smoking in public spaces, the efectiveness of their
compliance largely depends on human resources [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ], which are limited and not always able to provide
full-fledged monitoring in real time [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. In this context, the use of computer vision technologies,
in particular deep learning models, opens up new opportunities for creating automatic surveillance
systems capable of identifying smoking facts in video or images with high accuracy and speed [
        <xref ref-type="bibr" rid="ref7 ref8">7, 8</xref>
        ].
This approach not only helps to increase the eficiency of detecting violations, but can also become an
important tool in preventing the harmful efects of passive smoking on the population [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ], especially
in places of mass gathering of people [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ]. In addition, the implementation of such solutions is a step
towards the formation of smart cities [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ], where technologies contribute to the implementation of
state health policy [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ].
      </p>
      <p>
        The task of detecting smoking in public areas can also be contextualized within the framework of
the United Nations Sustainable Development Goals, as it involves the application of innovative data
management and monitoring technologies to promote public health [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ]. It aligns with SDG #3 “Good
Health and Well-being”, as it ofers an instrument for preventing smoking-related health risks and
reducing the consequences of passive smoking [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ]. At the same time, it corresponds to SDG #11
“Sustainable Cities and Communities”, since the use of advanced AI-based surveillance tools strengthens
urban safety and supports the development of sustainable and human-centered public spaces [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ]. From
this perspective, the advancement of interpretable deep learning solutions for smoking detection not
only enhances public health protection, but also integrates artificial intelligence into broader strategies
of sustainable urban growth.
      </p>
      <p>The main goal of the paper is to improve the accuracy of detecting smoking in images by applying
transfer learning using the Vision Transformers architecture with transparent AI decisions. The main
contribution of the paper is the method of detecting smoking using deep learning neural networks,
which difers from existing ones by using partial parameter fixation and modification of training data,
as well as using the explainability of neural network solutions in the form of a heat map. The proposed
approach involves freezing all layers of the model, except for the final classifier, which allows adapting
the model to a task with a limited amount of data and reducing the risk of overtraining.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related works</title>
      <p>
        The problem of automated smoking identification from digital images is studied in modern scientific
works. An improved method for detecting smoking in public places based on an improved version of
the YOLOv5-small algorithm is considered in [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ]. The main attention is paid to increasing the accuracy
and reliability of detecting cigarettes as objects in images with a complex background, small target size,
and visual similarity to other objects. The proposed solution is based on the use of hybrid transformer
and pyramidal structures to improve feature extraction, spatial generalization, and reduce the impact of
visual interference. The authors obtained Precision 0.935 and Recall 0.891 on a specialized dataset of
images with cigarettes.
      </p>
      <p>
        The article [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ] presents an approach to automated monitoring of compliance with the smoking ban
within a smart city by developing a smoking detection system based on artificial intelligence. The authors
propose a framework that combines the use of transfer learning with a pre-trained Inception-ResNetV2
model to classify images as containing or not containing smoking. In addition, the study contributes
to the field by creating a new dataset of images in indoor and outdoor conditions, which provides a
basis for further research. The proposed method demonstrates high performance on the experimental
dataset, achieving Accuracy 0.969, indicating its potential suitability for real-time applications.
      </p>
      <p>
        In [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ] the potential of deep learning for detecting hidden tobacco advertising in media content
is investigated. Multimodal model is proposed that combines image and text analysis using neural
networks, generative methods and expert reinforcement learning mechanisms. This approach allows to
detect smoking even with a limited amount of training data. The researchers [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ] provide an overview
of modern methods for classifying and detecting smokers using computer vision based on deep learning.
The study includes a critical analysis of existing models, used datasets and their efectiveness, and also
ofers a systematization of approaches used in the literature. To overcome the shortage of labeled data,
the authors present a new dataset CigDet, specially created for cigarette localization tasks, and evaluate
its efectiveness on diferent variants of YOLO models, where the best results are achieved by YOLOv9
(mAP = 83.5%). The conceptual architecture of the SURRONE system is also presented, which can be
used for real-world monitoring of smoking using drones, with the prospect of integration into public
health systems and surveillance technologies.
      </p>
      <p>
        In the study [
        <xref ref-type="bibr" rid="ref20">20</xref>
        ], the application of computer vision for automated detection of visual content related
to e-cigarettes on youth-oriented video platforms such as TikTok was demonstrated. The YOLOv7
model was used to detect vaping devices, hands, and vapor clouds in images obtained from posts with
popular hashtags. The model achieved high classification Accuracy (up to 0.929) and demonstrated
reliable generalization (1 = 0.81), which demonstrates the feasibility of using such approaches for
monitoring visual violations of tobacco policies in the online environment.
      </p>
      <p>
        A method for indoor smoking detection combining infrared and visible images with the YOLOv9
architecture to improve accuracy in low-visibility conditions is proposed in [
        <xref ref-type="bibr" rid="ref21">21</xref>
        ]. By using bispectral
fusion, data-level and feature-level optimization, and model training, the system achieves high average
accuracy (mAP@0.5 = 0.958) on a specialized dataset.
      </p>
      <p>Despite the availability of efective deep learning approaches to smoking detection, research into new
methods remains relevant due to the need to expand scientific understanding and technical variability.
The development of alternative architectures and data processing methods allows not only to reevaluate
existing hypotheses regarding feature representation, but also to adapt known solutions to specific
operating conditions.</p>
    </sec>
    <sec id="sec-3">
      <title>3. Method design</title>
      <sec id="sec-3-1">
        <title>3.1. Problem statement</title>
        <p>Despite significant advances in deep learning-based smoking detection, existing methods often face
limitations related to performance on small datasets and the need to adapt to various operating conditions.
Many modern deep learning models, in particular those used in computer vision, require large amounts
of labeled data to achieve high accuracy, which makes their application in real-world conditions dificult.
Despite this, the technical potential of transfer learning remains underexplored in the context of specific
tasks, such as detecting smoking in public places.</p>
        <p>There is a need to develop strategies to improve the adaptability of deep learning models, such
as freezing the parameters of all layers except the final classifier, which allows preserving the
generalization ability of the model, while reducing the risk of overtraining and improving eficiency in
resource-constrained conditions. The main scientific question is how transfer learning using the Vision
Transformers architecture can be adapted to the task of smoking detection with limited data volumes,
and whether this can lead to increased detection accuracy without the risk of overtraining. Accordingly,
there is a need not only to improve the Accuracy of smoking detection, but also to verify whether the
neural network correctly highlights the relevant areas of attention in images. Therefore, all AI decisions
must be transparent.</p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Proposed method</title>
        <p>The smoking detection in public spaces method is designed to identify smoking in public places using
neural network tools (Figure 1). The input data are images for analysis and a trained neural network
model (for more details about the model, see Section 3.3).</p>
        <p>
          In step 1, the image is preprocessed: it is resized to 224 × 224 pixels, normalized using ImageNet
statistics, and transformed into a tensor that meets the requirements of the input layer of the ViT
model [
          <xref ref-type="bibr" rid="ref22 ref23">22, 23</xref>
          ]. After that, the image is automatically partitioned into non-overlapping patches of
16 × 16 pixels, each of which is prepared for further vector encoding. This partitioning is an intrinsic
property of the ViT architecture and is implemented through a convolutional layer that simultaneously
performs partitioning and projection. In step 2, each patch is transformed into a fixed-length vector
using linear projection. These vectors form a sequence to which positional information is added.
A transform encoder operates on the entire sequence of patches, which uses multi-head attention
mechanisms to analyze spatial relationships between diferent regions of the image. The result of this
stage is a representation vector for the entire image – a classification token, which contains generalized
information about the image.
        </p>
        <p>
          The resulting vector representation obtained in step 3 is passed to the final linear layer, which was
trained for binary classification. The output value is interpreted as a logit, which characterizes the
probability of the presence of smoking in public places. A sigmoid activation function is used to obtain
a probabilistic result. Thus, the model generates a numerical estimate in range of [
          <xref ref-type="bibr" rid="ref1">0, 1</xref>
          ], which is
thresholded as the classes “Smoker” or “Non-Smoker”. In the study, the threshold was taken as 0.5 (if
less than 0.5, the class “Non-Smoker”, if greater than or equal to – “Smoker”).
        </p>
        <p>
          In order to interpret the model solution, step 4 is performed – the Grad-CAM method is applied [
          <xref ref-type="bibr" rid="ref24">24</xref>
          ].
This allows visualizing the spatial regions that most influenced the classification [
          <xref ref-type="bibr" rid="ref25">25</xref>
          ]. In the context of
Vision Transformer, a projection layer of patches is used as a reference for generating an importance
map. Based on the calculated gradients, a heat map is formed and superimposed on the image. It shows
which areas of the image were decisive for the decision, increasing transparency and confidence in the
model [
          <xref ref-type="bibr" rid="ref26">26</xref>
          ]. The output data are the class (Smoker or Non-Smoker), a numerical score of the presence
of smoking (in the range from 0 to 1), and an interpretive heat map.
        </p>
      </sec>
      <sec id="sec-3-3">
        <title>3.3. Neural network model</title>
        <p>The scheme of converting the input image into the probability of identifying smoking in public places
through the applied neural network architecture is shown in Figure 2. At the classification stage, the
input image with a size of 224 × 224 pixels is fed to the ViT model, pre-scaled and normalized according
to the statistical parameters of ImageNet. The image is divided into 196 non-overlapping patches with a
size of 16 × 16 pixels, which form a regular 14 × 14 grid. Each patch is passed through a convolutional
layer (patch embedding), which performs linear projection, resulting in a sequence of feature vectors of
ifxed dimension (768).</p>
        <p>A special classification token (CLS-token) is added to this sequence, which is designed to accumulate
global information about the image. Thus, the total number of input vectors is 197. These vectors
are sequentially passed to the transformer encoder, which consists of 12 serially connected blocks.
Each block contains two main components: a Multi-Head Self-Attention (MHSA) mechanism and a
Multi-Layer Perceptron Network (MLP), accompanied by layer normalization operations. The attention
mechanism provides the model with the ability to capture dependencies between remote patches,
allowing, in particular, to establish connections between semantically related regions of the image. MLP
layers are responsible for nonlinear feature transformation, and normalization improves the stability of
calculations.</p>
        <p>After passing through all the transformer blocks, the updated CLS-token acts as an aggregated
representation of the entire image. This token is passed to the final classification layer – a linear
projection, which outputs a single scalar value (logit). A sigmoid activation function is used to transform
the logit into a probabilistic interpretation. The final value is compared with a fixed threshold (0.5),
which allows determining a binary class: class “1” is interpreted as smoking signs presence (Smoker),
class “0” as their absence (Non-Smoker).</p>
        <p>
          In order to improve the training eficiency, partial freezing of the model parameters was applied,
in which only the weights of the last linear layer are optimized. This helps to speed up the training
process, as the number of parameters that need to be updated is reduced, which allows to significantly
reduce computational costs. The use of pre-trained weights, in particular through the use of SWAG
(Stochastic Weight Averaging Gaussian) for linear layers, improves the model’s ability to generalize
to new tasks, which is critically important when working with limited data sets. This approach also
reduces the risk of overtraining, since the model does not “remember” the specific features of training
set [
          <xref ref-type="bibr" rid="ref27">27</xref>
          ], but focuses on the general characteristics of the images. In addition, freezing most of the
parameters allows to reduce the requirements for computational resources, focusing the calculations
only on the optimization of the weights critical for classification, which increases the eficiency of using
computational resources when training large architectures. Such a modification helps to improve the
quality of classification and reduce the amount of necessary calculations, especially on small datasets.
        </p>
        <p>Also, an image preprocessing procedure is applied before feeding them to the model and is aimed at
ensuring structural uniformity and expanding the variability of the training data. In particular, scaling
images to a single size guarantees compliance with the architectural requirements of the model, while
stochastic augmentations, in particular mirroring and changing color characteristics, contribute to
increasing the model’s resistance to changes in external shooting conditions. Normalization of channel
intensities improves the stability and convergence speed of the optimization process by eliminating the
variability of the scales of the input features. Taken together, these transformations ensure both the
invariance of the model to variations in data and correct representation of information for efective
training.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Experiment</title>
      <sec id="sec-4-1">
        <title>4.1. Dataset</title>
        <p>
          The “Smoker Detection [Image] classification Dataset” [
          <xref ref-type="bibr" rid="ref28">28</xref>
          ] represents a balanced sample of visual data
consisting of two semantic categories – smokers and non-smokers. Each class includes the same number
of images, which ensures symmetry in the representation of the target categories and reduces the bias
of the model during training. The total number of samples in the dataset is 1120 (560 in each class).
        </p>
        <p>A feature of this corpus is the presence of a high degree of interclass similarity: images in the
non-smoker class were selected in such a way as to contain poses and gestures that are visually close
to smoking acts, in order to complicate the classification task. Such a selection strategy increases the
discriminatory power of the model and stimulates it to learn deeper contextual features. All images
were unified in resolution to ensure compatibility with the architectural requirements of deep learning
models.</p>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. Software implementation</title>
        <p>
          To validate the proposed method of smoking detection in public spaces, a software application was
implemented in the Python environment using the Google Colab cloud service [
          <xref ref-type="bibr" rid="ref29">29</xref>
          ], which provides
interactive code execution with hardware acceleration support (GPU/TPU) and facilitates experimentation
with image processing and deep model training. The application architecture is built on the basis of the
TensorFlow library [
          <xref ref-type="bibr" rid="ref30">30</xref>
          ] and its high-level API Keras, which made it possible to implement the Vision
Transformer model and organize the training process with the possibility of flexible hyperparameter
tuning.
        </p>
        <p>
          The model results were explained using the Grad-CAM technique, which was implemented by
constructing a custom computational graph using automatic diferentiation. The Matplotlib [
          <xref ref-type="bibr" rid="ref31">31</xref>
          ] and
OpenCV [
          <xref ref-type="bibr" rid="ref32">32</xref>
          ] libraries were used to visualize heat maps and interpret classification results. An examples
of image analysis by the developed software is shown in Figure 3.
        </p>
        <p>(a) Input Image 1 (Dataset sample)
(b) Input Image 2 (Author photo)</p>
        <sec id="sec-4-2-1">
          <title>The results obtained and their interpretations are presented below.</title>
        </sec>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Results and discussion</title>
      <p>The developed software was used to conduct a study of the developed method. Part of the input data of
the developed method is a transfer machine learning model. The results of the training are presented in
the form of a confusion matrix and a rock curve are shown in Figure 4. Classification report is shown
in Table 1. With these parameters, Accuracy value is 0.98.</p>
      <p>The results of the experimental study indicate the efectiveness of the proposed approach based on
the transfer learning of the Vision Transformer model. The achieved classification accuracy (according
to the Accuracy metric = 0.98) indicates the ability of the model to generalize knowledge on new data
with a minimum number of errors. High values of precision and recall (over 0.96) in combination
with balanced indicators for both classes indicate the absence of a bias of the model towards one of
the classes, which is critically important in tasks related to behavioral analytics. The hypothesis that
transfer learning using the Vision Transformers architecture can be adapted to the task of smoking
detection with limited amounts of data, which will lead to an increase in detection accuracy without
the risk of overtraining, is also confirmed by the given graph of the loss function (the neural network
was trained for 10 epochs, Figure 5).</p>
      <p>The loss function plot shows a steady decline with each epoch, while the neural network without
considered adaptations in the form of training sample transformations and weight freezing demonstrated
jumpy amplitudes and higher final values even with increasing number of epochs (up to 20). Analysis of
the confusion matrix and ROC curve confirms the ability of the model to accurately distinguish classes,
while maintaining stable sensitivity and specificity. This gives grounds to argue that the model not
only learns eficiently on a limited amount of adaptation data, but also demonstrates a high level of
generalization, which makes it suitable for application in real conditions.</p>
      <p>
        Since the goal was not only to improve the accuracy of smoking detection, but also to check whether
the neural network correctly highlights the relevant areas of attention in the images, heat maps were
also implemented for the analyzed images. An example of smoking identification and the corresponding
heat maps are shown in Figure 6. The results of heat map visualization confirm that the model not
only demonstrates high classification accuracy, but also forms semantically based spatial attention. The
active areas detected by the neural network during decision-making are mainly concentrated in areas of
the image containing signs of smoking - in particular, the mouth, face and cigarette area. This indicates
the ability of the model to form internal representations that correspond to meaningful visual patterns
associated with the target class. This result indicates that the network is not a “black box” [
        <xref ref-type="bibr" rid="ref33">33</xref>
        ], and its
behavior can be interpreted through visual attention mechanisms. The presence of a focus on relevant
regions of the image increases confidence in the model and confirms that the classification decision is
based on a meaningful analysis of the input signal, and not on distorted or random features.
(a) Sample 1
(b) Sample 2
      </p>
      <p>The obtained results were also compared with analogues. The comparison results are presented in
Table 2. The results obtained indicate a significant increase in the accuracy of recognizing the action
of smoking in public places when using the developed approach based on Vision Transformer with
the use of image transformations and weight freezing. The proposed model achieved an Accuracy of
0.98, which demonstrates an advantage over InceptionResNetV2, which, according to previous studies,
achieved an Accuracy of 0.969. In addition, compared to the YOLOv5-small approach, for which the
Precision and Recall values were 0.935 and 0.891, respectively, the developed model provides a significant
improvement in all classification metrics, achieving a balanced Precision, Recall and 1-measure value
of 0.98. This indicates a better ability of the model not only to accurately classify objects, but also to
reduce the number of both false positive and false negative activations. Thus, the proposed solution
outperforms both traditional CNN architectures and modern object detection approaches, demonstrating
the feasibility of using a transformative architecture for visual monitoring of violations in public spaces.</p>
      <p>Future research will be devoted to processing streaming video, in particular in real time, which will
allow expanding the application of the developed method in the context of video surveillance systems
and automated behavior monitoring. Special attention is planned to be paid to adapting the model to
conditions of low image quality, variability of angles and scene dynamics, which will allow increasing
its stability and generalizability in real operating conditions.</p>
    </sec>
    <sec id="sec-6">
      <title>6. Conclusion</title>
      <p>The paper proposes the transfer machine learning method for smoking detection in public spaces with
transparent AI decisions, which allows achieving the set goal of increasing the accuracy of detecting
smoking in images. In accordance with the goal, the accuracy increased compared to known analogues
by 0.011 in the Accuracy metric compared to the approach using InceptionResNetV2, as well as by
0.045 in the Precision metric and by 0.089 in the Recall metric compared to the approach based on
YOLOv5-small. In addition to the main goal of achieving accuracy, visual interpretations of neural
network decisions in the form of heat maps were used. Visual interpretations confirm the correctness
of the detected attention zones in the images, which encourages trust in the decisions made by artificial
intelligence, and also increases transparency, which is especially relevant in the field of automatic
control of compliance with social norms.</p>
      <p>The proposed method difers from existing ones by the use of partial parameter fixation (freezing all
layers of the model, except for the final classifier, which allows to adapt the model to a problem with a
limited amount of data and reduce the risk of overtraining and modification of training data), as well as
the use of explainability of neural network solutions in the form of a heat map.</p>
      <p>The implemented strategy of partial parameter freezing, in which only the weights of the final
classifier are optimized, allowed to significantly reduce computational costs and avoid overtraining, while
maintaining the model’s ability to generalize. The use of stochastic averaging of weights contributed to
increasing the stability of training and improving the results on validation samples. Comprehensive
image preprocessing, including normalization, scaling and augmentations, ensured the model invariance
to variability of shooting conditions.</p>
      <p>In further research, it is planned to focus on real-time streaming video processing to expand the
possibilities of applying the method in video surveillance systems. Particular attention will be paid to
increasing the model’s resilience to low image quality, changes in perspective, and scene dynamics.</p>
    </sec>
    <sec id="sec-7">
      <title>Acknowledgments</title>
      <p>The authors are grateful to Prof. Tetiana Hovorushchenko, Prof. Olexander Barmak, Prof. Iurii Krak
and other Program Committee members for organizing and conducting the workshop ExplAI-2025:
Advanced AI in Explainability and Ethics for the Sustainable Development Goals.</p>
    </sec>
    <sec id="sec-8">
      <title>Declaration on Generative AI</title>
      <sec id="sec-8-1">
        <title>The authors have not employed any Generative AI tools.</title>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>S.</given-names>
            <surname>Gupta</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Kumar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Gupta</surname>
          </string-name>
          ,
          <article-title>A comprehensive study on the harmful efects of the smoking on human beings, in: Challenges in Information, Communication and Computing Technology</article-title>
          , CRC Press, London,
          <year>2024</year>
          , pp.
          <fpage>577</fpage>
          -
          <lpage>582</lpage>
          . doi:
          <volume>10</volume>
          .1201/
          <fpage>9781003559085</fpage>
          -
          <lpage>99</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>T. G.</given-names>
            <surname>Pinto</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L. d. S.</given-names>
            <surname>Avanci</surname>
          </string-name>
          ,
          <string-name>
            <surname>A. C. M. Renno</surname>
            ,
            <given-names>D. C.</given-names>
          </string-name>
          <string-name>
            <surname>Hipolide</surname>
            ,
            <given-names>J. N. d.</given-names>
          </string-name>
          <string-name>
            <surname>Santos</surname>
            ,
            <given-names>P. R.</given-names>
          </string-name>
          <string-name>
            <surname>Cury</surname>
            ,
            <given-names>R. A.</given-names>
          </string-name>
          <string-name>
            <surname>Dedivitis</surname>
            ,
            <given-names>D. A.</given-names>
          </string-name>
          <string-name>
            <surname>Ribeiro</surname>
          </string-name>
          ,
          <article-title>The impact of genetic polymorphisms on genotoxicity (dna damage) induced by cigarette smoke in humans: A systematic review</article-title>
          ,
          <source>Journal of Applied Toxicology</source>
          (
          <year>2025</year>
          ). doi:
          <volume>10</volume>
          .1002/jat.4753.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>J.</given-names>
            <surname>Hart</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Lignou</surname>
          </string-name>
          ,
          <article-title>Generational smoking bans: inegalitarian without disadvantage?</article-title>
          ,
          <source>Journal of Medical Ethics</source>
          (
          <year>2025</year>
          ). doi:
          <volume>10</volume>
          .1136/jme-2024-110632.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>A.</given-names>
            <surname>Boretti</surname>
          </string-name>
          ,
          <article-title>Rethinking selective prohibitions: the inconsistency of a generational smoking ban in a permissive society</article-title>
          ,
          <source>Journal of Medical Ethics</source>
          (
          <year>2025</year>
          ). doi:
          <volume>10</volume>
          .1136/jme-2024-110577.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>S.</given-names>
            <surname>Howard</surname>
          </string-name>
          , G. Krishna,
          <article-title>Do smoking bans work?</article-title>
          ,
          <source>BMJ</source>
          (
          <year>2025</year>
          ). doi:
          <volume>10</volume>
          .1136/bmj.q2759.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>B.</given-names>
            <surname>Saunders</surname>
          </string-name>
          ,
          <article-title>A generational ban creates inequality between non-smokers</article-title>
          ,
          <source>Journal of Medical Ethics</source>
          (
          <year>2025</year>
          ). doi:
          <volume>10</volume>
          .1136/jme-2024-110602.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Lei</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Shi</surname>
          </string-name>
          ,
          <article-title>Smoking behavior detection algorithm based on yolov8-mnc</article-title>
          ,
          <source>Frontiers in Computational Neuroscience</source>
          <volume>17</volume>
          (
          <year>2023</year>
          ). doi:
          <volume>10</volume>
          .3389/fncom.
          <year>2023</year>
          .
          <volume>1243779</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>K.</given-names>
            <surname>Bobrovnikova</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Lysenko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Savenko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Gaj</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Savenko</surname>
          </string-name>
          ,
          <article-title>Technique for IoT malware detection based on control flow graph analysis</article-title>
          ,
          <source>Radioelectronic and Computer Systems</source>
          <year>2022</year>
          (
          <year>2022</year>
          )
          <fpage>141</fpage>
          -
          <lpage>153</lpage>
          . doi:
          <volume>10</volume>
          .32620/reks.
          <year>2022</year>
          .
          <volume>1</volume>
          .11.
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>M.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <article-title>A novel finetuned yolov8 transfer learning model for smoking behavior detection</article-title>
          , in: 2024 International Conference on Image Processing,
          <source>Computer Vision and Machine Learning (ICICML)</source>
          , IEEE,
          <year>2024</year>
          , pp.
          <fpage>1944</fpage>
          -
          <lpage>1949</lpage>
          . doi:
          <volume>10</volume>
          .1109/icicml63543.
          <year>2024</year>
          .
          <volume>10957847</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Fu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Ran</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Xiao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Yuan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Zhao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>He</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Mei</surname>
          </string-name>
          ,
          <article-title>Gd-yolo: An improved convolutional neural network architecture for real-time detection of smoking and phone use behaviors</article-title>
          ,
          <source>Digital Signal Processing</source>
          (
          <year>2024</year>
          ). doi:
          <volume>10</volume>
          .1016/j.dsp.
          <year>2024</year>
          .
          <volume>104554</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>S.</given-names>
            <surname>Dalal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>U. K.</given-names>
            <surname>Lilhore</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Radulescu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Simaiya</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Jaglan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Sharma</surname>
          </string-name>
          ,
          <article-title>A hybrid lbp-cnn with yolov5- based fire and smoke detection model in various environmental conditions for environmental sustainability in smart city</article-title>
          ,
          <source>Environmental Science and Pollution Research</source>
          (
          <year>2024</year>
          ).
          <source>doi:10.1007/ s11356-024-32023-8.</source>
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>E.</given-names>
            <surname>Sezgin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. B.</given-names>
            <surname>Kocaballi</surname>
          </string-name>
          ,
          <article-title>The era of generalist conversational ai to support public health communications (preprint)</article-title>
          ,
          <source>Journal of Medical Internet Research</source>
          (
          <year>2024</year>
          ). doi:
          <volume>10</volume>
          .2196/69007.
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>T.</given-names>
            <surname>Hovorushchenko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Moskalenko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Osyadlyi</surname>
          </string-name>
          ,
          <article-title>Methods of medical data management based on blockchain technologies</article-title>
          ,
          <source>Journal of Reliable Intelligent Environments</source>
          <volume>9</volume>
          (
          <year>2023</year>
          )
          <fpage>5</fpage>
          -
          <lpage>16</lpage>
          . doi:
          <volume>10</volume>
          . 1007/s40860-022-00178-1.
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>R.</given-names>
            <surname>Gupta</surname>
          </string-name>
          , G. Bhatt,
          <string-name>
            <given-names>S.</given-names>
            <surname>Goel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Singh</surname>
          </string-name>
          ,
          <article-title>Prioritizing tobacco control &amp; its cessation under sustainable development goals with a focus on india</article-title>
          ,
          <source>Indian Journal of Medical Research</source>
          <volume>157</volume>
          (
          <year>2023</year>
          )
          <article-title>381</article-title>
          . doi:
          <volume>10</volume>
          .4103/ijmr.ijmr_
          <volume>3030</volume>
          _
          <fpage>21</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>I.</given-names>
            <surname>Krak</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Barmak</surname>
          </string-name>
          , E. Manziuk,
          <article-title>Using visual analytics to develop human and machine-centric models: a review of approaches and proposed information technology</article-title>
          ,
          <source>Computational Intelligence</source>
          <volume>38</volume>
          (
          <year>2022</year>
          )
          <fpage>921</fpage>
          -
          <lpage>946</lpage>
          . doi:
          <volume>10</volume>
          .1111/coin.12289.
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Feng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Hou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Hu</surname>
          </string-name>
          ,
          <article-title>An improved smoking behavior detection algorithm via incorporating an interference information filtering network</article-title>
          ,
          <source>Engineering Applications of Artificial Intelligence</source>
          <volume>136</volume>
          (
          <year>2024</year>
          ). doi:
          <volume>10</volume>
          .1016/j.engappai.
          <year>2024</year>
          .
          <volume>109050</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>A.</given-names>
            <surname>Khan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Khan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Hassan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Zheng</surname>
          </string-name>
          ,
          <article-title>Cnn-based smoker classification and detection in smart city application</article-title>
          ,
          <source>Sensors</source>
          <volume>22</volume>
          (
          <year>2022</year>
          )
          <article-title>892</article-title>
          . doi:
          <volume>10</volume>
          .3390/s22030892.
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>R.</given-names>
            <surname>Lakatos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Pollner</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Hajdu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Joó</surname>
          </string-name>
          ,
          <article-title>A multimodal deep learning architecture for smoking detection with a small data approach</article-title>
          ,
          <source>Frontiers in Artificial Intelligence</source>
          <volume>7</volume>
          (
          <year>2024</year>
          ). doi:
          <volume>10</volume>
          .3389/ frai.
          <year>2024</year>
          .
          <volume>1326050</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>A.</given-names>
            <surname>Khan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. A. M.</given-names>
            <surname>Elhassan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Khan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Deng</surname>
          </string-name>
          ,
          <article-title>Deep learning-based smoker classification and detection: An overview and evaluation</article-title>
          ,
          <source>Expert Systems with Applications</source>
          <volume>267</volume>
          (
          <year>2025</year>
          ). doi:
          <volume>10</volume>
          . 1016/j.eswa.
          <year>2024</year>
          .
          <volume>126208</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>D.</given-names>
            <surname>Murthy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. R.</given-names>
            <surname>Ouellette</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Anand</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Radhakrishnan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N. C.</given-names>
            <surname>Mohan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Kong</surname>
          </string-name>
          ,
          <article-title>Using computer vision to detect e-cigarette content in tiktok videos</article-title>
          ,
          <source>Nicotine &amp; Tobacco Research</source>
          <volume>26</volume>
          (
          <year>2024</year>
          )
          <fpage>S36</fpage>
          -
          <lpage>S42</lpage>
          . doi:
          <volume>10</volume>
          .1093/ntr/ntad184.
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>A. A. N. M.</given-names>
            <surname>Lavu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. A. I.</given-names>
            <surname>Jonayed</surname>
          </string-name>
          , M. T. Hossain,
          <article-title>Indoor smoking detection method based on dual spectral fusion image and yolo framework</article-title>
          , LC
          <source>International Journal of STEM 5</source>
          (
          <year>2024</year>
          )
          <fpage>13</fpage>
          -
          <lpage>35</lpage>
          . doi:
          <volume>10</volume>
          .5281/zenodo.14028770.
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>J.</given-names>
            <surname>Maurício</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Domingues</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Bernardino</surname>
          </string-name>
          ,
          <article-title>Comparing vision transformers and convolutional neural networks for image classification: A literature review</article-title>
          ,
          <source>Applied Sciences</source>
          <volume>13</volume>
          (
          <year>2023</year>
          )
          <article-title>5521</article-title>
          . doi:
          <volume>10</volume>
          .3390/app13095521.
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <given-names>C.</given-names>
            <surname>Meng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Lin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Gan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Ouyang</surname>
          </string-name>
          , Rts-vit:
          <article-title>Real-time share vision transformer for image classification</article-title>
          ,
          <source>IEEE Journal of Biomedical and Health Informatics</source>
          (
          <year>2025</year>
          )
          <fpage>1</fpage>
          -
          <lpage>12</lpage>
          . doi:
          <volume>10</volume>
          . 1109/jbhi.
          <year>2024</year>
          .
          <volume>3525054</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [24]
          <string-name>
            <given-names>M.</given-names>
            <surname>Jayamohan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Yuvaraj</surname>
          </string-name>
          ,
          <article-title>A novel human action recognition using grad-cam visualization with gated recurrent units</article-title>
          ,
          <source>Neural Computing and Applications</source>
          (
          <year>2025</year>
          ).
          <source>doi:10.1007/ s00521-025-10978-0.</source>
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          [25]
          <string-name>
            <given-names>O.</given-names>
            <surname>Kovalchuk</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Slobodzian</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Sobko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Molchanova</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Mazurets</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Barmak</surname>
          </string-name>
          , I. Krak,
          <string-name>
            <given-names>N.</given-names>
            <surname>Savina</surname>
          </string-name>
          ,
          <article-title>Visual analytics-based method for sentiment analysis of covid-19 ukrainian tweets</article-title>
          ,
          <source>in: Lecture Notes on Data Engineering and Communications Technologies</source>
          , volume
          <volume>149</volume>
          ,
          <year>2023</year>
          , pp.
          <fpage>591</fpage>
          -
          <lpage>607</lpage>
          . doi:
          <volume>10</volume>
          .1007/978-3-
          <fpage>031</fpage>
          -16203-9_
          <fpage>33</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          [26]
          <string-name>
            <given-names>J.</given-names>
            <surname>Narkhede</surname>
          </string-name>
          ,
          <article-title>Comparative evaluation of post-hoc explainability methods in ai: Lime, shap, and grad-cam</article-title>
          ,
          <source>in: 2024 4th International Conference on Sustainable Expert Systems (ICSES)</source>
          , IEEE,
          <year>2024</year>
          , pp.
          <fpage>826</fpage>
          -
          <lpage>830</lpage>
          . doi:
          <volume>10</volume>
          .1109/icses63445.
          <year>2024</year>
          .
          <volume>10762963</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          [27]
          <string-name>
            <given-names>E.</given-names>
            <surname>Manziuk</surname>
          </string-name>
          ,
          <string-name>
            <given-names>I.</given-names>
            <surname>Krak</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Barmak</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Mazurets</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Kuznetsov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Pylypiak</surname>
          </string-name>
          ,
          <article-title>Structural alignment method of conceptual categories of ontology and formalized domain</article-title>
          ,
          <source>in: CEUR Workshop Proceedings</source>
          , volume
          <volume>3003</volume>
          ,
          <year>2021</year>
          , pp.
          <fpage>11</fpage>
          -
          <lpage>22</lpage>
          . URL: http://ceur-ws.
          <source>org/</source>
          Vol-
          <volume>3003</volume>
          /paper2.pdf.
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          [28]
          <string-name>
            <given-names>S.</given-names>
            <surname>Kapadnis</surname>
          </string-name>
          , Smoking dataset,
          <year>2023</year>
          . URL: https://www.kaggle.com/datasets/sujaykapadnis/ smoking, accessed:
          <fpage>2025</fpage>
          -11-27.
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          [29]
          <string-name>
            <given-names>I.</given-names>
            <surname>Krak</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Sobko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Molchanova</surname>
          </string-name>
          ,
          <string-name>
            <given-names>I.</given-names>
            <surname>Tymofiiev</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Mazurets</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Barmak</surname>
          </string-name>
          ,
          <article-title>Method for neural network cyberbullying detection in text content with visual analytic</article-title>
          ,
          <source>in: CEUR Workshop Proceedings</source>
          , volume
          <volume>3917</volume>
          ,
          <year>2025</year>
          , pp.
          <fpage>298</fpage>
          -
          <lpage>309</lpage>
          . URL: https://ceur-ws.
          <source>org/</source>
          Vol-
          <volume>3917</volume>
          /paper57.pdf.
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          [30]
          <string-name>
            <given-names>M.</given-names>
            <surname>Molchanova</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Didur</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Mazurets</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Sobko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Zakharkevich</surname>
          </string-name>
          ,
          <article-title>Method for construction and demolition waste classification using two-factor neural network image analysis</article-title>
          ,
          <source>in: CEUR Workshop Proceedings</source>
          , volume
          <volume>3970</volume>
          ,
          <year>2025</year>
          , pp.
          <fpage>168</fpage>
          -
          <lpage>182</lpage>
          . URL: https://ceur-ws.
          <source>org/</source>
          Vol-
          <volume>3970</volume>
          /PAPER14.pdf.
        </mixed-citation>
      </ref>
      <ref id="ref31">
        <mixed-citation>
          [31]
          <string-name>
            <given-names>J. D.</given-names>
            <surname>Hunter</surname>
          </string-name>
          ,
          <article-title>Matplotlib: A 2d graphics environment</article-title>
          ,
          <source>Computing in Science &amp; Engineering</source>
          <volume>9</volume>
          (
          <year>2007</year>
          )
          <fpage>90</fpage>
          -
          <lpage>95</lpage>
          . doi:
          <volume>10</volume>
          .1109/
          <string-name>
            <surname>MCSE</surname>
          </string-name>
          .
          <year>2007</year>
          .
          <volume>55</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref32">
        <mixed-citation>
          [32]
          <string-name>
            <given-names>G.</given-names>
            <surname>Bradski</surname>
          </string-name>
          , The OpenCV Library, Dr.
          <source>Dobb's Journal of Software Tools</source>
          (
          <year>2000</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref33">
        <mixed-citation>
          [33]
          <string-name>
            <given-names>I.</given-names>
            <surname>Krak</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Zalutska</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Molchanova</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Mazurets</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Manziuk</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Barmak</surname>
          </string-name>
          ,
          <article-title>Method for neural network detecting propaganda techniques by markers with visual analytic</article-title>
          ,
          <source>in: CEUR Workshop Proceedings</source>
          , volume
          <volume>3790</volume>
          ,
          <year>2024</year>
          , pp.
          <fpage>158</fpage>
          -
          <lpage>170</lpage>
          . URL: https://ceur-ws.
          <source>org/</source>
          Vol-
          <volume>3790</volume>
          /paper14.pdf.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>