<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Method for Construction and Demolition Waste Classification Using Two-Factor Neural Network Image Analysis ⋆</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Zakharkevich</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Khmelnytskyi National University</institution>
          ,
          <addr-line>11, Instytuts'ka str., Khmelnytskyi, 29016</addr-line>
          ,
          <country country="UA">Ukraine</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Method for construction and demolition waste classification using two-factor neural network image analysis has been developed, which allows increasing the classification accuracy to 97.8% for 10 classes of construction waste. High accuracy of object recognition and prevention of misclassifications is achieved due to use of two-factor approach to identification and classification of the construction and demolition waste. Approach consists in combining the use of YOLO11 model for object identification and set of binary ResNet50V2 classifiers for the classification of each of the 10 classes of construction and demolition waste. Method takes into account the results of binary trained neural networks for the classification of construction and demolition waste, and in cases of ambiguous classification, it additionally uses the prediction of YOLO11 arbiter neural network. To ensure the diversity of training dataset and improve the generalization capabilities of classifiers, dataset was expanded using transformations such as rotation, scaling, color change, and reflection; this also significantly increased the accuracy of separation and classification of construction and demolition waste. Comparison of the developed method and known analogues revealed minimum increase in accuracy of 5.2%, maximum increase in accuracy of 32.5%.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;construction and demolition waste</kwd>
        <kwd>construction waste sorting</kwd>
        <kwd>classification</kwd>
        <kwd>ResNet50</kwd>
        <kwd>YOLO111</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>waste
management methods, often
accompanied by low levels of sorting and disposal, are one of the main causes of environmental
degradation, water, air and soil pollution [5]. This creates risks of geological instability, increases
the number of dangerous zones, such as landfills, and contributes to the spread of diseases among
the population due to exposure to toxic substances. In addition, inefficient waste management
exacerbates the problem of depletion of natural resources, as many materials that could be recycled
end up in landfills [6, 7].</p>
      <p>In this context, recycling of such waste appears as an important mechanism to reduce the
volume of landfills, while reducing the need for energy and natural resources required for the
extraction and production of new materials [8, 9]. Innovative solutions such as neural network
technologies, automated sorting systems and recycling infrastructure aim to minimize these
negative impacts, ensuring not only the protection of the ecosystem, but also transition to a
circular economy [10, 11].</p>
      <p>The application of neural network technologies in the classification of construction and
demolition waste contributes to the achievement of the UN Sustainable Development Goals, in
particular sustainable urban development (SDG 11), responsible consumption (SDG 12), sustainable
infrastructure development (SDG 9) and combating climate change (SDG 13) through effective
waste management, increased recycling rates and reduced emissions [12, 13]. The aim of the study
is to develop a method for classifying the remains of destroyed buildings and construction waste
using neural network image analysis, which will allow to increase the accuracy of classification.
The main contributions of the article can be summarized as follows:


</p>
      <p>Method for construction and demolition waste classification using two-factor neural
network image analysis is proposed, which is based on combination of binary classifiers
and the YOLO neural network, which allows to increase the accuracy of multi-class
classification.</p>
      <p>Approach to refine predictions in case of classification ambiguity is developed, which
reduces the number of errors and ensures identification accuracy at the level of 97.8% for 10
classes of construction waste.</p>
      <p>Experimental validation of the method is carried out, which confirmed its effectiveness in
real conditions with heterogeneous data.</p>
      <p>The next section presents an overview of related works in the field of construction waste and
production waste. Section 4 offers an overview of the experiment to investigate the effectiveness of
the proposed approach. Section 5 offers an overview of the obtained results with comparisons and
discussions. The last section is conclusions.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related Works</title>
      <p>Studies have shown that construction and demolition waste (C&amp;DW) constitutes a significant
proportion of municipal solid waste (MSW) accumulating in landfills worldwide, posing a
significant environmental challenge [14].</p>
      <p>The study [15] is devoted to the development of a highly efficient method for fine-grained
classification of construction and demolition waste (C&amp;DW), which is an important step for the
development of a recycling system and achieving carbon neutrality in the waste management
sector. C&amp;DWNet models built on ResNet using knowledge transfer and a cyclic learning
coefficient were proposed to classify ten types of construction waste. The results showed that the
knowledge transfer method can reduce the training time and improve the efficiency of the model.
The average training time increased with the number of layers of the architecture: from
C&amp;DWNet-18 (946.7 s) to C&amp;DWNet-152 (1186.6 s). The best Accuracy (73.6%) was achieved on
C&amp;DWNet-152, however, the C&amp;DWNet-18 model showed the optimal balance between training
time, accuracy, precision and F1-metric. The use of t-distributed stochastic neighbor embedding
allowed to clearly distinguish the types of waste.</p>
      <p>In [16], an approach to construction waste management is presented using a robotic system to
automatically sort and collect construction materials, such as nails and screws, which are difficult
to detect manually. The robot uses neural network and computer vision technologies, as well as the
Faster R-CNN method for real-time detection and collection of target objects in the work
environment.</p>
      <p>In the study [17], deep learning approaches for effective waste identification and classification
based on an open set of images from the Middle East region were studied. Images of garbage and
waste are divided into five categories: paper, plastic, glass, metal and other. Each image contains
only one class. Convolutional neural network (CNN) algorithms were used to classify waste, as
well as pre-trained models MobileNetV2, ResNet50V2, and DenseNet169. The highest accuracy
among the pre-trained models was demonstrated by ResNet50V2 (98.95%), while the proposed CNN
model achieved an accuracy of 88.5%, which exceeds the results of previous studies on the same
dataset.</p>
      <p>The paper [18] describes the development of a methodology to improve the sorting of
construction and demolition waste (BDW) using machine learning, using an RGB camera to
recognize waste fragments. The main goal of the research is to improve the efficiency of waste
sorting through the use of advanced feature extraction methods, which improves the speed and
accuracy of classification. The paper compares three classifiers: convolutional neural network
(CNN), gradient boosting decision trees (GB), and multilayer perceptron (MLP), where the feature
extraction method for GB and MLP showed better results in terms of speed and accuracy compared
to the traditional CNN. The results show that the new methodology provides accuracy up to 92.3%,
compared to 85.9% for CNN. The paper also provides additional materials, including datasets,
codes, and models, which contributes to the transparency and reproducibility of the results.</p>
      <p>In [19], a method for detecting construction and demolition waste (CDW) in urban development
is considered, taking Beijing as an example. Given the rapid demolition of old buildings, waste has
become an important component of the urban pollution problem. Since CDW landfills have
unstable contours, it is important to determine their location in a timely and accurate manner to
achieve accurate mapping and effective waste management. A method based on change detection
and deep learning was proposed to solve this problem. ZY-3 multispectral images from 2016 and
2019, as well as difference images obtained using change detection methods, were used to prepare
the initial samples. Sample expansion using the post-classification method allowed to increase the
sample by 25.4%, which improved the results. This extended learning environment was used to
train the DeepLabV3+ model, and the digital terrain model information was also used to
distinguish different types of CDW, such as demolition waste, landfills, and large-scale dumps. The
CDW detection accuracy was 91.67%, and the Kappa coefficient was 0.8642. Comparison of the
results with the original samples showed that expanding the sample using change data improves
the accuracy of the deep learning models, which is also confirmed by the results for the PSPNet and
UNet models. This study is the first to effectively distinguish the three main forms of CDW and
significantly reduce the errors in the classification of CDW and bare land.</p>
      <p>The literature review identified several key issues in the classification of construction and
demolition waste. First, most works are focused on classification within already defined objects,
which is a simplified task, since one image may contain different fragments, often belonging to
different classes, and objects may be superimposed on each other. Second, in multi-class
classification, there is still low identification accuracy, which complicates the effective solution of
the problem.</p>
    </sec>
    <sec id="sec-3">
      <title>3. Method Design</title>
      <p>The proposed approach allows to increase the accuracy of classification of construction and
demolition waste by applying the developed method, which converts input data in the form of a
photo image for analysis, a trained neural network for highlighting fragments of construction
remains in the photo image and their basic classification, and a set of trained neural networks for
binary classification of fragments of construction remains into output data in the form of an image
with highlighted fragments of construction remains and defined classes with probabilities of their
belonging. The scheme of the method for construction and demolition waste classification using
two-factor neural network image analysis is shown in Fig. 1.</p>
      <p>The input data of the method are photo images for analysis, a trained neural network for
extracting fragments of construction remains in the photo image, and trained neural networks for
binary classification of the selected fragments. YOLO11 [20] was chosen as the neural network for
extracting fragments of construction remains in the photo image, and the trained neural networks
for binary classification of the selected fragments have ResNet50 architectures [21].</p>
      <p>For neural network extraction of objects in the image, the step 1 of preprocessing of the photo
image first occurs [22]. The image is scaled to a size of 640x640 pixels. After scaling, the step 2 of
extracting fragments of construction remains in the photo image directly occurs.</p>
      <p>For each selected fragment, a preprocessing stage is also performed for the possibility of further
classification, which includes scaling to a size of 224x224 (while preserving the aspect ratio) and
normalizing the image by subtracting the mean value and dividing by the standard deviation [23]</p>
      <p>At step 3, each selected fragment is classified by each of the 10 trained neural networks for
classifying fragments of construction remains in the photo image. The following classes are
classified in the study: brick, concrete, foam, plastic, general waste, gypsum board, pipes, plastic,
stone, tile, wood. Examples of fragments are shown in Fig. 2 (a-k). In the case where the neural
networks give the same results, step 4 is performed, which consists of additional verification by an
arbitrator. The arbitrator is the YOLO11 classification.</p>
      <p>As a result, the output data are the selected fragments of construction remains in the photo and
their defined names with membership probabilities.</p>
      <p>The developed method was used in this study to analyze separate frames, but it can be
integrated into automated demolition waste sorting systems and applied to analyze a video stream
in real time.</p>
      <p>a) brick
b) concrete
c) foam
d) general waste</p>
      <sec id="sec-3-1">
        <title>e) gypsum board</title>
      </sec>
      <sec id="sec-3-2">
        <title>f) pipes</title>
      </sec>
      <sec id="sec-3-3">
        <title>g) plastic</title>
      </sec>
      <sec id="sec-3-4">
        <title>h) stone j) tile k) wood</title>
        <p>Since one of the input data is a set of ResNet50 neural network models for binary classification,
Figure 3 shows an example of obtaining a neural network model for brick identification. A similar
approach is used to obtain trained models for the remaining 9 classes.</p>
        <p>First, to train binary ResNet50 models from the general dataset for classifying different types of
construction waste, subdatasets are formed. All fragments of the subdataset are scaled to 224x224
size while preserving the aspect ratio and converted to RGB format. Since the key problem of
multi-class classification is the reduction of identification accuracy, it was decided to use 10 neural
network classifiers, separate for each type of construction waste. Datasets are formed according to
the rule: all images from the general dataset that correspond to the target category are transferred
to the subdataset in the target category. The opposite category of the subdataset is formed by
combining in equal parts other types of construction waste and construction residues. Further, each
of the fragments is rotated by 45 degrees 4 times, 4 new samples are formed, and for each new
inverted sample, a random color change is applied according to the brightness and contrast
parameters. Thus, an extended subdataset is formed.</p>
        <p>ResNet50 models are trained on the obtained subdatasets, which are validated by the metrics
Accuracy, Precision, Recall, F1, after which a typical binary classifier model is saved.</p>
        <p>So, a method is proposed that, instead of using the standard multi-class classification, which is
typical for the YOLO neural network, uses a set of binary classifiers, each of which determines the
belonging of the object to one of the classes. The method takes into account the results of binary
trained neural networks for classifying fragments of construction remains, and in cases of
ambiguity (equality of estimates between classes), it additionally uses the YOLO forecast. In
situations where YOLO and one of the binary classifiers coincide in the estimate, this class is
determined as dominant. This approach reduces the number of erroneous classifications and
provides higher accuracy of object recognition.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Experiment</title>
      <p>To research the effectiveness of proposed approach, the software was developed (Fig. 4) that uses
pre-trained YOLO11 model and set of ResNet50 models for finding and classifying objects in an
image. The PySide [24], TensorFlow [25], and Ultralytics [26] libraries were used to create the
software.</p>
      <p>The following datasets were used to train the neural networks, which are the input data of the
proposed method and are used in the application:</p>
      <p>1) “Construction and Demolition Waste Object Detection Dataset” [27]. The dataset contains
3,129 high-resolution images (1920 × 1200 × 3, RGB) with 16,545 annotated samples covering 10
waste categories. The data were collected at a waste sorting plant in Cyprus. The annotations are
provided in PASCAL VOC XML format, including the coordinates of bounding boxes and polygons
for accurate object detection. The dataset is divided into subsets for training, validation, and
testing.</p>
      <p>2) “Dataset-of-stacked-construction-solid-waste” [28] to supplement the images of concrete and
wood. The dataset contains 4 types of solid construction waste: concrete, brick, wood, and rubber.
Each document contains 25 sample images and label files.</p>
      <p>3) “RGB-D construction solid waste” [29]. An instance segmentation dataset for solid waste
identification. The dataset includes 4 waste types: concrete, rubber, wood, and brick. It is also used
for supplementation.</p>
      <p>From the considered datasets, a general dataset was formed, which has the following structure
and dimension: “brick” – 1370 samples, “concrete” – 1407 samples, “foam” – 746 samples,
“general_w” – 742 samples, “gypsum_board” – 1184 samples, “pipes” – 715 samples, “plastic” – 675
samples, “stone” – 744 samples, “tile” – 1344 samples, “wood” – 1131 samples. This formed dataset
was used to create subdatasets for training binary ResNet50 classifiers.</p>
      <p>For further training of YOLO11, the following distribution was used: “brick” – 1370 samples
contained in 309 files; “concrete” – 1407 samples contained in 311 files; “foam” – 746 samples
contained in 126 files; “general_w” – 742 samples contained in 359 files; “gypsum_board” – 1184
samples contained in 126 files; “pipes” – 715 samples contained in 196 files; “plastic” – 675 samples
contained in 265 files; “stone” – 744 samples contained in 177 files; “tile” – 1344 samples contained
in 308 files; “wood” – 1131 samples contained in 504 files.</p>
      <p>To compare the effectiveness of the studied neural networks, the metrics Accuracy, Precision,
Recall, F1-Score [30] and specialized metrics were used, which allowed for a comprehensive
assessment of the proposed approaches. The effectiveness study was conducted in the following
areas:
study of the impact of the elapsed training epochs on the accuracy of binary classifiers;
comparison of the ResNet50 and MobileNetV2 architectures;
comparison of the efficiency of YOLO11 for classifying production waste objects with the
developed method.






</p>
    </sec>
    <sec id="sec-5">
      <title>5. Results and discussion</title>
      <p>To identify construction and demolition waste objects, pre-trained neural network YOLO11 was
used, which was further trained on the above-described dataset. This neural network also acts as an
arbiter to confirm the predictions of individual neural network binary models [31].</p>
      <p>During training YOLO11, the following transformations are performed on the input data:
horizontal mirroring;
random scaling;
random shifting;
random color changes (brightness, contrast).</p>
      <p>These transformations occur over the full image, the fragments are not processed separately.
Such transformations allow to increase the correctness of the selection of construction and
demolition waste fragments.</p>
      <p>To classify the selected fragments, a pre-trained ResNet50 model was taken, which is part of the
TensorFlow library and was further trained on subdatasets, separate for each of the 10 models. To
study the influence of the passed training epochs on the accuracy of binary classifiers, each of the
binary neural network models ResNet50 was trained on different numbers of epochs – from 5 to 20.
To train ResNet50, the following manipulations are performed on each fragment:


the fragments are rotated 45 degrees 4 times, creating 4 new samples;
a random color change (brightness and contrast) is applied to the new rotated sample.</p>
      <p>Before that, the fragments are scaled to 224x224 while preserving the aspect ratio (the original
image is centered, the "empty space" is filled with black) and converted to RGB format, if it was
BGR or RGBA. The results of the experiments are given in Table 1.</p>
      <p>The data in Table 1 show that increasing the number of training epochs has a positive effect on
the performance metrics for most classes. Larger number of epochs allows the model to better
adapt to the patterns in the data, although after a certain point the effect can fade or even worsen
due to overfitting [32], as seen in the classes "Gypsum Board" and "Plastic". The metrics can also be
affected by the characteristics of the data itself [33], such as the distribution between classes, the
complexity of the samples, and the level of noise. Thus, choosing the optimal number of epochs is
critical and depends on the specifics of the task and the data [34].</p>
      <p>To compare the results, a similar experiment was conducted with the MobileNetV2 architecture
[35], which is also a pre-trained neural network model. The training results are presented in Table
2.</p>
      <p>The impact of training epochs number on metrics for the MobileNet neural network-Score is
observed at the 20th epoch, while for the "Plastic" class, the qualitative indicators remain lower,
regardless of the number of epochs, probably due to the more complex data structure.</p>
      <p>The stability of the metrics at the 15th and 20th epochs for many classes indicates that the
network achieves optimal learning, although some fluctuations, for example in the "Pipes" class,
may be caused by overfitting or insufficient generalization. The overall result indicates the
importance of fine-tuning the number of epochs for each specific class to avoid performance loss.</p>
      <p>A comparison of the best results from Tables 1 - 2 of alternative neural network options for
binary classification is shown in Figure 5.</p>
      <p>The results of the performance comparison of the ResNet50 and MobileNetV2 models
demonstrate significant differences, which are due to the architectural features of these neural
networks. ResNet50, being a deeper and more complex model, shows higher Accuracy indicators
and harmoniously balanced Precision, Recall and F1-Score for most classes, in particular for
"Foam", "Stone" and "Wood". This indicates its ability to effectively generalize patterns even in
complex data.</p>
      <p>MobileNetV2, on the contrary, is focused on optimizing computational resources and has less
depth, which explains its lower performance in the cases of complex classes such as "Plastic" or
"Tile". Its advantages become noticeable in tasks with less complex data, where it is able to achieve
acceptable metric values with minimal resource costs.</p>
      <p>The pre-trained YOLO11 neural network is also capable of performing not only the separation
of construction debris and construction residues, but also the identification of the separated objects.
After further training on the specified dataset, the neural network obtained the metric values
shown in Table 3. The metrics results for YOLO11 demonstrate significant variability in
classification performance for different classes of construction and demolition waste, confirming
the feasibility of implementing binary classifiers to improve accuracy and consistency of results.
The macrometrics show a moderate overall Accuracy of 0.601 and F1-Score of 0.7508, however,
there is a clear gap between high performance for some classes (e.g., "brick" and "concrete") and
significantly lower results for others (e.g., "foam", "pipes", "plastic"). Most critically, for classes such
as "foam" (Accuracy 0.373, F1-Score 0.544), "pipes" (Accuracy 0.289, F1-Score 0.45) and "plastic"
(Accuracy 0.294, F1-Score 0.455), Recall values are either not high enough or excessively exceed
Precision. This indicates problems with the differentiation of these classes, which may be due to
overlapping features between classes, imbalance in the amount of data, or the complexity of the
objects themselves to be classified.</p>
      <p>The introduction of binary classifiers allows you to optimize the performance for each class
separately, adapting the model to the specifics of each type of object. For example, the use of
separate binary classifiers for "plastic" and "pipes" allowed you to focus on the characteristic
features of these classes, minimizing the impact of the heterogeneity of features of other classes in
the overall model. This, in turn, improved the Recall and F1-Score for classes with low
performance, while increasing the overall level of consistency of the metrics.</p>
      <p>The mAP50 and mAP50-95 metrics are used to evaluate the quality of the YOLO11 model in
classifying and localizing construction debris and construction waste objects in images. The overall
mAP50 score is 0.746, indicating a high ability of the model to detect objects at a moderate
threshold of match between the predicted and actual frame (IoU ≥ 50%). At the same time,
mAP5095 with an average value of 0.669 shows reduced performance under stricter evaluation conditions
that take into account localization accuracy at higher IoU thresholds. This difference is expected,
since mAP50-95 is a more stringent metric that reflects the ability of the model to accurately
localize objects. Analysis of the results by individual classes shows that for objects with clear visual
features, such as "brick", "concrete", "gypsum board" and "tile", the model demonstrates
consistently high performance, even under strict conditions. This indicates a reliable ability of
YOLO11 to recognize objects with minimal variations in shape or texture. At the same time, classes
with heterogeneous characteristics, such as "plastic", "pipes" and "stone", are characterized by
significantly lower indicators. This indicates the need for optimization, in particular, increasing the
number of training samples, improving data quality or applying specialized approaches to their
processing.</p>
      <p>However, despite the classification problems, this neural network copes with the task of object
selection at a fairly high level, as evidenced by the MioU metric of 0.897.</p>
      <p>Regarding the task of classifying the selected objects, a comparison of the obtained metrics for
YOLO11 and ResNet50 and their delta results is given in Table 4. A comparison of the metrics of
YOLO11 and ResNet50 shows significant advantages of ResNet50 in object classification accuracy
and overall model performance. The data in the table shows that ResNet50 significantly
outperforms YOLO11 in all key metrics – Accuracy, Precision, Recall and F1-Score. The largest
gains are observed in metrics for classes such as "foam", "pipes" and "plastic", where the difference
reaches more than 0.6, indicating significant improvements in the ability of ResNet50 to correctly
classify these objects and determine their belonging to the corresponding classes.</p>
      <p>The average value of the indicators (MACRO) also demonstrates significant advantages of
ResNet50, where the overall increase in Accuracy is 0.374, Precision – 0.169, Recall – 0.272, and
F1Score – 0.223. This indicates that ResNet50 is not only more accurate in class detection, but also
better at reducing missed (False Negative) or incorrectly identified (False Positive) objects.</p>
      <p>Particular attention should be paid to the classes with low performance of YOLO11, such as
"foam", "pipes", "plastic" and "stone". For these classes, YOLO11 shows significantly lower
Accuracy, which may be due to the difficulty of object detection due to non-uniform texture or
shape features. ResNet50 significantly improves the results for these classes, indicating its ability to
better capture object features even in complex cases.</p>
      <p>However, for some classes, such as "brick", "gypsum board" and "stone", the Precision of
YOLO11 exceeds ResNet50 or has minimal loss. This may indicate that YOLO11 does a good job of
identifying specific, well-defined objects, although the overall performance of the model is lost due
to low Recall, which indicates frequent omissions. That is why this model was used for
classification only as an arbiter neural network, and its main role is to select objects.</p>
      <p>A comparison with existing scientific research was also made. Comparing the results obtained
with a similar study of the classification of construction debris and construction waste conducted
in [15], an improvement in Accuracy to 0.975 was achieved, compared to the authors' value of 0.736
for classification by ten classes. In the study [17], ResNet50V2 was also used, which achieved an
Accuracy of almost 0.99, however, the study was conducted only for five classes, and the objects in
the images were contained one at a time, the authors did not conduct a study of object
identification, but only classification. In the study [18], the classification of construction waste into
four classes was carried out and the Accuracy of 0.923 was achieved, which is lower than the
results obtained by the proposed method. The proposed method also performed better than in [19]
(the authors obtained Accuracy of 0.917).</p>
      <p>Thus, the proposed implementation of binary classifiers is reasonable approach to improve the
quality of classification in tasks with high variability of performance between classes.</p>
    </sec>
    <sec id="sec-6">
      <title>6. Conclusions</title>
      <p>The method for construction and demolition waste classification using two-factor neural network
image analysis has been developed, which allows increasing the classification accuracy to 97.8% for
10 classes of construction and demolition waste. This effect is achieved by using a two-factor
approach to identifying and classifying the remains of destroyed buildings and construction waste,
which consists in combining the use of YOLO11 for object identification and as a neural
networkarbitrator and a set of binary classifiers ResNet50V2 for classifying each of the 10 classes of
construction waste. The method takes into account the results of binary trained neural networks
for classifying fragments of construction remains, and in cases of ambiguity (equality of estimates
between classes), it additionally uses the prediction of the YOLO11 neural network-arbitrator. In
situations where YOLO11 and one of the binary classifiers coincide in the estimate, this class is
determined as dominant. This approach reduces the number of false classifications and provides
higher object recognition accuracy.</p>
      <p>As data for further training of YOLO11 and binary classifiers ResNet50V2, a composite dataset
based on the datasets "Construction and Demolition Waste Object Detection Dataset",
"Dataset-ofstacked-construction-solid-waste" and "RGB-D construction solid waste" was used.</p>
      <p>The use of transformations such as rotation, scaling, color change and reflection significantly
increases the accuracy of construction waste separation and classification. This provides a variety
of training dataset and promotes generalization of models. Comparison of the developed method
and known analogues showed an increase in accuracy in the range from 5.2% to 32.5%, reaching an
indicator of up to 97.8%.</p>
      <p>The developed method contributes to the achievement of the UN Sustainable Development
Goals, in particular SDG 11, SDG 12, SDG 9 and SDG 13 through effective waste management,
increased recycling and reduced emissions. The presented approach can be used to automate the
sorting of construction waste, which will reduce time and increase the efficiency of waste
processing. Further development includes refining the models to better handle complex classes
such as "Plastic" and "General Waste", as well as implementing active learning mechanisms to
adapt to new data. Another direction for further research is method's integration into automated
waste sorting systems and real-time analysis of the method's performance. The priority direction of
further research is to expand the dataset with additional image classes and scaling developed
method up to a wider range of building materials.</p>
    </sec>
    <sec id="sec-7">
      <title>Declaration on Generative AI</title>
      <p>The authors have not employed any Generative AI tools.
[3] R. Kazmi, M. Chakraborty, Use of different forms of waste promoting a circular economy in
the construction industry: A systematic literature review, in: International Conference on
Environ-mental Geotechnology, Recycled Waste Materials and Sustainable Engineering,
Springer, Singa-pore, 2025, pp. 193–204. doi:10.1007/978-981-97-9777-6_17
[4] S. Jha, S. Nanda, O. Zapata, B. Acharya, A. K. Dalai, A review of systems thinking perspectives
on sustainability in bioresource waste management and circular economy, Sustainability 16
(2024) 10157. doi:10.3390/su162310157.
[5] F. R. Sayem, M. S. B. Islam, M. Naznine, M. Nashbat, M. Hasan-Zia, A. K. A. Kunju, M. E.</p>
      <p>Chow-dhury, Enhancing waste sorting and recycling efficiency: Robust deep learning-based
approach for classification and detection, Neural Computing and Applications (2024) 1–17.
doi:10.1007/s00521-024-10855-2.
[6] D. Caro, C. Lodato, A. Damgaard, J. Cristóbal, G. Foster, F. Flachenecker, D. Tonini,
Environmental and socio-economic effects of construction and demolition waste recycling in
the European Un-ion, Science of the Total Environment 908 (2024) 168295.
doi:10.1016/j.scitotenv.2023.168295.
[7] D. V. Yevle, P. S. Mann, Artificial intelligence-based classification for waste management: A
sur-vey based on taxonomy, classification &amp; future direction, Computer Science Review 56
(2025) 100723. doi:10.1016/j.cosrev.2024.100723.
[8] Y. Lu, Y. Ge, G. Zhang, A. Abdulwahab, A. A. Salameh, H. E. Ali, B. N. Le, Evaluation of waste
man-agement and energy saving for sustainable green building through analytic hierarchy
process and artificial neural network model, Chemosphere 318 (2023) 137708.
doi:10.1016/j.chemosphere.2022.137708.
[9] N. N. I. Prova, Garbage intelligence: Utilizing vision transformer for smart waste sorting, in:
2024 Second International Conference on Intelligent Cyber Physical Systems and Internet of
Things (ICoICI), IEEE, 2024, pp. 1213–1219. doi:10.1109/ICoICI62503.2024.10696177.
[10] M. Farghali, A. I. Osman, Revolutionizing waste management: Unleashing the power of
artificial intelligence and machine learning, in: Advances in Energy from Waste, Woodhead
Publishing, 2024, pp. 225–279. doi:10.1016/B978-0-443-13847-8.00007-5
[11] R. Alsabt, W. Alkhaldi, Y. A. Adenle, H. M. Alshuwaikhat, Optimizing waste management
strate-gies through artificial intelligence and machine learning – An economic and
environmental im-pact study, Cleaner Waste Systems 8 (2024) 100158.
doi:10.1016/j.clwas.2024.100158
[12] Z. Zhou, D. Syamsunur, X. Fan, J. Li, A preliminary investigation on the relationship between
the sustainable construction field and the Sustainable Development Goals: A proposed
research framework, in: E3S Web of Conferences, vol. 617, EDP Sciences, 2025, p. 03014.
doi:10.1051/e3sconf/202561703014.
[13] M. A. W. M. A. Wibowo, M. A. A. M. A. Ammar, Lean construction and sustainability: A
review of research trends and implications for the United Nations SDGs, in: E3S Web of
Conferences, vol. 605, EDP Sciences, 2025, p. 03048. doi:10.1051/e3sconf/202560503048
[14] D. Sirimewan, M. Bazli, S. Raman, S. R. Mohandes, A. F. Kineber, M. Arashpour, Deep
learning-based models for environmental management: Recognizing construction, renovation,
and demo-lition waste in-the-wild, Journal of Environmental Management 351 (2024) 119908.
doi:10.1016/j.jenvman.2023.119908.
[15] K. Lin, Y. Zhao, T. Zhou, X. Gao, C. Zhang, B. Huang, Q. Shi, Applying machine learning to
fine clas-sify construction and demolition waste based on deep residual network and
knowledge transfer, Environment, Development and Sustainability 25 (2023) 8819–8836.
doi:10.1007/s10668-022-02740-6.
[16] Z. Wang, H. Li, X. Zhang, Construction waste recycling robot for nails and screws: Computer
vi-sion technology and neural network approach, Automation in Construction 97 (2019) 220–
228. doi:10.1016/j.autcon.2018.11.009
[17] M. I. B. Ahmed, R. B. Alotaibi, R. A. Al-Qahtani, R. S. Al-Qahtani, S. S. Al-Hetela, K. A.
AlMatar, G. Krishnasamy, Deep learning approach to recyclable products classification: Towards
sustainable waste management, Sustainability 15 (2023) 11138. doi:10.3390/su151411138.
[18] V. Nežerka, T. Zbíral, J. Trejbal, Machine-learning-assisted classification of construction and
demolition waste fragments using computer vision: Convolution versus extraction of selected
features, Expert Systems with Applications 238 (2024) 121568. doi:10.1016/j.eswa.2023.121568.
[19] X. Zhao, Y. Yang, F. Duan, M. Zhang, G. Jiang, X. Yan, W. Zhao, Identification of construction
and demolition waste based on change detection and deep learning, International Journal of
Remote Sensing 43 (6) (2022) 2012–2028. doi:10.1080/01431161.2022.2054296.
[20] YOLOv11 Model Documentation, 2024. URL: https://docs.ultralytics.com/models/yolo11.
[21] ResNet50 Model Documentation, 2024. URL:
https://pytorch.org/vision/main/models/generated/torchvision.models.resnet50.html.
[22] S. Sonawane, N. N. Patil, A Review of Pre-processing Techniques for Weed-Plant Detection
and Classification in Precision Agriculture, Lecture Notes in Electrical Engineering 1146 (2024)
401–417. doi:10.1007/978-981-97-0037-0_24.
[23] P. Dehbozorgi, O. Ryabchykov, T. Bocklitz, A Systematic Investigation of Image
PreProcessing on Image Classification, IEEE Access 12 (2024) 64913–64926.
doi:10.1109/ACCESS.2024.3395063.
[24] PySide6, 2024. URL: https://pypi.org/project/PySide6.
[25] TensorFlow, 2024. URL: https://www.tensorflow.org.
[26] Ultralytics, 2024. URL: https://www.ultralytics.com.
[27] Construction and Demolition Waste Object Detection Dataset, 2024. URL:
https://data.mendeley.com/datasets/wds85kt64j/3.
[28] Dataset-of-stacked-construction-solid-waste, 2024. URL:
https://github.com/Bazenr/Dataset-ofstacked-construction-solid-waste.
[29] RGB-D construction solid waste, 2024. URL:
https://www.kaggle.com/datasets/bazenr/stackedconstruction-solid-waste.
[30] I. Krak, O. Zalutska, M. Molchanova, O. Mazurets, E. Manziuk and O. Barmak, Method for
neural network detecting propaganda techniques by markers with visual analytic, in: CEUR
Workshop Proceedings, 3790 (2024) pp.158-170. URL:
https://ceur-ws.org/Vol-3790/paper14.pdf.
[31] I. Krak, V. Didur, M. Molchanova, O. Mazurets, O. Sobko, O. Zalutska and O. Barmak, Method
for political propaganda detection in internet content using recurrent neural network models
en-semble, in: CEUR Workshop Proceedings, 3806 (2024) pp.312-324. URL:
https://ceurws.org/Vol-3806/S_36_Krak.pdf.
[32] I. Krak, O. Zalutska, M. Molchanova, O. Mazurets, R. Bahrii, O. Sobko and O. Barmak, Abusive
speech detection method for Ukrainian language used recurrent neural network, in: CEUR
Work-shop Proceedings, 3688 (2024) pp.16-28. doi: 10.31110/COLINS/2024-3/002.
[33] O. Sobko, O. Mazurets, M. Molchanova, I. Krak and O. Barmak, 2025. Method for analysis and
formation of representative text datasets, in: CEUR Workshop Proceedings, 3899 (2024)
pp.8498. URL: https://ceur-ws.org/Vol-3899/paper9.pdf.
[34] E. Manziuk, I. Krak, O. Barmak, O. Mazurets, V. Kuznetsov and O. Pylypiak, 2021. Structural
alignment method of conceptual categories of ontology and formalized domain, in: CEUR
Work-shop Proceedings, 3003, pp.11–22. URL: http://ceur-ws.org/Vol-3003/paper2.pdf.
[35] MobileNetV2,2024.URL:
https://www.tensorflow.org/api_docs/python/tf/keras/applications/MobileNetV2.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>K.</given-names>
            <surname>Hu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. T. A.</given-names>
            <surname>Gillani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Tao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Tariq</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <article-title>Eco-friendly construction: Integrating demolition waste into concrete masonry blocks for sustainable development</article-title>
          ,
          <source>Construction and Building Materials</source>
          <volume>460</volume>
          (
          <year>2025</year>
          )
          <article-title>139797</article-title>
          . doi:
          <volume>10</volume>
          .1016/j.conbuildmat.
          <year>2024</year>
          .
          <volume>139797</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Gao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <article-title>Machine learning in construction and demolition waste management: Pro-gress, challenges, and future directions</article-title>
          ,
          <source>Automation in Construction</source>
          <volume>162</volume>
          (
          <year>2024</year>
          )
          <article-title>105380</article-title>
          . doi:
          <volume>10</volume>
          .1016/j.autcon.
          <year>2024</year>
          .
          <volume>105380</volume>
          ..
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>