<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>A Distance-metric Approach to Expert Agreement In Multi-level Publications Classification⋆</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Ihor Turkin</string-name>
          <email>i.turkin@khai.edu</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Andriy Chukhray</string-name>
          <email>a.chukhray@khai.edu</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Oleksandr Liubimov</string-name>
          <email>oleksandr.liubimov@gmail.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Lina Volobuieva</string-name>
          <email>l.volobuieva@khai.edu</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>National Aerospace University "Kharkiv Aviation Institute"</institution>
          ,
          <addr-line>17 Vadym Manko St, Kharkiv, 61070</addr-line>
          ,
          <country country="UA">Ukraine</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>ProfIT AI'25: 5th International Workshop of IT-professionals on Artificial Intelligence</institution>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2026</year>
      </pub-date>
      <abstract>
        <p>In secondary scientific research, there is often a need to classify publications by content. However, the results of diferent experts may difer significantly due to diferent experiences, context, or individual biases. This creates dificulties in the formation of systematic reviews and scientific analytics. The article proposes a method for agreeing expert assessments that takes into account the multi-level classification structure (main and additional classes) and the relative weight of the selected categories. To quantify the diferences, a distance metric was used, built by analogy with the Levenshtein distance, with the determination of the cost of editing operations (Replace, Rotate, Add). The proposed approach allows finding a compromise agreed result even with significant diferences in assessments. The method was tested on a case of classification of 203 scientific publications on the architecture of on-board computers of CubeSat nanosatellites. The results showed that the use of the method reduces the number of conflict cases and increases the stability of the classification compared to simple majority voting. The proposed approach can be used not only in scientometrics, but also in expert systems and group decision-making tasks.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;expert classification</kwd>
        <kwd>consensus evaluation</kwd>
        <kwd>distance metric</kwd>
        <kwd>systematic review</kwd>
        <kwd>artificial intelligence</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Traditionally, methods such as majority voting or calculation of agreement coeficients (e.g., Cohen’s
Kappa, Fleiss’ Kappa [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]) have been used to reconcile expert opinions. However, these approaches have
significant limitations, as:
• Replace (class replacement);
• Rotate (change of class order);
• Add (adding a new class).
      </p>
      <p>
        This approach allows:
• quantify the diferences between estimates;
• build a normalized metric of the quality of expert evaluation;
• find a balanced option that minimizes the distance to the estimates of all experts.
Unlike simple methods, our approach preserves the multi-layered nature of the classification and allows
us to work with interdisciplinary publications, when several areas are considered simultaneously in
one article. To test the method, a subject area related to the development of the architecture of CubeSat
onboard computers was chosen [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. This is a relevant direction, since modern satellite missions require
innovative hardware and software solutions, and the number of studies in this area is rapidly growing.
As a test dataset, 203 publications over the past ten years were selected from the IEEE Xplore and
Scopus databases. Two independent experts classified these works according to the specified classes,
after which the proposed method of agreeing on estimates was applied. The results show that the new
method significantly reduces the number of conflicting cases and produces a more stable and objective
classification. In particular, only 18 out of 203 articles required re-review after automated reconciliation.
Overall, this work contributes to two areas:
1. Methodological — a new method of coordinating expert assessments has been proposed, which
can be used not only in scientometrics, but also in peer review and expert systems.
2. Applied — the efectiveness of the method was demonstrated on a real case of classification of
publications about CubeSat on-board computers, which confirms its practical value in
interdisciplinary research.
      </p>
      <p>Thus, the article lays the foundation for further development of methods for agreeing expert
classifications in the context of artificial intelligence and machine learning, where the task of building consensus
is key to increasing the accuracy and objectivity of decisions.</p>
    </sec>
    <sec id="sec-2">
      <title>2. State-of-the-art</title>
      <p>
        The problem of consensus among experts is a classic problem in the field of collective decision-making,
bibliometrics, and artificial intelligence. The scientific literature describes a number of approaches that
allow assessing the degree of consensus between experts and for§ming a collective decision.
Traditional statistical methods. The most common statistical measures of consistency are given
below:
• Kappa coeficient (Cohen’s Kappa, Fleiss’ Kappa) is used to measure the level of agreement
between two or more experts, taking into account chance coincidences. This approach allows
us to determine how much better the experts’ estimates are than chance guessing, but does not
provide a mechanism for forming an integrated result.
• Kendall’s W [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] is used to analyze rankings provided by experts. It shows the level of concordance
between orders, but does not take into account the structure of the classes themselves and may
be insensitive to multilevel assessments.
• Methods based on arithmetic mean or majority voting [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] allow for rapid integration of estimates,
but often lead to loss of information, especially in cases where the publication is interdisciplinary
in nature and may belong to several classes simultaneously.
      </p>
      <p>
        Consensus-based and collective learning approaches. In the field of machine learning, the
problem of consensus estimation is often considered as an ensemble learning problem [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. The idea
is to combine the results of several “classifiers” to improve the accuracy of the final solution. Some
well-known methods include:
• bagging and boosting — use iterative learning and weight adjustment of classifier results;
• stacking — combines multiple models through a metaclassifier;
• Consensus clustering [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] is used in clustering problems, where the goal is to construct a consistent
partition of a set of objects based on several independent clusterings.
      </p>
      <p>These approaches show high eficiency, but their direct application to expert assessments has limitations:
experts often operate not only with one “label”, but with ordered sets of classes (main class, additional
classes), which requires preserving the hierarchy.</p>
      <p>
        Methods in bibliometrics and scientometrics. In bibliometric studies, there is also a need to
develop agreed classifications. Some studies suggest using the Delphi method [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ], where experts
gradually revise their assessments until an acceptable level of agreement is reached. The advantage is
that consensus can be gradually reached, but the disadvantage is that it is time-consuming. Another
direction is related to the application of multi-criteria decision-making (MCDM) methods [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ], such as
AHP (Analytic Hierarchy Process), TOPSIS, or ELECTRE. They allow building agreed solutions based
on weight coeficients and a multi-level structure of criteria, but require formalization of preferences of
all experts, which is not always possible in publication classification problems.
      </p>
      <p>Identified limitations. Analysis of the literature allows us to highlight several key problems:
1. Information loss: simple methods (majority voting) ignore secondary classes.
2. Low sensitivity to hierarchy: Consistency metrics (Kappa, Kendall’s W) work with nominal or
ranked data, but do not take into account ordered sets of classes.
3. High labor intensity: Delphi-like approaches are time-consuming and do not scale for large sets
of publications.
4. Lack of versatility: ensemble methods are efective in ML, but are not directly adapted to work
with human expert assessments containing semantically rich categories.</p>
      <p>Conclusion. Therefore, in modern literature, there is no universal approach that would allow:
• integrate multi-level expert classifications;
• take into account the relative weight of classes;
• maintain the interdisciplinarity of publications.</p>
      <p>This confirms the relevance of developing a new method for coordinating expert assessments, which
will be presented in this paper.</p>
    </sec>
    <sec id="sec-3">
      <title>3. Methodology</title>
      <p>The following procedure for processing and agreeing expert assessments in a generalized formulation
is proposed. Suppose two experts have classified Y publications. Let us assume that the cardinality of
the set of classes proposed to the expert is X, and the experts must assign each publication to no more
than Xm ordered classes, with Xm&lt;X. Then, two ratings of one publication are two vectors of classes
defined by experts:
 =&lt; 1, 2, ...,  &gt;,  =&lt; 1, 2, ...,  &gt;
(1)
In each vector of evaluation of the publication belonging to a certain class, the positions of the vectors
can contain elements from a set of predefined list of classes or an empty value ( ∅). We believe that the
assignment of the main class (vector positions – 1, 1) is mandatory, the following vector positions are
iflled only in the case when experts consider the publication to correspond to several classes. In this
context ∅ means the absence of an expert assessment in the relevant position, not a separate class. In
formulating a method for reconciling the results of expert classification of publications, the first step is
to determine the distance metric between the two estimates.</p>
      <p>When constructing a distance metric, we follow the axiomatic definitions of the metric:
• the distance between any two elements cannot be negative;
• the distance is zero only when the elements are the same;
• the distance from the first vector to the second is equal to the distance from the second vector to
• triangle inequality - the straight line distance between two vectors is no greater than the
circumthe first;
navigation through the third vector.</p>
      <p>(∅).</p>
      <p>As an axiom, we also assume that the distance is equal to one if the vectors are completely diferent
and do not intersect. In this case, in the case of complete disagreement or lack of intersection of
the classification results from diferent experts, it will be possible to achieve at least some common
understanding only through repeated expert evaluation.</p>
      <p>As a metric of the distance D(A, B) between two expert estimates, we will take the minimum total cost
of editing the vectors in such a way as to achieve their complete coincidence, similar to the Levenshtein
distance. When editing, it is allowed to apply a predefined set of operations:
•  – replace the class at position i in the vector;
• , – permutation of two classes in the score vector;
•  – add a new class to the score vector, which can only be used in the case of an empty value
We also set the following constraints on the cost of editing operations for arbitrary  (Table 1).
Limit on the cost of editing operations at random</p>
      <p>Name of the operation</p>
      <p>Limitation
∑︀=1  = 1
∀, ;  &gt;  ⇒  &lt; 
∀, , ;  &gt;  &gt;  ⇒ , ≤
∑︀=1− 1 ∑︀</p>
      <p>=+1 , = 21
∀, , ;  &gt;  &gt;  ⇒ , &lt; ,
 =</p>
      <p>, + ,
Replace
Rotate</p>
      <p>Add</p>
      <p>Since the components of the publication classification vector are initially ordered according to their
importance, the distance calculation algorithm can be characterized by the recurrent relation:
 =  − 1 + , − 1 + ,| &gt; , − 1 +</p>
      <p>* (1, 2, ...,  ) =  =1︁( (, )
The proposed solution has drawbacks, it only partially automates the search for the average value and
may require further expert coordination, since there remains a multivariate range of possible candidates
for representing classes in the average vector. As an example of the need for further expert coordination,
consider the following situation which has repeatedly arisen when processing expert assessments on
1. Each publication was allowed to be assigned to no more than three classes. The first main class is
2. The costs of editing operations are defined as follows:
• 1 = 0.6, 2 = 0.25, 3 = 0.15;
• 1,2 = 0.2, 1,3 = 0.25, 2,3 = 0.05;
• 2 = 0.25, 3 = 0.15.</p>
      <p>Since the Add weights correspond to the Replace weights, a sensitivity analysis (SA) was conducted
for Replace and Rotate using an “One at a Time” (OAT) perturbation, with the other weight vectors
ifxed as above for expert estimates (5). The datasets for the study were generated with a step size
of 0.05 for each weight, and all the data sets generated satisfy the constraints given in Table 1.
Based on the results of the SA, the defined cost indicators were selected from a small subset of
the gold standard and can be considered generalizable to diferent expert groups.
3. The two peer reviews of the publication are as follows:</p>
      <p>= ⟨1, 2, 3⟩,  = ⟨2, 1, 4⟩.</p>
      <p>In this case, the distance between the two estimates is the sum of the values of the 1,2 and
3 operations, which is 0.35. There are two alternative options for the mean vector S*, the
distance from which to the input vectors will not exceed 0.2: 1* = ⟨1, 2, 4⟩,
2* = ⟨2, 1, 3⟩. If it is desirable to avoid a second examination and both experts have
the same level of confidence in their competence, then the best solution is to choose the average vector
randomly. For a more detailed analysis of this limitation, let us define the probability distribution
function of the event that, as a result of independent classification of publications, experts selected q
common classes. That is:
• There is an alphabet A, which includes X unique characters, namely a list of classes to which it is
proposed to attribute each publication;
{1, 2, . . . , };
set of characters);
• expert evaluations are filled independently, without repetitions of symbols (classes) with a limited
length  &lt; . These are two subsets where 1, 2 ⊂ , |1| = 1, |2| = 2, 1, 2 ∈
• The order of the characters does not matter (the probability of intersection depends only on the
• all subsets of fillings are equally likely;
• the lengths of the filled lines are also equally likely with the restriction that the length of the line
is randomly chosen from the range 1 . . . .</p>
      <p>Then, with fixed 1, 2 ∈ {1, 2, . . . , }, the conditional probability of occurrence of q intersections
is:
 (, 1, 2, ) =
︀( 1 )︀(
− 1 )︀
2− 
︀( 2 )︀
where (︀  )︀ = !(− )! is the number of combinations of selecting y elements from a set containing x
!
elements.</p>
      <p>After averaging over all 1, 2 ∈ {1, 2, . . . , }, taking into account that 1, 2 are random and
uniformly distributed, we obtain a lower bound for the quantitative assessment of the coincidence of
expert assessments. In essence, this is the probability of a random variable that describes the number of
common elements (intersections) between two random subsets of the alphabet X, with random lengths
uniformly distributed from one to . In other words, this is the probability of a situation where,
instead of assessing, experts indicated random classes:
 (, , ) =</p>
      <p>1
2 1=1 2=1
  ( 1 ) ︁( − 1 )︁
∑︁ ∑︁ 2− 
( 2 )
.</p>
      <p>If the expert estimates completely coincide, the probability will be as follows:</p>
      <p>(, ) = ∑︁  (, , ) · ,
=1
(5)
(6)
(7)
(8)
 + 1
2</p>
      <p>The normalized metric of expert assessment quality (, ) is constructed based on the assumption
that the minimum value  = 0 corresponds to random assignment of classes, the maximum value
 = 1 can be achieved only with complete coincidence of expert assessments. A negative value of Q is
possible if experts, for some reason, intentionally indicated incorrect answers.</p>
      <p>Then the quality of the assessment through the metric of coincidence of expert assessments is determined
as follows:
(, ) =
 (, ) −  (, )
 () −  (, )
,
where (, ) – the calculated average value of the number of intersections in expert assessments
of the membership of Y publications to X classes, with the restriction that each publication can be
assigned to no more than  classes.</p>
      <p>(9)
(10)
(11)
where  is the number of common classes in two expert assessments of one publication.</p>
      <p>,
. (, )
()</p>
      <p>X</p>
      <p>1</p>
      <p>Thus, firstly, the lower and upper limits for assessing the consistency of expert evaluation results
are determined, and, as a result, a normalized metric of the quality of expert evaluation is proposed.
Secondly, a method is proposed to coordinate expert assessments and find a compromise.</p>
    </sec>
    <sec id="sec-4">
      <title>4. Experimental Results</title>
      <sec id="sec-4-1">
        <title>4.1. Data Set</title>
        <p>
          To test the proposed method, a sample of 203 scientific publications devoted to the architecture of
onboard computers (OBC) for CubeSat nanosatellites was formed [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ]. The data sources were the
electronic libraries IEEE Xplore and Scopus in the period 2015–2024. The PRISMA [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ] methodology
was used to select articles, which included:
• primary search by keywords (“CubeSat”, “Onboard Computer”, “Fault Tolerance”, “Software
        </p>
        <p>Architecture”, etc.);
• removing duplicates and irrelevant publications;
• verification of compliance with inclusion criteria.</p>
        <p>The result was a representative sample of works covering both hardware and software aspects of
building a CubeSat OBC.</p>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. Experts</title>
        <p>Two experts were involved in the classification process:
• Expert 1 — a specialist in the industrial IT industry, specializing in building embedded real-time
systems.
• Expert 2 — a representative of the university environment who researches on-board systems
architectures in scientific projects.</p>
        <p>Each expert classified all 203 publications according to a defined system of categories. For each article,
experts were required to indicate a main class and up to two additional classes.</p>
      </sec>
      <sec id="sec-4-3">
        <title>4.3. Classification System</title>
        <p>Six main classes have been identified, reflecting key areas of CubeSat OBC research, and they are listed
below.</p>
        <p>1. Hardware Architectures.
2. Software Systems.
3. Fault Tolerance &amp; Reliability.
4. Power &amp; Resource Optimization.
5. Communications and interfaces (Communication &amp; I/O).</p>
        <p>6. Intelligent functions (AI &amp; Autonomy).</p>
        <p>Each publication could be assigned to several classes simultaneously, reflecting its interdisciplinary
nature.</p>
      </sec>
      <sec id="sec-4-4">
        <title>4.4. Comparable Methods</title>
        <p>Three approaches were compared to assess efectiveness:
1. Majority voting — choosing the class that is most frequently found among expert assessments.
2. Cohen’s Kappa — assessment of the consistency of two experts without building an integrated
classification.</p>
        <p>3. Proposed method — distance metric + search for a compromise solution.</p>
      </sec>
      <sec id="sec-4-5">
        <title>4.5. Results</title>
        <p>• According to the majority voting method, 57 publications (28%) turned out to be conflicting,
where experts chose diferent main classes;
• According to Cohen’s Kappa coeficient, the level of agreement was 0.62 (average agreement);
• The proposed method allowed for the agreement of most cases, leaving only 18 publications (9%)
that required re-examination due to too high a distance between the assessments.
In addition, the method was particularly useful in cases of interdisciplinary articles. For example, if one
paper described both hardware solutions and fault tolerance algorithms, traditional methods "forced"
the selection of only one class. Instead, our method allowed us to maintain a multi-level evaluation
structure (main + additional classes), which provided a more objective reflection of the content.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Discussion</title>
      <p>The results obtained show that the proposed method of harmonizing expert classifications of publications
has a number of advantages compared to traditional approaches:
• reduces the number of conflicts between experts by more than three times (from 57 to 18
publications);
• provides a better interpretation of results in the case of interdisciplinarity;
• allows you to quantify the distance between assessments and use this metric as a criterion for
re-examination.</p>
      <p>Thus, the method has proven its efectiveness as a tool for harmonizing expert classifications, combining
formalism with practical applicability.
5.1. Advantages of the Method
1. Preservation of multi-level classifications. Unlike most existing approaches, the method does not
reduce the assessment to just one "correct" class, but allows for a hierarchical structure (main
class + additional ones). This is especially important in interdisciplinary research, where one
article can simultaneously belong to several directions.
2. Quantifying diferences. Using the distance metric allows us to calculate the "degree of divergence"
between expert assessments. This allows us to more objectively determine when the diference is
insignificant (for example, a permutation of classes) and when it is critical (replacement with a
completely diferent class).
3. Flexibility in customization. The cost of Replace, Rotate, and Add operations can be tailored to
specific tasks, giving more or less weight to individual aspects of the classification. This makes
the method versatile for diferent subject areas.
4. Reducing the number of conflict cases. In the experiment, the number of publications requiring
re-examination decreased from 57 (28%) to 18 (9%). This indicates an increase in the eficiency of
the systematic review process.</p>
      <sec id="sec-5-1">
        <title>5.2. Limitations</title>
        <p>Despite the obvious advantages, the method also has certain limitations:
• dependence on the choice of weighting factors. If the values of operations (Replace, Rotate, and
Add) are set incorrectly, the result may be skewed. Therefore, it is necessary to calibrate the
parameters for each subject area;
• a small number of experts in the study. The presented experiment involved only two experts. For
greater reliability, tests with a wider group of evaluators are needed;
• the need for manual intervention in complex cases. If the distance between the ratings exceeds a
certain threshold, the article still needs to be re-examined. It is currently impossible to completely
eliminate the human factor.</p>
      </sec>
      <sec id="sec-5-2">
        <title>5.3. Potential Areas of Application</title>
        <p>The proposed method can be used not only for classifying publications, but also in a wider range of
tasks:
• bibliometrics and scientometrics: construction of systematic reviews, formation of taxonomies of
scientific areas, analysis of interdisciplinary research;
• peer review and expert evaluation: coordination of decisions of multiple reviewers when
evaluating articles or grant applications;
• medical diagnostic systems: combining the conclusions of several expert doctors to form a more
reliable diagnosis;
• decision support systems: use in project management, multi-criteria evaluations, or collective
strategy selection;
• artificial intelligence and machine learning: adaptation of the method as a tool for building
ensembles in classification problems, where various algorithms act as "experts".</p>
      </sec>
      <sec id="sec-5-3">
        <title>5.4. Generalization</title>
        <p>Thus, the proposed method can be considered as a hybrid approach that combines the ideas of classical
consistency metrics (Kappa, Kendall’s W) with the concepts of ensemble learning. Its main advantage
lies in the ability to work with ordered sets of classes and in preserving the multilayer structure of
estimates, which is especially relevant for the analysis of complex interdisciplinary data.</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>6. Conclusions</title>
      <p>In this paper, a method for harmonizing the results of expert classification of scientific publications was
proposed and tested, which combines approaches from bibliometrics and artificial intelligence. Unlike
traditional methods, such as majority voting or consistency coeficients (Cohen’s Kappa, Kendall’s
W), the developed approach takes into account the multi-level structure of classifications (main and
additional classes) and allows for quantitative assessment of discrepancies between expert assessments.</p>
      <p>The proposed method is based on the introduction of a distance metric between expert assessments,
which takes into account the operations of replacement (Replace), rotation (Rotate), and addition (Add).
This makes it possible to form a compromise solution that minimizes the distance to the assessments
of all experts, as well as to identify publications that require re-examination. The method was tested
on a corpus of 203 publications on the architecture of CubeSat onboard computers. Two experts with
diferent professional experiences participated in the study. The results showed that the number of
conflict cases decreased more than threefold (from 57 to 18 publications), and the level of classification
stability increased significantly. The main conclusions of the study can be formulated as follows.
1. The proposed method provides a more objective and interpretable agreement of expert
assessments.
2. It allows you to maintain the interdisciplinary nature of classifications without reducing
publications to a single class.
3. The use of a quantitative distance metric opens up opportunities for automating the matching
process.</p>
      <p>Further research directions:
• expanding the method to work with a larger number of experts;
• integration of weighting factors depending on the expert’s qualifications or industry specifics;
• application in the field of peer review, medical expert systems, and other domains where the
integration of the opinions of several specialists is important;
• development of machine learning-based tools for automated reconciliation support.
Thus, the developed approach contributes to the development of the methodology for analyzing scientific
publications and can become an efective tool for improving the quality of systematic reviews in various
ifelds of knowledge.</p>
    </sec>
    <sec id="sec-7">
      <title>7. Authors’ Contribution</title>
      <p>Problem formulation – I. Turkin; review and analysis of information sources – A. Chukhray; development
of a method for processing and coordinating expert assessments – O. Liubimov, L. Volobuieva; conducting
research, evaluation, and visualization of results – I. Turkin, A. Chukhray.</p>
    </sec>
    <sec id="sec-8">
      <title>Acknowledgments</title>
      <p>The study was funded by the National Research Foundation of Ukraine in the framework of the research
project NRFU.2023.04/0143 - "Experimental development and validation of the on-board computer of a
dual-purpose unmanned aerial vehicle". The authors express their gratitude to the graduate student
of the Department of Software Engineering of the National Aerospace University, Valkovy VS, for his
work as an expert in the classification of scientific publications.</p>
    </sec>
    <sec id="sec-9">
      <title>Declaration on Generative AI</title>
      <p>During the preparation of this work, the authors used Grammarly in order to: Grammar and spelling
check. After using these tool(s)/service(s), the author(s) reviewed and edited the content as needed and
take(s) full responsibility for the publication’s content.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>J.</given-names>
            <surname>Cohen</surname>
          </string-name>
          ,
          <article-title>A coeficient of agreement for nominal scales</article-title>
          ,
          <source>Educational and Psychological Measurement</source>
          <volume>20</volume>
          (
          <year>1960</year>
          )
          <fpage>37</fpage>
          -
          <lpage>46</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>I.</given-names>
            <surname>Turkin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Liubimov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Volobuieva</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Valkovyi</surname>
          </string-name>
          ,
          <article-title>Hardware and software of cubesat nanosatellites' on-board computers: a systematized literature review</article-title>
          ,
          <source>Aerospace Technic and Technology</source>
          <volume>0</volume>
          (
          <year>2025</year>
          )
          <fpage>116</fpage>
          -
          <lpage>137</lpage>
          . doi:https://doi.org/10.32620/aktt.
          <year>2025</year>
          .
          <volume>3</volume>
          .11.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>M. G.</given-names>
            <surname>Kendall</surname>
          </string-name>
          , Rank Correlation Methods, Grifin,
          <year>1970</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>T. G.</given-names>
            <surname>Dietterich</surname>
          </string-name>
          ,
          <article-title>Ensemble methods in machine learning</article-title>
          ,
          <source>in: Multiple Classifier Systems</source>
          , Springer, Berlin, Heidelberg,
          <year>2000</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>15</lpage>
          . doi:
          <volume>10</volume>
          .1007/3-540-45014-
          <issue>9</issue>
          _
          <fpage>1</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>Z.-H.</given-names>
            <surname>Zhou</surname>
          </string-name>
          , Ensemble Methods:
          <article-title>Foundations and Algorithms</article-title>
          , CRC Press,
          <year>2012</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>A.</given-names>
            <surname>Strehl</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Ghosh</surname>
          </string-name>
          ,
          <article-title>Cluster ensembles-a knowledge reuse framework for combining multiple partitions</article-title>
          ,
          <source>Journal of Machine Learning Research</source>
          <volume>3</volume>
          (
          <year>2002</year>
          )
          <fpage>583</fpage>
          -
          <lpage>617</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>H.</given-names>
            <surname>Linstone</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Turof</surname>
          </string-name>
          , The Delphi Method,
          <string-name>
            <surname>Addison-Wesley</surname>
          </string-name>
          ,
          <year>2002</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>T. L.</given-names>
            <surname>Saaty</surname>
          </string-name>
          ,
          <article-title>Decision making with the analytic hierarchy process</article-title>
          ,
          <source>Int. J. Services Sciences</source>
          <volume>1</volume>
          (
          <year>2008</year>
          )
          <fpage>83</fpage>
          -
          <lpage>98</lpage>
          . doi:
          <volume>10</volume>
          .1504/IJSSCI.
          <year>2008</year>
          .
          <volume>017590</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <surname>PRISMA</surname>
          </string-name>
          , PRISMA statement, https://www.prisma-statement.org/,
          <year>2020</year>
          . Accessed on 2025-
          <volume>09</volume>
          -03.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>