<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Continuous-categorical feature dependence with Determination-based correlation coeficient ⋆</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Andrii Berko</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Vladyslav Holdovanskyi</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Lviv Polytechnic National University</institution>
          ,
          <addr-line>12 Stepana Bandery Street, Lviv, 79014</addr-line>
          ,
          <country country="UA">Ukraine</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>The paper explores the application of the Determination-Based Correlation Coeficient (DBCC) for analyzing dependencies between continuous and categorical features in classification tasks. While DBCC has previously been validated in regression contexts, its suitability for classification problems remains underexplored. This study evaluates DBCC across various dependency structures, including step, multi-class, and sigmoidal relationships, demonstrating its robustness and efectivity. The findings suggest that DBCC provides a computationally eficient and flexible alternative to traditional methods such as ANOVA, point-biserial correlation, and mutual information. Additionally, classification performance metrics such as Accuracy, F1-score, and ROC-AUC are examined to assess the practical implications of using Grid-Mean Algorithm in feature selection and model evaluation. Future research directions include refining the methodology and extending its application to high-dimensional datasets.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Analysis</kwd>
        <kwd>Correlation coeficient</kwd>
        <kwd>Classification</kwd>
        <kwd>Categorical data</kwd>
        <kwd>Machine Learning</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Understanding the relationship between continuous and categorical features is a fundamental aspect
of many classification tasks in machine learning. While numerous techniques exist for dependency
analysis, their efectiveness varies based on data structure, underlying assumptions, and computational
eficiency. In our previous work, we introduced and analyzed the Determination-Based Correlation
Coeficient (DBCC) in the context of regression tasks, demonstrating its ability to capture complex
dependencies and maintain robust performance across diverse datasets [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. However, classification
problems require diferent approaches, as feature relationships often exhibit step or sigmoidal patterns
that traditional correlation metrics struggle to interpret [
        <xref ref-type="bibr" rid="ref2 ref3 ref4 ref5 ref6 ref7">2, 3, 4, 5, 6, 7</xref>
        ]. Given these strengths, a natural
question arises: can DBCC also be applied to classification problems, where feature relationships often
exhibit stepwise or sigmoidal patterns that traditional correlation metrics struggle to interpret?
      </p>
      <p>
        In this study, we explore the use of DBCC for assessing feature dependence in classification tasks,
aiming to provide a robust and interpretable alternative to existing methods. We evaluate DBCC on
synthetic datasets with stepwise, multi-class, and sigmoidal dependencies, comparing its performance
against established dependency measures. Additionally, we examine classification metrics such as
accuracy, F1-score, log loss, and ROC-AUC [
        <xref ref-type="bibr" rid="ref10 ref11 ref12 ref13 ref8 ref9">8, 9, 10, 11, 12, 13</xref>
        ] to assess how DBCC contributes to
feature selection and model evaluation.
      </p>
      <p>
        Accurate dependency estimation is crucial for classification models. Recent studies have highlighted
the importance of efective dependency estimation in classification. Techniques such as ANOVA and
point-biserial correlation rely on distributional assumptions that limit their applicability in real-world
scenarios [
        <xref ref-type="bibr" rid="ref14 ref15 ref16">14, 15, 16</xref>
        ]. Additionally, mutual information (MI) provides a flexible alternative but often
requires careful parameter selection and exhibits higher computational complexity [
        <xref ref-type="bibr" rid="ref17 ref18">17, 18</xref>
        ]. The Akaike
Information Criterion (AIC) has been utilized in model selection but does not directly quantify feature
dependence [19, 20, 21, 22]. Given these limitations, exploring alternative approaches like DBCC is
essential for improving classification model performance.
      </p>
    </sec>
    <sec id="sec-2">
      <title>2. Materials</title>
      <p>In our previous studies, we conducted an in-depth investigation of the Determination-Based Correlation
Coeficient (DBCC) and explored its application for predicting regression metrics based on data
characteristics. Our findings demonstrated that DBCC is highly efective at capturing nonlinear dependencies,
making it a powerful tool for assessing complex relationships within datasets. Given that classification
tasks frequently involve interactions between continuous and categorical features, which often follow
stepwise or sigmoidal patterns, we hypothesize that DBCC can also provide valuable insights in this
context. DBCC is calculated using the following formula:
 2 = 1 −
∑︀=0 [ − ()]2
∑︀=0 [ −  ()]2
Then for each  ∈ [, ], the grid is:
where () – is the mean value within the grid cell that contains . This grid-based mean is defined
as follows: Let the interval [, ] be divided into  equally sized subintervals of width ℎ = ( − )/ .
ℎ =
︂{
∈[,+1](),  =  + ℎ,  = 0, . . . , ,
ℎ =
 −  }︂

To rigorously assess the efectiveness of DBCC in classification tasks, we designed and executed a
series of controlled experiments using synthetic datasets. These experiments aimed to analyze how
well DBCC captures diferent types of dependencies between continuous and categorical features. Our
methodology consisted of the following key steps:
• Generating datasets with diferent types of dependencies between continuous and categorical
features.
• Computing the DBCC correlation coeficient and interpreting its results.</p>
      <p>• Analyzing the outcomes and assessing the adequacy of DBCC in various scenarios.
We categorized dependencies into five primary types to ensure comprehensive testing:
• Step function.
• Twice-changing step function.
• No dependency.
• Multi-class categorical feature.</p>
      <p>• Sigmoidal dependency with varying levels of noise.</p>
      <p>Each synthetic dataset contained 10,000 data points, ensuring stable and statistically meaningful
estimates. To manage the discretization of continuous variables, we set the grid parameter to n =
100. This parameter controls the level of granularity in the approximation process: lower values
smooth out dependencies, whereas higher values increase noise and inconsistencies. Through extensive
testing, we determined that n = 100 provides an optimal balance between approximation accuracy and
computational eficiency.</p>
      <p>Step Function. In the first experiment, the continuous variable was sampled from a normal
distribution with a mean of 50 and a standard deviation of 15. The categorical variable was assigned a
value of 1 if the condition  &gt; 50 was met and 0 otherwise (see Figure 1). This dependency structure
is frequently observed in classification tasks where a clear threshold separates the classes.</p>
      <p>The computed DBCC value was 0.975, indicating an almost perfect correspondence between the
continuous feature and the categorical label. The approximation accurately captured the underlying
1.0
1.0
0.8
0.6
c
n
u
f
0.4
dependency, though minor uncertainty was observed around the threshold due to the efects of
gridbased discretization.</p>
      <p>Double Step Function. For the second experiment, we considered a more complex twice-changing
step function. In this scenario, the categorical feature changed its value at two distinct thresholds,
simulating real-world classification problems where decision boundaries are non-monotonic (see Figure
2).</p>
      <p>The resulting DBCC value was 0.996, demonstrating DBCC’s ability to efectively capture
nonmonotonic relationships. This result is particularly noteworthy, as many traditional correlation measures
struggle to detect such patterns.
1.0
0.8
1.0
0.8
40 x 60
0.2 noise 0.3
0
20
80
100
0.0
0.1
0.4
0.5</p>
      <p>No dependency. To examine DBCC’s behavior when no dependency exists between the variables,
we generated independent random features (see Figure 3). As expected, DBCC returned a near-zero
correlation value of 0.009, correctly indicating the absence of a relationship. This outcome confirms that
DBCC does not falsely identify correlations where none exist, reinforcing its reliability as a statistical
measure.</p>
      <p>Multiclass Categorical Feature. A distinguishing characteristic of DBCC is its capacity to handle
multi-class categorical variables. To test this capability, we constructed a dataset where the categorical
feature could assume three distinct values (see Figure 4). The analysis yielded a DBCC value of 0.984,
illustrating DBCC’s robustness in capturing relationships in multi-class classification settings.
2.00
true data
estimated data
1.75
1.50
1.25
c
fun1.00
0.75
0.50
0.25
0.00
true data
estimated data
−10.0
−7.5
−5.0
−2.5
0.0
2.5
5.0
7.5</p>
      <p>10.0
x
0
20
40
60
80
100
0.0
0.1
0.3
0.4
0.5
x</p>
      <p>Sigmoidal Dependence.To simulate more intricate nonlinear dependencies, we introduced a
sigmoidal function:
 =</p>
      <p>1
where  controls the steepness of the transition between class labels. As  increases, the function
becomes more gradual, introducing greater uncertainty and blurring class boundaries (see Figure 5).
Conversely, as  approaches zero, the function approximates a power-law relationship.</p>
      <p>Our results showed that as  increased, the DBCC value decreased accordingly, reflecting the growing
ambiguity in classification. This behavior suggests that DBCC is sensitive to uncertainty in decision
boundaries and can serve as an efective diagnostic tool for evaluating the separability of classes in
noisy data.</p>
      <p>DBCC results showed that as class boundaries became more blurred (higher  ), the correlation
coeficient decreased, reflecting the growing uncertainty in classification.</p>
      <p>The findings indicate that DBCC is an efective tool for assessing dependencies in classification tasks.
It demonstrates high accuracy even in cases of complex and nonlinear dependencies. In future research,
we plan to evaluate the efectiveness of this approach on real-world datasets.</p>
      <p>Comparison with Other Methods. To provide a broader perspective, we compared DBCC with
widely used methods for analyzing dependencies between continuous and categorical variables. The
key characteristics and performance metrics of these methods are summarized in Table 1.</p>
      <p>Our preliminary results suggest that DBCC ofers a unique balance between sensitivity to nonlinear
patterns and robustness against noise, making it a valuable addition to the toolkit of data scientists
working with classification problems.</p>
      <p>Extending Grid-Mean Algorithm for Classification Metrics. In our other study, we explored
the application of a grid-based algorithm for the approximate computation of regression metrics. The
fundamental idea behind this approach was that by discretizing the continuous feature space into
grid intervals and computing mean values within each interval, we could eficiently approximate
regression-based evaluation metrics. Following the same logic, we propose extending this methodology
to classification metrics, leveraging the Grid-Mean Algorithm as a machine learning model capable of
estimating classification performance indicators.</p>
      <p>Beyond correlation analysis, the grid-based structure of DBCC allows for the approximation of
classification metrics by utilizing the probability distribution of categorical labels within each grid
interval. Since each segment of the grid contains averaged feature values, it can serve as a foundation for
estimating the probability of categorical outcomes, thereby supporting class prediction. This capability
enables us to derive classification performance metrics directly from the grid without requiring an
explicit predictive model.</p>
      <p>We considered four primary classification metrics that are widely used for evaluating model
performance:
• Accuracy – The proportion of correctly classified values relative to the total number of samples.</p>
      <p>This is a fundamental metric that provides an overall measure of predictive performance.</p>
      <p>Accuracy =</p>
      <p>True Positive + True Negative</p>
      <p>True Positive + True Negative + False Positive + False Negative
• F1-Score – The harmonic mean of precision and recall, balancing the trade-of between false
positives and false negatives. It is particularly useful in cases with imbalanced class distributions.
where –  predicted probability of the i-point, and –  label of this point.</p>
      <p>To evaluate the efectiveness of this approach, we conducted experiments using a sigmoidal function
with varying  values:
 =</p>
      <p>1
where controls the level of noise in the data. Higher values of result in smoother transitions and
greater uncertainty near the decision boundary, while lower values produce sharper class separations.
This setup allows us to assess how well the Grid-Mean Algorithm can approximate classification metrics
under diferent levels of ambiguity in class labels.</p>
      <p>For each experimental setting, we compared the actual and approximated values of classification
metrics, ensuring a comprehensive evaluation of this method’s accuracy. The results (see Figure 6),
demonstrate that the approximation error never exceeded 0.05 across all considered scenarios. This
level of accuracy indicates that the proposed approach provides a reliable and computationally eficient
means of estimating classification performance without the need for full-fledged model training.</p>
      <p>Our findings indicate that DBCC is an efective tool for assessing dependencies in classification tasks.
It demonstrates high accuracy, even in cases of complex and nonlinear dependencies. Moreover, its
grid-based approach extends to approximate classification metric computation, making it a versatile
tool in data analysis.</p>
    </sec>
    <sec id="sec-3">
      <title>3. Conclusions</title>
      <p>This study demonstrates the efectiveness of the Determination-Based Correlation Coeficient (DBCC) in
evaluating dependencies between continuous and categorical features in classification tasks. Accurately
assessing such dependencies is essential for improving classification models, as it influences feature
selection, model interpretability, and predictive performance. Traditional methods often struggle to
capture complex relationships, making the development of more flexible and eficient dependency
measures a crucial research direction.</p>
      <p>Through extensive experimentation on synthetic datasets, our results confirm that DBCC efectively
identifies various types of relationships, including stepwise, sigmoidal, and multi-class dependencies,
while also correctly recognizing the absence of correlation where applicable. The ability to handle
both simple and intricate dependency structures makes DBCC a valuable tool for feature evaluation.
Furthermore, compared to conventional approaches such as ANOVA, point-biserial correlation, and
mutual information, DBCC ofers a favorable balance between computational eficiency, interpretability,
and adaptability to diverse data distributions.</p>
      <p>Key Advantages of DBCC:
• Efective Handling of Nonlinear and Complex Dependencies: Unlike traditional methods that
often assume linearity or specific distributions, DBCC can capture intricate relationships without
imposing restrictive assumptions. This makes it particularly useful in real-world classification
tasks, where dependencies are rarely straightforward.
• Robustness Across Various Dependency Structures: DBCC performs well across diferent types
of relationships, including stepwise transitions, non-monotonic dependencies, and multi-class
interactions. This generalizability enhances its applicability in diverse classification scenarios.
• High Computational Eficiency: One of DBCC’s strengths is its ability to process large datasets
eficiently, making it scalable for applications involving extensive feature analysis. Unlike
computationally expensive techniques such as mutual information, DBCC provides a streamlined
approach that balances precision with speed.</p>
      <p>Future Research Directions. To further enhance DBCC’s efectiveness, several avenues for future
research should be explored. One key area is refining the methodology to improve precision and
robustness across more diverse data structures. Advanced techniques for parameter tuning and optimization
could enhance its sensitivity in detecting subtle dependencies, ensuring more reliable results even in
challenging classification settings.</p>
      <p>Another important direction is extending DBCC’s applicability to high-dimensional datasets, where
feature interactions become increasingly complex. Many real-world classification tasks involve
numerous variables, and a robust dependency measure must be capable of eficiently handling such
multi-feature relationships. Developing scalable implementations of DBCC that leverage parallel
computing and optimized algorithms could significantly enhance its usability in large-scale machine learning
workflows.</p>
      <p>Moreover, integrating DBCC into automated feature selection frameworks could greatly benefit the
ifeld by providing a systematic way to identify relevant features based on their dependency structures.
By incorporating DBCC into feature engineering pipelines, practitioners can streamline the process of
selecting informative variables, ultimately improving model performance while reducing computational
overhead.</p>
      <p>Finally, exploring the potential of DBCC for approximate computation of classification metrics
presents an exciting opportunity. If DBCC can serve as a proxy for evaluating classification performance,
it could enable rapid model assessment without requiring extensive training and validation steps. Such
an approach would be particularly valuable in iterative machine learning workflows, where quick
feedback is essential for model optimization.</p>
      <p>By addressing these research directions, DBCC can be further established as a powerful and versatile
tool for analyzing complex dependencies in modern data science, contributing to more efective and
interpretable classification models.</p>
    </sec>
    <sec id="sec-4">
      <title>Declaration on Generative AI</title>
      <p>During the preparation of this work, the authors used OpenAI GPT-4 in order to: check grammar and
spelling, improve formal writing style, and enhance overall readability of the text. After using this tool,
the authors reviewed and edited the content as needed and takes full responsibility for the publication’s
content.
Cryptographic Hardware and Embedded Systems - CHES 2009, volume 5747 of Lecture Notes in
Computer Science, Springer, 2009, pp. 429–443.
[19] Q. Liu, M. A. Charleston, S. A. Richards, B. R. Holland, Performance of akaike information criterion
and bayesian information criterion in selecting partition models and mixture models, Systematic
Biology 72 (2023) 92–105. doi:10.1093/sysbio/syac081.
[20] J. E. Cavanaugh, A. A. Neath, The akaike information criterion: Background, derivation, properties,
application, interpretation, and refinements, Wiley Interdisciplinary Reviews: Computational
Statistics 11 (2019) e1460. doi:10.1002/wics.1460.
[21] H. Cheng, B. Sterner, Error statistics using the akaike and bayesian information criteria, Erkenntnis
(2024). doi:10.1007/s10670-024-00897-2.
[22] C. Sutherland, D. Hare, P. J. Johnson, et al., Practical advice on variable selection and reporting
using akaike information criterion, Proceedings of the Royal Society B: Biological Sciences 290
(2023) 20231261. doi:10.1098/rspb.2023.1261.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>A.</given-names>
            <surname>Berko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Alieksieiev</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Holdovanskyi</surname>
          </string-name>
          ,
          <article-title>Determination-based correlation coeficient</article-title>
          ,
          <source>in: Proceedings of the 6th International Workshop on Modern Machine Learning Technologies (MoMLeT-2024)</source>
          , Lviv-Shatsk, Ukraine,
          <year>2024</year>
          . May 31 - June 1,
          <year>2024</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>S. B.</given-names>
            <surname>Kotsiantis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>I. D.</given-names>
            <surname>Zaharakis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. E.</given-names>
            <surname>Pintelas</surname>
          </string-name>
          ,
          <article-title>Machine learning: a review of classification and combining techniques</article-title>
          ,
          <source>Artificial Intelligence Review</source>
          (
          <year>2007</year>
          ).
          <source>doi: 10.1007/s10462-007-9052-3.</source>
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>P.</given-names>
            <surname>Pukach</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Liubinskyy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Hladun</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Holdovanskyi</surname>
          </string-name>
          ,
          <article-title>The classifier models usage for the recruitment process forecasting for applicants of higher education to universities of ukraine</article-title>
          ,
          <source>in: Data-Centric Business and Applications</source>
          , Springer,
          <year>2024</year>
          . doi:
          <volume>10</volume>
          .1007/978-3-
          <fpage>031</fpage>
          -60815-
          <issue>5</issue>
          _
          <fpage>4</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>Q.</given-names>
            <surname>An</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Rahman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. J.</given-names>
            <surname>Kang</surname>
          </string-name>
          ,
          <article-title>A comprehensive review on machine learning in healthcare industry: Classification, restrictions, opportunities and challenges</article-title>
          ,
          <source>Sensors</source>
          <volume>23</volume>
          (
          <year>2023</year>
          ). doi:
          <volume>10</volume>
          . 3390/s23094178.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>A. F. A. H.</given-names>
            <surname>Alnuaimi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T. H. K.</given-names>
            <surname>Albaldawi</surname>
          </string-name>
          ,
          <article-title>An overview of machine learning classification techniques</article-title>
          ,
          <source>BIO Web of Conferences</source>
          <volume>97</volume>
          (
          <year>2024</year>
          )
          <article-title>00133</article-title>
          . doi:
          <volume>10</volume>
          .1051/bioconf/20249700133.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>R. I.</given-names>
            <surname>Mukhamediev</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Popova</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Kuchin</surname>
          </string-name>
          , et al.,
          <article-title>Review of artificial intelligence and machine learning technologies: Classification, restrictions, opportunities and challenges</article-title>
          ,
          <source>Mathematics</source>
          <volume>10</volume>
          (
          <year>2022</year>
          )
          <article-title>2552</article-title>
          . doi:
          <volume>10</volume>
          .3390/math10152552.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>J. E.</given-names>
            <surname>Black</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. K.</given-names>
            <surname>Kueper</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T. S.</given-names>
            <surname>Williamson</surname>
          </string-name>
          ,
          <article-title>An introduction to machine learning for classification and prediction</article-title>
          , Family
          <string-name>
            <surname>Practice</surname>
          </string-name>
          (
          <year>2022</year>
          ). doi:
          <volume>10</volume>
          .1093/fampra/cmac104.
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>L.</given-names>
            <surname>Ferrer</surname>
          </string-name>
          ,
          <article-title>Analysis and comparison of classification metrics</article-title>
          ,
          <source>arXiv preprint arXiv:2209.05355</source>
          ,
          <year>2023</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>Z. D.</given-names>
            <surname>Vujović</surname>
          </string-name>
          ,
          <article-title>Classification model evaluation metrics</article-title>
          ,
          <source>International Journal of Advanced Computer Science and Applications</source>
          <volume>12</volume>
          (
          <year>2021</year>
          ). doi:
          <volume>10</volume>
          .14569/IJACSA.
          <year>2021</year>
          .
          <volume>0120670</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>J.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. H.</given-names>
            <surname>Gandomi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Holzinger</surname>
          </string-name>
          ,
          <article-title>Evaluating the quality of machine learning explanations: A survey on methods and metrics</article-title>
          ,
          <source>Electronics</source>
          <volume>10</volume>
          (
          <year>2021</year>
          )
          <article-title>593</article-title>
          . doi:
          <volume>10</volume>
          .3390/ electronics10050593.
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>M.</given-names>
            <surname>Steurer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. J.</given-names>
            <surname>Hill</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Pfeifer</surname>
          </string-name>
          ,
          <article-title>Metrics for evaluating the performance of machine learningbased automated valuation models</article-title>
          ,
          <source>Journal of Property Research</source>
          <volume>38</volume>
          (
          <year>2021</year>
          )
          <fpage>99</fpage>
          -
          <lpage>129</lpage>
          . doi:
          <volume>10</volume>
          .1080/ 09599916.
          <year>2020</year>
          .
          <volume>1858937</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>J.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <article-title>Analysis of statistic metrics in diferent types of machine learning</article-title>
          ,
          <source>in: Highlights in Science, Engineering and Technology IFMPT</source>
          , volume
          <volume>88</volume>
          ,
          <year>2024</year>
          , p.
          <fpage>182</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>S. A.</given-names>
            <surname>Hicks</surname>
          </string-name>
          ,
          <string-name>
            <given-names>I.</given-names>
            <surname>Strümke</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Thambawita</surname>
          </string-name>
          , et al.,
          <article-title>On evaluation metrics for medical applications of artificial intelligence</article-title>
          ,
          <source>Scientific Reports</source>
          <volume>12</volume>
          (
          <year>2022</year>
          )
          <article-title>5979</article-title>
          . doi:
          <volume>10</volume>
          .1038/s41598-022-09954-8.
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>C. E.</given-names>
            <surname>Flores</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K. L.</given-names>
            <surname>Flores</surname>
          </string-name>
          ,
          <article-title>Kruskal-wallis, friedman and mood nonparametric tests applied to business decision making, Espirales</article-title>
          .
          <source>Revista Multidisciplinaria de Investigación Científica</source>
          <volume>6</volume>
          (
          <year>2022</year>
          )
          <fpage>1</fpage>
          -
          <lpage>20</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>D. C.</given-names>
            <surname>Howell</surname>
          </string-name>
          , Fundamental Statistics for the Behavioral Sciences, 5th ed., Duxbury Press, Pacific Grove,
          <year>2004</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>D. J.</given-names>
            <surname>Sheskin</surname>
          </string-name>
          , Handbook of Parametric and Nonparametric Statistical Procedures, 2nd ed.,
          <source>Chapman &amp; Hall</source>
          , London,
          <year>2000</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>A.</given-names>
            <surname>Kraskov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Stögbauer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Grassberger</surname>
          </string-name>
          , Estimating mutual information,
          <source>Physical Review E</source>
          <volume>69</volume>
          (
          <year>2004</year>
          )
          <article-title>066138</article-title>
          . doi:
          <volume>10</volume>
          .1103/PhysRevE.69.066138.
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>N.</given-names>
            <surname>Veyrat-Charvillon</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.-X.</given-names>
            <surname>Standaert</surname>
          </string-name>
          ,
          <article-title>Mutual information analysis: How, when and why?</article-title>
          , in:
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>