=Paper= {{Paper |id=Vol-3735/paper_02 |storemode=property |title=An Empirical Study on the Robustness of Knowledge Injection Techniques Against Data Degradation |pdfUrl=https://ceur-ws.org/Vol-3735/paper_02.pdf |volume=Vol-3735 |authors=Andrea Rafanelli,Matteo Magnini,Andrea Agiollo,Giovanni Ciatto,Andrea Omicini |dblpUrl=https://dblp.org/rec/conf/woa/RafanelliMACO24 }} ==An Empirical Study on the Robustness of Knowledge Injection Techniques Against Data Degradation== https://ceur-ws.org/Vol-3735/paper_02.pdf
                         An Empirical Study on the Robustness of Knowledge
                         Injection Techniques Against Data Degradation
                         Andrea Rafanelli1,2,† , Matteo Magnini3,*,† , Andrea Agiollo3,† , Giovanni Ciatto3 and
                         Andrea Omicini3
                         1
                           Department of Computer Science, University of Pisa, Pisa, Italy
                         2
                           Department of Information Engineering, Computer Science and Mathematics, University of L’Aquila, L’Aquila, Italy
                         3
                           Department of Computer Science and Engineering (DISI), University of Bologna, Bologna, Italy


                                     Abstract
                                     Symbolic knowledge injection (SKI) represents a promising paradigm for bridging symbolic knowledge and
                                     sub-symbolic predictors in intelligent autonomous agents. Given the wide availability of SKI methods from the
                                     literature, we observe that SKI effectiveness is commonly measured in terms of predictive performance variation
                                     – e.g., accuracy improvement – introduced by SKI. However, other aspects such as the injection mechanism’s
                                     ability to maintain its performance and generalisation capability, despite encountering unexpected or anomalous
                                     input during the training process, are equally relevant. Accordingly, in this paper we propose a new metric
                                     to evaluate the robustness of SKI techniques, defined as a measure of performance degradation in response to
                                     systematic dataset variations. The proposed metric enables precise quantification of the robustness degree across
                                     injection approaches and different perturbations. Details on generating and quantifying perturbations are also
                                     provided. We evaluate the effectiveness of our metric through several experiments, where we apply multiple SKI
                                     techniques to three datasets and measure how robustness varies as perturbations increase.

                                     Keywords
                                     Symbolic Knowledge Injection, Robustness, Neural Networks




                         1. Introduction
                         In recent years, the success of sub-symbolic approaches such as neural networks (NNs) enabled compu-
                         tational agents with more and more smart behaviours ranging from speech recognition [1] to object
                         detection [2], and many more [3]. However, while reaching considerable success in tasks requiring
                         perception, sub-symbolic mechanisms still face challenges in tasks requiring reasoning and / or inter-
                         pretability of decision processes and results [4]. Moreover, the autonomous agents’ knowledge and
                         behaviours are commonly represented symbolically both at the conceptual and technological level,
                         hindering the integration of sub-symbolic components inside autonomous agents capable of complex
                         reasoning. The integration of symbolic and sub-symbolic systems (a.k.a. neuro-symbolic) has emerged
                         as a possible solution to the aforementioned problems gaining significant attention in the field of
                         autonomous agents [5] and general artificial intelligence (AI) systems [6]. In this context, most recent
                         efforts focus on the integration of NN with symbolic AI techniques aiming to augment NNs with
                         some reasoning capability—or making them more interpretable, knowledgeable, predictable, or any
                         combination of those. Among the many methods for mixing NN with symbolic AI [7], in this paper we
                         focus on those involving the injection of symbolic knowledge into NN [8].
                            Symbolic knowledge injection (SKI) [8] techniques entail the incorporation of structured and explicit
                         knowledge – available at the intelligent agent level – into sub-symbolic systems, such as NN, to improve
                         their reasoning and decision-making capabilities. Symbolic knowledge can take the form of logic

                         WOA 2024: 25th Workshop "From Objects to Agents", 7th-10th July, 2024, Forte di Bard (Italy)
                         *
                           Corresponding author.
                         †
                           These authors contributed equally.
                         $ andrea.rafanelli@phd.unipi.it (A. Rafanelli); matteo.magnini@unibo.it (M. Magnini); andrea.agiollo@unibo.it (A. Agiollo);
                         giovanni.ciatto@unibo.it (G. Ciatto); andrea.omicini@unibo.it (A. Omicini)
                          0000-0001-8626-2121 (A. Rafanelli); 0000-0001-9990-420X (M. Magnini); 0000-0003-0531-1978 (A. Agiollo);
                         0000-0002-1841-8996 (G. Ciatto); 0000-0002-6655-3869 (A. Omicini)
                                  © 2022 Copyright for this paper by its authors. Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0).


CEUR
                  ceur-ws.org
Workshop      ISSN 1613-0073
Proceedings
rules, as well as knowledge graphs, or expert knowledge in a specific domain. The available symbolic
knowledge can represent the agent’s beliefs, some well-known concepts, a societal norm or any desirable
rule for the sub-symbolic system to be considered. In any case, the goal of SKI is altering the sub-
symbolic predictor to make it compliant w.r.t. some given symbolic knowledge. Commonly, this is
attained by extending either the structure or the training process of a sub-symbolic predictor. Letting a
NN exploit prior knowledge may lead to several benefits, e.g., making the NN (i) more robust to data
quality degradation (e.g., noisy or corrupted inputs, lack of data, etc.), (ii) more predictable – hence,
more trustworthy –, (iii) less data-requiring, and, sometimes, (iv) less time-requiring to train.
   SKI methods are often measured by predictive performance variation, such as accuracy improvements,
but other aspects like efficiency [5], or the robustness of the injection mechanism also play a role. Robust
mechanisms can effectively incorporate new knowledge into NN without causing errors or deteriorating
their overall performance. In safety-critical applications, like autonomous vehicles or medical diagnosis
systems, a non-robust mechanism may introduce errors, leading to poor performances. A robust
mechanism can seamlessly adapt to new knowledge, ensuring the quality of performance and enhancing
the trustworthiness of NN systems.
   Robustness can be assessed by evaluating the injected model’s ability to maintain its performance in
the presence of data perturbations, such as noisy or corrupted training samples. Therefore, measuring
the robustness of injection mechanisms is a crucial stage in developing neuro-symbolic agents, and it
should be included in the process of evaluating any SKI approach. However, available metrics focus
specifically on purely sub-symbolic mechanisms, such as NN [9], while trustworthiness measurements
in the SKI realm are currently lacking [10]. Moreover, such metrics focus heavily on robustness under
the adversarial attack paradigm, while ignoring the robustness of the learning algorithm itself when
faced with bugged training data. Therefore, in this paper, we focus specifically on the SKI realm and
consider robustness as the resilience of its training approach over imperfect, bugged or missing data.
Accordingly, we present a comprehensive modelling of a new robustness metric for SKI, as well as an
empirical evaluation of this metric through PSyKI [11]—a software library for SKI algorithms. Our
findings provide compelling evidence that the introduction of a robustness metric for SKI is a crucial
step towards enhancing the reliability and transparency of AI systems. To the best of our knowledge,
this study represents the first attempt to introduce a robustness metric for SKI, along with the necessary
software tools for its practical implementation and evaluation.


2. Background
Robustness Robustness is a crucial challenge for NN [12], and relates to their capability to maintain
performance despite the presence of input perturbations. In general, the more a neural network is
robust the more reliable and accurate its predictions are likely to be.
    To increase the neural network robustness, a number of techniques have been proposed in the
literature, such as regularisation techniques [13], data augmentation techniques [14], and also methods
such as adversarial training [15]. The strong interest behind NN robustness is motivated by both (i) the
need to address certain limitations of NN – such as be sensitive to random perturbations [16, 17] –, and
(ii) the development of trustworthy AI (TAI).

Trustworthy AI TAI refers to the development and deployment of AI systems that are transparent,
responsible, ethical, and safe. It gained momentum in recent years, also because of the growing
popularity of AI, and the consequent attention of policymaker—cf. the guidelines set by the European
Union [18].
  Robustness is a key component of TAI, as it ensures that the AI system can function consistently and
accurately in a variety of scenarios. Along this line, research about NN trustworthiness has greatly
advanced, with more and more new possible solutions being sought. In this regard, neuro-symbolic
contributions were proposed as means to achieve more trustworthy models. In practice, these methods
aim at combining the strengths of NN and symbolic AI, mitigating the weaknesses of both—e.g., model
opacity for NN, or limited scalability for symbolic systems.

Neuro-symbolic integration Neuro-symbolic methods [19] can increase the robustness of NN, for
instance by providing them with symbolic representations of knowledge. Symbolic knowledge is charac-
terised by the use of facts and rules – commonly, in logic form –, which can facilitate the representation
of knowledge in a structured – and, possibly, intensional – manner. This can be particularly useful for
applications where training data may be noisy or incomplete. In fact, the lack of quality data may be
compensated by concise – yet general – logic rules describing the phenomena for which data is missing.
Therefore, by incorporating symbolic knowledge into NN, the general expectation is that they should
improve their performance, even in presence of issues in the training data.

Symbolic Knowledge Injection This paper focuses on symbolic knowledge injection (SKI) as an ap-
proach for neuro-symbolic integration. SKI techniques, in general, are arbitrary algorithmic procedures
that affect how sub-symbolic predictors draw their inferences, either so that predictions are computed
as a function of or made consistent with some given symbolic knowledge. To enhance reasoning and
decision-making capabilities in sub-symbolic systems, SKI requires incorporating structured knowledge,
such as logical rules, knowledge graphs, or expert knowledge [8].
   There are several methods for knowledge injection [8], including constraining-based methods,
structuring-based methods, and embedding symbolic knowledge within NN—the latter laying outside
the scope of this paper. In all such cases, SKI is performed by modifying the sub-symbolic predictor to
conform with input knowledge, and then trained and exploited “as usual”. Accordingly, in the remainder
of this paper, we denote any injection procedure as 𝑁   ^ = ℐ(𝑁, 𝐾, 𝐷), where 𝑁 denotes the predictor
before injection, 𝐾 is the symbolic knowledge to be injected, 𝐷 is the training dataset, ℐ is the injection
procedure, and 𝑁  ^ is the predictor with injected knowledge. Informally, we may refer to 𝑁       ^ as the
educated predictor, and 𝑁 as the uneducated one.

Performance scores Effectiveness of SKI is commonly measured by comparing the performance
gain (e.g. accuracy, F1-score, MSE, etc.) achieved by the SKI predictor w.r.t. its uneducated counterpart.
This metric indicates the quality of the SKI approach in predictive performance but does not capture
aspects related to injection mechanism quality, such as robustness. Therefore, in the remainder of this
paper, we identify a metric to measure the actual robustness of the SKI mechanism compared to its
uneducated counterpart.


3. Robustness Score for SKI
In counterfactual analysis [20], it is a common practice to exploit controlled perturbation on the training
data to assess the robustness of predictors. Taking inspiration from this technique, in this section we
introduce the notion of statistical robustness of SKI procedure. Accordingly, we consider an injection
mechanism as robust if the predictive performance of the educated predictors is poorly affected by
perturbations of the training data—as long as the perturbation magnitude is small. The remainder
of this section provides a general description of robustness, before delving into the details of which
perturbations types can be applied to training data and how their magnitude may be measured.

Data perturbation We define data perturbation as altering a training dataset 𝐷 by adding, removing,
or editing its entries, and denote it by Δ𝐷. Accordingly, we denote the perturbed dataset as 𝐷′ =
𝐷 ∘ Δ𝐷, where ∘ is the perturbation application operator. We also denote by ‖Δ𝐷‖ ∈ R≥0 the
magnitude of the perturbation—i.e., the scalar value quantifying the amount of changes induced by the
perturbation.
Robustness score Let D = {Δ𝐷1 , . . . , Δ𝐷𝑛 } be a set of potential data perturbations to be applied
to some dataset 𝐷, let 𝑁 be a predictor of any sort – trained on 𝐷 –, and let 𝑁Δ𝐷 be the predictor
having the same hyperparameters of 𝑁 , yet being trained upon the perturbed dataset 𝐷 ∘ Δ𝐷. Under
such hypotheses, we define the robustness score of 𝑁 w.r.t. D, as follows:

                                          1 ∑︁        𝜋(𝑁Δ𝐷 , 𝐷 ∘ Δ𝐷)
                            𝜌𝑁,𝐷 (D) =         ‖Δ𝐷‖ ·                                                     (1)
                                          𝑛              𝜋(𝑁, 𝐷)
                                            Δ𝐷∈D

where 𝜋 is a performance metric of choice, such as accuracy, computing the performance of the input
predictor w.r.t. the input dataset.
   As the reader may notice, robustness is directly proportional to (i) the magnitude of the perturbations,
and (ii) the ratio among the performance of the perturbed predictors and the performance of the
unperturbed predictor. We would like to emphasise that, although it is true that every component in
Equation (1) has a positive value, this does not imply that every data perturbation leads to a positive
impact on performance. Indeed, our robustness measure incorporates a performance ratio, where both
the numerator and denominator are bounded between 0 and 1 (as the performance metric itself is
considered to range from 0 to 1). As a result, the performance ratio can be slightly above 1 when
the perturbed model performs better – i.e., positive perturbation where 𝜌𝑁^ ,𝐷 (D) > 𝜌𝑁,𝐷 (D) –,
while it falls below 1 when the perturbed model performs worse—i.e. negative perturbation where
𝜌𝑁^ ,𝐷 (D) < 𝜌𝑁,𝐷 (D). Practically speaking, when positive perturbations occur, our metric produces a
ratio approaching 1. This indicates merely a marginal gain, as ratios higher than 1 denote performance
improvements over the unperturbed model. Ratios below 1 point to decrements in performance. In
other words, the robustness of a predictor increases when relatively big perturbations in the training
data have a relatively small effect on the overall predictor’s performance, i.e. the more robust the model,
the less its performance is disrupted by significant modifications to its training inputs.
   The robustness of some injection mechanism can be measured by applying Equation (1) to some edu-
cated predictor 𝑁^ = ℐ(𝑁, 𝐾, 𝐷), attained by injecting the knowledge 𝐾, on some uneducated predictor
𝑁 , then trained upon 𝐷—which we denote denote by 𝜌𝑁^ ,𝐷 (D). One may also be interested in under-
standing whether some injection mechanism makes the predictor more or less robust than its uneducated
counterpart. To this end, we define the robustness gain score as follows: 𝑅𝑁,𝐷 (ℐ) = 𝜌𝑁^ ,𝐷 (D)/𝜌𝑁,𝐷 (D).
Here, the robustness gain is a positive measure that indicates if the injection mechanism ℐ produces
a more robust predictor (𝑅𝑁,𝐷 (ℐ) > 1) w.r.t. its uneducated counterpart over data perturbation.
Meanwhile, injection mechanisms suffering data perturbations result in 𝑅𝑁,𝐷 (ℐ) < 1, as they pro-
duce an educated model 𝑁    ^ less effective for dealing with perturbed data. Therefore, 𝑅𝑁,𝐷 (ℐ) is an
easy-to-understand measure for analysing the robustness quality of a selected SKI mechanism.

3.1. Measuring Data Perturbations
Equation (1) defines robustness by relying on the possibility of measuring the magnitude ‖Δ𝐷‖ of any
given perturbation Δ𝐷. While being easy to model in theory, the problem is hard to tackle in practice.
As far as this paper is concerned, we focus on the simpler goal of measuring the difference among any
two datasets 𝐴, 𝐵. To serve this purpose, we leverage on the Kullback-Leibler (KL) divergence [21].
In theory, the KL-divergence is a statistical operator aimed at measuring the difference among any
two probability distributions 𝛼 and 𝛽. In the particular case where (i) 𝛼 and 𝛽 are multivariate normal
distributions having the same dimensionality, and (ii) two datasets 𝐴, 𝐵 are sampled from 𝛼 and 𝛽
respectively, the KL-divergence can be computed as follows:

                     1 [︁                             (︁ det 𝜎 )︁
                                                               𝐵
                                                                                                  ]︁
         𝜓(𝐴, 𝐵) =           𝜎 −1
                          tr(𝜎 𝐵  𝜎 𝐴 ) − dim(𝐴) + ln             +  𝜇
                                                                    (𝜇 𝐵 − 𝜇 𝐴 ) ⊤ −1
                                                                                  𝜎 𝐵  𝜇
                                                                                      (𝜇 𝐵 − 𝜇 𝐴 )        (2)
                     2                                   det 𝜎 𝐴
where 𝜇 𝐴 (resp. 𝜇 𝐵 ) represents the mean of 𝐴 (resp. 𝐵), and 𝜎 𝐴 (resp. 𝜎 𝐵 ) represents its covariance
matrix, det 𝜎 𝐴 (resp. det 𝜎 𝐵 ) is its determinant, and tr(𝜎
                                                            𝜎 𝐴 ) (resp. tr(𝜎
                                                                            𝜎 𝐵 )) its trace. Finally, dim(𝐴)
represents the dimensionality of 𝐴—i.e., the number of features characterising each of its entries.
  The normality hypothesis may appear restrictive, as datasets 𝐴 and 𝐵 may, in the general case,
be sampled from unknown distributions. However, any unknown probability distribution can be
approximated by a mixture of normal distributions – as the latter are universal approximators for
probability densities [22, Sec. 3.9.6] –, and the computation of the KL-divergence for a mixture of
normal distributions can be approximated to that of a single normal distribution—as discussed in Joyce
[21, Sec. 2.2.2]. So, in practice, Equation (2) can be exploited to estimate the difference among any two
datasets 𝐴 and 𝐵.

KL-divergence for classification problems. For classification problems, the KL divergence is
commonly computed per-class, as different classes may come with different distributions. In case 𝐾
classes are available, we consider datasets as partitioned into disjoint sub-sets, on a per-class basis: i.e.,
𝐴 = 𝐴1 ∪ . . . ∪ 𝐴𝐾 and 𝐵 = 𝐵1 ∪ . . . ∪ 𝐵𝐾 , where 𝐴𝑘 (resp. 𝐵𝑘 ) is the set of samples belonging to
class 𝑘 in 𝐴 (resp. 𝐵).
   We here propose to define the overall divergence score between 𝐴 and         𝐵 as the weighted average
                                                                          1 ∑︀𝐾
of class-specific Kullback-Leibler divergences, namely: Ψ(𝐴, 𝐵) = |𝐴|         𝑘=1 𝜓(𝐴𝑘 , 𝐵𝑘 ) · |𝐴𝑘 |. where
|𝐴| (resp. |𝐴𝑖 |) represents the cardinality of 𝐴 (resp. 𝐴𝑖 ). The weighted average is necessary to tackle
very unbalanced datasets with diverging distributions over unpopular labels. Therefore, the proposed
approach allows to quantify the distribution changes for each label, while also focusing on the points
where the actual change occurred, handling cases where the perturbation may significantly affect only
some classes.
   Accordingly, we estimate the magnitude of a perturbation Δ𝐷 transforming 𝐷 into 𝐷′ = 𝐷 ∘ Δ𝐷
by means of the overall divergence among 𝐷 and 𝐷′ : ‖Δ𝐷‖ = Ψ(𝐷, 𝐷′ ). The proposed metric lays in
R≥0 : it is equal to zero for two identical datasets – i.e., when the perturbation Δ𝐷 has no effect –, and
it gets higher in value as 𝐷′ differs from 𝐷—i.e., proportionally to the effect of ‖Δ𝐷‖.
   It is worth highlighting that the proposed perturbation metric ‖·‖ can be applied to any sort of
perturbation, as it does not take into account how the perturbation Δ𝐷 affects the dataset or its samples.
For this reason, the metric can be exploited to measure the many sorts of perturbations discussed in
Section 3.2.

3.2. Perturbation Strategies
Here we discuss three major perturbation strategies for altering a dataset, namely: (i) sample drop,
(ii) noise addition, and (iii) label flipping. These strategies mimic common issues that can degrade
a dataset’s quality, as they involve randomly deleting samples, adding random noise, and randomly
changing labels.

Sample Drop This perturbation strategy mimics the effect of an intelligent agent lacking training data.
Hence, it aims at selectively mutilating a dataset in a controlled way. This strategy is commonly exploited
to test if and to what extent neuro-symbolic approaches are effective to deal with data scarcity [23].
   The basic idea behind sample drop is to randomly remove some samples from 𝐷 to obtain 𝐷′ . The
whole process is stochastic and controlled by a single parameter, namely the dropping probability –
denoted by 𝑝 ∈ ]0, 1[ –, which is shared among all data entries in 𝐷. Generally speaking, 𝑝 = E[𝑋𝑑 ]
represents the mean of the Bernoulli-distributed random variable 𝑋𝑑 ∼ ℬ(𝑝), dictating whether the
data entry 𝑑 ∈ 𝐷 should (P(𝑋𝑑 = 0)) or should not (P(𝑋𝑑 = 1)) be included in 𝐷′ . Under such
hypotheses, the perturbed dataset 𝐷′ is such that 𝐷′ = {𝑑 | ∀𝑑 ∈ 𝐷 s.t. 𝑋𝑑 = 0}.
   We consider constructing the set of data perturbations D = {Δ𝐷1 , . . . , Δ𝐷𝑛 } by applying the
sample drop process on 𝐷 multiple times – namely, 𝑛 times – with increasing dropping probabilities
0 < 𝑝1 < . . . < 𝑝𝑛 < 1.

Noise Addition This perturbation strategy mimics the situation where the data sampling / acquisition
process of the autonomous agent is affected by error—e.g., due to sensor noise. Hence, it aims at
degrading a dataset in a controlled way. This strategy is commonly exploited to test if and to what
extent neuro-symbolic approaches are effective to deal with unreliable (e.g., corrupted, inconsistent)
data [24].
  The basic idea behind noise addition is to add some random noise to the data entries in 𝐷, to obtain 𝐷′ .
The whole process is stochastic, and it is controlled by a single parameter, namely the noise intensity –
denoted by 𝑣 ∈ R≥0 –, which is shared among all data entries in 𝐷. Generally speaking1 , 𝑣 ·1 = cov(𝑉𝑑 )
represents the covariance matrix of the 0-mean, multi-variate, normally distributed random variable
𝑉𝑑 ∼ 𝒩 (0, 𝑣 · 1), representing the random noise to be applied to each entry 𝑑 ∈ 𝐷. Under such
hypotheses, the perturbed dataset 𝐷′ can be described as 𝐷′ = {𝑑 + 𝑉𝑑 | ∀𝑑 ∈ 𝐷}. As the reader
may notice, the perturbed dataset is characterised by the same number of samples of 𝐷—meaning that
|𝐷| = |𝐷′ |.
  We consider constructing the set of data perturbations D = {Δ𝐷1 , . . . , Δ𝐷𝑛 } by applying the
noise addition process on 𝐷 multiple times – namely, 𝑛 times – with increasing noise intensities
0 < 𝑣1 < . . . < 𝑣𝑛 .

  About discrete features. Normal noise can be applied to any continuous feature in a dataset.
However, when dealing with datasets with ordinal or categorical features, it is necessary to adjust
the noise addition process to account for discontinuity. Accordingly, to deal with discrete features,
the additive normal noise is discretised by replacing the distribution 𝒩 (0, 𝑣 · 1) with its discrete
counterpart 𝒩¯ [25]. Overall, this means that for ordinal features, additive noise is more likely to choose
an additional ordinal value close to the original one, whereas for categorical features, additive noise
may choose an additional ordinal value with uniform probability—analogously to the label flipping case
described below.

Label Flipping This perturbation strategy mimics the situation where some data labelling process is
affected by error—e.g., human operator mislabelling some data entries. Hence, it aims at selectively
flipping the labels of some entries in a dataset. The underlying assumption is that the dataset consists
of input and output features, and the output ones are either binary or categorical. Therefore, this
perturbation strategy mostly makes sense for classification tasks.
   The basic idea behind label flipping is to randomly alter the output features of some data entries in
𝐷, to produce 𝐷′ . Flipping, in particular, requires each output feature to be randomly re-assigned with
some value in the output domain—of course different from the original one.
   Without loss of generality, we consider the case of a single output feature having 𝐾 admissible
values—i.e., 𝐾 classes represented as a single categorical attribute. Hence, for an 𝑚-dimensional dataset
𝐷, we let 𝑑 ≡ (𝑥𝑑,1 , . . . , 𝑥𝑑,𝑚−1 , 𝑦𝑑 ) ∈ 𝐷 denote the generic data entry in 𝐷, where the many 𝑥𝑑,𝑗 are
the values of the input features of 𝑑, and 𝑦𝑑 ∈ {1, . . . , 𝐾} is the value of its output feature.
   Similarly to the cases above, we model label flipping as a stochastic process controlled by the flipping
probability parameter – denoted by 𝑓 ∈ ]0, 1[ –, representing the likelihood that the label of any given
data entry may flip. Parameter 𝑓 controls the probability distribution of the random variable 𝑌𝑑 , which
represents the novel value of the output feature for data entry 𝑑 ∈ 𝐷, after flipping. In particular,
we define the probability density of 𝑌𝑑 as follows: P(𝑌𝑑 = 𝑘) = 𝑓 /(𝐾 − 1) if 𝑘 ̸≡ 𝑦𝑑 ; otherwise
P(𝑌𝑑 = 𝑦𝑑 ) = (1 − 𝑓 ). In other words, the value of the output feature may flip with probability 𝑓 ,
and in that case each value different than the original one is equally likely to be selected. Under such
hypotheses, the perturbed dataset 𝐷′ is such that 𝐷′ = {(𝑥𝑑,1 , . . . , 𝑥𝑑,𝑚−1 , 𝑦𝑑′ ) | ∀𝑑 ∈ 𝐷 : 𝑌𝑑 = 𝑦𝑑′ }.
   We consider constructing the set of data perturbations D = {Δ𝐷1 , . . . , Δ𝐷𝑛 } by applying the label
flipping process on 𝐷 multiple (i.e., 𝑛) times, with increasing flipping probability 0 < 𝑓1 < . . . < 𝑓𝑛 < 1.

  About regression problems. Label flipping is not directly applicable to regression problems, as
the output feature is not categorical. However, it may be modelled as a particular sort of noise addition

1
    1 is the 𝑚 × 𝑚 identity matrix, 0 is the 𝑚 × 1 zero vector, where 𝑚 is dimensionality of the dataset.
affecting only output feature(s). Nevertheless, the application of label flipping to regression problems is
beyond the scope of this paper.


4. Experiments
In this section we present the setups and results of the experiments for evaluating the effectiveness
of the proposed robustness metric. For sake of reproducibility, the source code of our experiments is
public and available.2

Datasets We rely on three different datasets from the UCI repository,3 namely the “Breast Cancer
Wisconsin” (BCW) dataset,4 the “Primate Splice Junction Gene Sequences” (PSJGS) dataset,5 and “Census
Income” (CI) dataset.6 Subsequently, we present in detail each of the datasets.

   Breast Cancer The BCW dataset contains 699 instances of breast cancer clinical exams, each with
9 features summarising biological characteristics. The feature BareNuclei has 16 missing values, which
are replaced with the value zero. The dataset is divided into benign and malignant classes, with 458
(65.5%) benign samples and 241 (34.5%) malignant samples.

  Splice Junction The PSJGS dataset contains 3,190 gene sequence instances, each described by 60
nucleotides. Values are adenine (a), cytosine (c), guanine (g), and thymine (t). However, other symbols
may be used to indicate that more than one specific nucleotide can appear in the same position. For
this reason, we perform a multi-hot encoding of the features, resulting in 240 boolean features. The
dataset has three classes, exon-intron (ei), intron-exon (ie), and none of the previous (n), with 767
(24%), 768 (24.1%), and 1655 (51.9%) instances respectively.

   Census Income The CI dataset contains 48,842 instances representing individuals, with 14 fea-
tures representing different attributes. We opt to remove Education as EducationNumeric is an
ordered representation of the same feature. Fnlwgt and Race features are avoided to avoid social
bias. The remaining nominal features (WorkingClass, MaritalStatus, Occupation, Relationship,
NativeCountry) are one-hot encoded. Classes are binary and represent the yearly income of a person,
less/equal than/to 50, 000 USD and more than 50, 000 USD, with 37, 155 (76.1%) and 11, 687 (23.9%)
instances respectively.

Injected Knowledge For our experiments we exploit specific knowledge for each dataset, using
Prolog syntax [26] for classification logic rules The PSJGS dataset already has a knowledge base
in the literature, with biological rules provided by human domain experts [27]. We use this very
knowledge as well, with the addition of an explicit third rule to classify the n class: class(𝑋   ¯ , n) ←
¬class(𝑋 , ei)∧¬class(𝑋 , ie). where 𝑋 is an abbreviation for the full sequence of variables X−30 , . . . ,
         ¯                 ¯              ¯
X+30 that represent the input data of a 60-basis long DNA sequence. The rule simply states that if a
record is not classified as ei or ie then it is classified as n. For the remaining datasets – BCW and CI –
there are no predefined rules in the literature. Therefore, we define a custom set of logic rules for both
of them.




2
  https://anonymous.4open.science/r/ski-data-robustness-experiments
3
  https://archive.ics.uci.edu/ml/index.php
4
  https://archive.ics.uci.edu/dataset/15/breast+cancer+wisconsin+original
5
  https://archive.ics.uci.edu/dataset/69/molecular+biology+splice+junction+gene+sequences
6
  https://archive.ics.uci.edu/dataset/20/census+income
Injectors We rely on PSyKI7 (Platform for Symbolic Knowledge Injection) for the SKI algorithms in
the experiments. PSyKI enables sub-symbolic predictors incorporate prior symbolic knowledge. It is
compatible with Tensorflow-deployed8 predictors and allows for various SKI algorithms to be used:

       • KINS, Knowledge Injection via Network Structuring [28] is a structuring-based injection mech-
         anism where an ordinary multi-layer neural network is extended with ad-hoc neural modules
         aimed at mimicking the provided symbolic knowledge. The weights of the neural modules are
         trained together with the weights of the original neural network.

       • KILL, Knowledge Injection via Lambda Layer [29] is a guided-learning-based approach that
         constrains the training phase as follows. Before each back-propagation step, a penalty value
         that represents the degree of violation w.r.t. the input knowledge is added to the loss function.
         Therefore, the training phase is forced to minimise the loss function and the penalty value at the
         same time—hence maximising the compliance w.r.t. the knowledge to be injected.

       • KBANN, Knowledge-Based Artificial Neural Network [27] is one of the first structuring-based
         SKI algorithms proposed in the literature. It creates a NN from a set of propositional rules: each
         piece of each formula is converted into partial neural structures, to be then composed in a single
         network. The main difference w.r.t. to KINS is that the network created by KBANN is entirely
         constructed from the knowledge, while KINS extends an existing network with ad-hoc modules.

4.1. Experimental Setup
For our experiments, we leverage three datasets, three input knowledge bases, three different injectors,
and three perturbation strategies. For each dataset we train an uneducated model, which is then applied
to the three injectors using the corresponding knowledge base. The uneducated model has two hidden
layers, one input, and one output layer, with rectified linear units (ReLU) as the activation function,
and softmax for the output layer. The number of neurons per hidden layer changes depending on the
dataset—namely, [16, 8] for BCW, [32, 16] for CI, and [64, 32] for PSJGS. Training involves categorical
cross-entropy loss function, a batch size of 32 elements, and an epoch limit of 100.
   Data perturbation is applied multiple times for each dataset-model configuration, using sample drop,
noise addition, and label flipping. More precisely, the drop probability (𝑑) parameter varies from 0.0
to 0.95 with a step of 0.05, the flipping probability (𝑓 ) parameter varies from 0.0 to 0.90 with a step of
0.08, and the noise intensity (𝑣) parameter varies from 0.0 to 1.0 with a step of 0.1. Each experiment is
repeated 30 times to draw statistically relevant comparisons.
   We point out that our robustness analysis exclusively focuses on symbolic knowledge injection
(SKI) approaches. In this context, we did not find any existing previous measures allowing a direct
comparison. The only related available metric [30] focuses on generalisation capability rather than
training robustness. Therefore, we here consider to solely analyse the goodness of the proposed
robustness metric, aiming at checking if it allows to identify robust SKI mechanisms.

4.2. Results and Discussion
Here we present the results of our experiments (Figure 1) over different perturbation strategies.
   To better understand the real differences in performance between educated and uneducated
predictors—which are fairly comparable—we compare the average accuracy of each predictor w.r.t. the
uneducated one. For this purpose, we employ the Mann-Whitney U Test [31], a non-parametric test for
differences between two groups. In this setting, a p-value ≥ 0.05 indicates that there are no significant
differences between the two average accuracy distributions, whereas a p-value < 0.05 indicates the
opposite.


7
    https://github.com/psykei/psyki-python
8
    https://www.tensorflow.org/?hl=it
                                                           (a)                                                                                                               (b)                                                                                                                   (c)
                     0.96
                                                breast-cancer with data drop                                                                                    splice-junction with data drop                                                            0.85
                                                                                                                                                                                                                                                                                  census-income with data drop
                                                                                                                                        0.95
                     0.94                                                                                                               0.90




                                                                                                                                                                                                                                       Accuracy on test set
                                                                                                                     Accuracy on test set
  Accuracy on test set

                                                                                                                                                                                                                                                          0.85
                     0.92                                                                                                               0.85
                                                                                                                                        0.80                                                                                                              0.84
                     0.90
                                                                                                                                        0.75
                     0.88                                                                                                                                                                                                                                 0.84          UNEDUCATED
                                  UNEDUCATED                                                                                            0.70          UNEDUCATED                 p-values
                                                                                                                                                                                                                                                                                                    p-values
                                                               p-values
                                                                                                                                                      KINS                            0.09                                                                              KINS                             0.13
                     0.86         KINS                              0.01                                                                                                                                                                                                KILL                             0.46
                                  KILL                              0.97                                                                0.65          KILL                            0.36                                                                0.83          KBANN                            0.0
                     0.84         KBANN                             0.0                                                                               KBANN                           0.0




                                                                                                                                                                                                                                                                 293 33 (d=0%)
                                                                                                                                                                                                                                                                 27605 (d= 5%)




                                                                                                                                                                                                                                                                  6510 (d=70%)
                                                                                                                                                                                                                                                                  4882 (d=85%)
                                                                                                                                                                                                                                                                  3254 (d=80%)
                                                                                                                                                                                                                                                                  1626 (d=95%)
                                                                                                                                                                                                                                                                      8 (d 0 % )
                                                                                                                                                                                                                                                                              %)
                                                                                                                                                                                                                                                                 22721 (d=20%)
                                                                                                                                                                                                                                                                 21193 (d=25%)
                                                                                                                                                                                                                                                                 19565 (d=30%)
                                                                                                                                                                                                                                                                 17937 (d=35%)
                                                                                                                                                                                                                                                                 16209 (d=40%)
                                                                                                                                                                                                                                                                 14681 (d=45%)
                                                                                                                                                                                                                                                                 13053 (d=50%)
                                                                                                                                                                                                                                                                 11325 (d=55%)
                                                                                                                                                                                                                                                                  9766 (d=60%)
                                                                                                                                                                                                                                                                  8148 (d=75%)
                                                                                                                                                                                                                                                                 26077 (d=10%)
                                                                                                                                                                                                                                                                 24449 (d=15%)
                                                                                                                                               1708 (d=10%)
                                                                                                                                               1591 (d=25%)
                                                                                                                                               1485 (d=20%)

                                                                                                                                               1272 (d=30%)
                                                                                                                                               1176 (d=45%)
                                                                                                                                               1060 (d=40%)
                                                                                                                                                957 (d=50%)
                                                                                                                                                851 (d=55%)
                                                                                                                                               1389 (d=35%)




                                                                                                                                                744 60 %)
                                                                                                                                               191 0 (d=5%)
                                                                                                                                               1804 (d=1 %)




                                                                                                                                                638 (d=65%)
                                                                                                                                                532 (d=70%)
                                                                                                                                                425 (d=75%)
                                                                                                                                                319 (d=80%)
                                                                                                                                                213 (d=85%)
                                                                                                                                                106 (d=90%)
                                                                                                                                                    (d= %)
                                                                                                                                                            )
                                                                                                                                                        95%
                            419 (d=5%)
                            396 (d=10 )




                             47 ( 85%)
                             23 (d=90%)
                                  d=9 )
                                     5%)
                            373 (d=15%)
                            350 (d=20%)
                            326 (d=25%)
                            303 (d=30%)
                            280 (d=35%)
                            256 (d=40%)
                            233 (d=45%)
                            210 (d=50%)
                            186 (d=55%)
                            163 (d=60%)
                            140 (d=65%)
                            117 (d=70%)
                             93 ((d=75%)
                             70 (d=80%)




                                                                                                                                                                                                                                                                          =95
                             443 (d=0%




                                      %




                                                                                                                                                         5




                                                                                                                                                                                                                                                                  30961 (d=




                                                                                                                                                                                                                                                                           6
                                                                                                                                                2027 (d=0




                                                                                                                                                       =
                                   =




                                                                                                                                                      d
                                  d




                                                                                                                                                    (
                                                                                                                                                  3




                                                                                                                                                                                                                                                                  325




                                                                                                                                                                                                                                                                    9
                             466




                                                                                                                                                212
                            Cardinality of the training set and dropping probability                                                           Cardinality of the training set and dropping probability                                                          Cardinality of the training set and dropping probability

                                                           (d)                                                                                                                (e)                                                                                                                  (f)
                                           breast-cancer with noise addition                                                                                   splice-junction with noise addition                                                        0.85
                                                                                                                                                                                                                                                                                census-income with noise addition
                                                                                            UNEDUCATED                                  0.94                                                                   UNEDUCATED                                                                                                        UNEDUCATED
                                                                                            KINS                                                                                                               KINS                                       0.85                                                                   KINS
                     0.94                                                                   KILL                                        0.92                                                                   KILL                                                                                                              KILL
                                                                                            KBANN                                                                                                              KBANN                                      0.85                                                                   KBANN
  Accuracy on test set




                                                                                                                     Accuracy on test set




                                                                                                                                                                                                                                       Accuracy on test set
                                                                                                                                        0.90                                                                                                              0.84
                     0.92
                                                                                                                                        0.88                                                                                                              0.84
                     0.90                                                                                                               0.86                                                                                                              0.84

                                                                                                                                        0.84                                                                                                              0.84
                             p-values                                                                                                            p-values                                                                                                          p-values
                     0.88         0.74                                                                                                                0.65                                                                                                0.84          0.55
                                  0.95                                                                                                  0.82          0.74                                                                                                              0.47
                                  0.07                                                                                                                0.69                                                                                                0.83          0.74
                                    0.0   0.1     0.2    0.3   0.4        0.5   0.6   0.7   0.8     0.9       1.0                       0.80            0.0   0.1   0.2    0.3   0.4        0.5    0.6   0.7   0.8     0.9       1.0                                      0.0   0.1   0.2    0.3    0.4        0.5   0.6   0.7   0.8     0.9       1.0
                                                        Noise intensity (v)                                                                                               Noise intensity (v)                                                                                               Noise intensity (v)

                                                           (g)                                                                                                               (h)                                                                                                                   (i)
                                            breast-cancer with label flipping                                                                                  splice-junction with label flipping                                                                               census-income with label flipping
                                                                                                  p-values                                                                                                           p-values                                                                                                          p-values
                                                                                                       0.95                             0.90                                                                              0.9                             0.80                                                                              1.0
                                                                                                       1.0                                                                                                                1.0                                                                                                               0.74
                     0.80                                                                              0.95                             0.80                                                                              0.03                                                                                                              0.9
                                                                                                                                                                                                                                                          0.70
  Accuracy on test set




                                                                                                                     Accuracy on test set




                                                                                                                                                                                                                                       Accuracy on test set
                                                                                                                                        0.70
                                                                                                                                                                                                                                                          0.60
                     0.60                                                                                                               0.60
                                                                                                                                                                                                                                                          0.50
                                                                                                                                        0.50
                     0.40                                                                                                                                                                                                                                 0.40
                                                                                                                                        0.40
                                  UNEDUCATED                                                                                            0.30          UNEDUCATED                                                                                          0.30          UNEDUCATED
                     0.20         KINS                                                                                                                KINS                                                                                                              KINS
                                  KILL                                                                                                  0.20          KILL                                                                                                0.20          KILL
                                  KBANN                                                                                                               KBANN                                                                                                             KBANN
                                                                                                                                        0.10
                                  0.00 0.09 0.18 0.27 0.36 0.45 0.54 0.63 0.72 0.81 0.90                                                              0.00 0.09 0.18 0.27 0.36 0.45 0.54 0.63 0.72 0.81 0.90                                                            0.00 0.09 0.18 0.27 0.36 0.45 0.54 0.63 0.72 0.81 0.90
                                                    Flipping probability f                                                                                            Flipping probability f                                                                                            Flipping probability f


Figure 1: Average accuracy over different datasets with different perturbation strategies


Table 1
Robustness relative scores. Bold numbers are the ones greater than 1 (i.e., the educated model is more robust
than the uneducated one).
                                                                                𝑅𝑁,𝐷 (ℐ) drop                                                                                𝑅𝑁,𝐷 (ℐ) noise                                                                                       𝑅𝑁,𝐷 (ℐ) flip
                         Dataset
                                                               KINS                         KILL                    KBANN                                       KINS                              KILL               KBANN                                       KINS                         KILL                    KBANN
                         BCW                             1.0493                       1.0318                        1.0382                                    0.9960                         0.9985                     1.0109                                   0.9994                1.0184                               0.9520
                         PSJGS                           1.0045                        0.9968                        0.8425                                   0.9950                         0.9984                     1.0145                                   0.9962                1.0026                              1.6749
                         CI                               0.9998                      1.0039                        1.0043                                    0.9992                        1.0012                       0.9965                                  0.9897                1.1703                               0.9815


Drop In the data drop experiments (Figure 1 a-c), KBANN is the only predictor that shows significant
differences w.r.t. the uneducated one. Specifically, KBANN demonstrates improved performance on
the BCW (Figure 1a) and CI (Figure 1c) datasets, but its performance on the PSJGS (Figure 1b) dataset
rapidly declined when 60% of the data are dropped (𝑑 = 0.6). For the other educated predictors, KINS
shows slightly better performances than the uneducated model. In the BCW dataset its p-value = 0.01
indicates that there are significant differences in terms of performance compared to the uneducated
model.
   Looking at Table 1, one can see the influence of the performance values on the robustness score, e.g.
for KBANN in the PSJGS dataset. Apart from this, the educated models improve robustness in 6 out of 9
experiments.

Noise The noise experiments (Figure 1 d-f) reveal that SKI mechanisms are more sensitive to noisy
data than missing data. In fact, the average accuracy of predictors trained over noisy data drops more
quickly than the data drop scenario. This is due to the downward trends exhibited by all predictors
since the early stages of the curves. Indeed, Table 1 confirms that educated models enhance robustness
in only 3 experiments out of 9.
   KBANN outperforms other SKI methods and the uneducated model in the BCW (Figure 1d), but
performs poorly in the CI (Figure 1f) dataset. In PSJGS (Figure 1e) dataset, it has low performances too
w.r.t. the uneducated predictor —except for high noise levels, where it is subject to relatively slower
performance decline. This explains why in Table 1 KBANN’s robustness score shows a gain over the
uneducated one. The opposite trend can be observed for the CI scenario.
   Along this line, it is worth of notice observing that KINS – despite being a SKI approach based
on neural structuring like KBANN – performs noticeably lower than KBANN in terms of robustness,
regardless of the dataset. The discrepancy may be caused by the trainable nature of KINS’ neural
modules, which makes them more prone to overfitting noisy data. It is noticeable that KILL consistently
exhibits similar (or worse) behaviour across all datasets than the uneducated model. This may imply
that the penalties imposed by KILL during training for performing injection are not compensating for
noise perturbations.

Label-flipping The label flipping experiments (Figure 1 g-i) show that predictors behave similarly
in both BCW (Figure 1g) and CI (Figure 1i) datasets. This similarity between all predictors can also
be observed when looking at p-values which are all close to 1—i.e. no significant difference between
educated and uneducated predictors. In both, BCW and CI, there is a quick degradation in performance
starting from 54% of flipped labels (𝑓 = 0.54). Conversely, for the PSJGS dataset (Figure 1h), the
performance decline pattern is linear. One notable exception is KBANN, that exhibits an impressive
performance, retaining nearly 70% accuracy even for highest flipping probability values (𝑓 = 0.9)—as
opposed to other models whose accuracy drop to 20%. Such behaviour is further emphasised by Table 1,
where KBANN demonstrates a remarkable improvement in robustness w.r.t. the uneducated model.
   Except for KBANN, the other predictors exhibit similar performance trends w.r.t. the uneducated
model across all three scenarios. Looking at Table 1, KBANN generally experiences a slight decrease in
robustness compared to the uneducated model—except for the PSJGS case. On the other hand, KILL
emerges as the predictor that best withstands the perturbations caused by label flipping.

Discussion and take-home message These experiments show that among the perturbations ana-
lysed, data drop elicits the greatest robustness gains from SKI methods, indicating their ability to
compensate for missing information through knowledge integration. The PSJGS dataset is an exception,
indeed the injected knowledge is far from optimal, being good for classifying only one class out of three.
In this case, poor prior knowledge combined with missing data may hinder the model from learning
underlying patterns. Regarding noise perturbations, the SKI methods employed prove insufficient to
improve robustness. For label flipping perturbations, the constraining method—KILL—demonstrates
good robustness across all three datasets, suggesting that penalty added during training is sufficient to
compensate for label flipping. The exception to KBANN on PSJGS likely stems from its strong adherence
to the injected knowledge, which provides useful joint information on splice junctions and mitigates
the influence of flipped labels.
   In summary, the experimental findings are: (i) SKI approaches demonstrate heightened robustness
when data is limited, owed to the alternate guidance of integrated knowledge. (ii) Loss-manipulating
SKI techniques like KILL better tolerate label corruptions because their algorithm de-emphasises
possibly-flawed labels during backpropagation. (iii) SKI structuring methods like KBANN are more
robust than constrained-layer types like KINS since injected knowledge is maintained intact.


5. Conclusions
In this paper we propose a novel metric to assess the robustness of SKI mechanisms. We define three
different types of data perturbation strategies along with a way to assess their intensity. We also provide
a formula to calculate the robustness score of a predictor, gauging its robustness when the data is
perturbed. Furthermore, we introduce a comparison to understand whether the SKI predictor is better
(or not) than its uneducated counterpart from a robustness perspective.
   Overall, our experiments indicate that our metric can be used to effectively determine the robustness
of the performance variations obtained through SKI. These findings serve as a solid foundation that
paves the way for a range of potential evaluations of model robustness, including those in the realms of
symbolic and sub-symbolic integration. By evaluating the robustness of educated predictors, we gain a
better understanding of their performance under varying circumstances and enhance their ability to
manage uncertainty and complexity inherent in real-world scenarios, thus leading to trustworthy and
effective AI systems.

Acknowledgements
This work has been partially supported by: (i) “FAIR–Future Artificial Intelligence Research”, Spoke
8 “Pervasive AI” (PNRR, M4C2, Investimento 1.3, Partenariato Esteso PE00000013), funded by the EC
under the NextGenerationEU programme; and by (ii) “ENGINES — ENGineering INtElligent Systems
around intelligent agent technologies” project funded by the Italian MUR program “PRIN 2022” (G.A.
20229ZXBZM).


References
 [1] A. B. Nassif, I. Shahin, I. B. Attili, M. Azzeh, K. Shaalan, Speech Recognition Using Deep Neural
     Networks: A Systematic Review, IEEE Access 7 (2019) 19143–19165. doi:10.1109/ACCESS.2019.
     2896880.
 [2] Z.-Q. Zhao, P. Zheng, S.-t. Xu, X. Wu, Object Detection With Deep Learning: A Review, IEEE
     Transactions on Neural Networks and Learning Systems 30 (2019) 3212–3232. doi:10.1109/
     TNNLS.2018.2876865.
 [3] A. Agiollo, A. Omicini, GNN2GNN: Graph neural networks to generate neural networks, in:
     J. Cussens, K. Zhang (Eds.), Uncertainty in Artificial Intelligence, volume 180 of Proceedings of
     Machine Learning Research, ML Research Press, Maastricht, The Netherlands, 2022, pp. 32–42. URL:
     https://proceedings.mlr.press/v180/agiollo22a.html, proceedings of the Thirty-Eighth Conference
     on Uncertainty in Artificial Intelligence, UAI 2022, 1-5 August 2022, Eindhoven, The Netherlands.
 [4] J. Zhang, B. Chen, L. Zhang, X. Ke, H. Ding, Neural, symbolic and neural-symbolic reasoning on
     knowledge graphs, AI Open 2 (2021) 14–35. doi:10.1016/J.AIOPEN.2021.03.001.
 [5] A. Agiollo, A. Rafanelli, M. Magnini, G. Ciatto, A. Omicini, Symbolic knowledge injection meets
     intelligent agents: QoS metrics and experiments, Autonomous Agents and Multi-Agent Systems
     37 (2023) 27:1–27:30. doi:10.1007/s10458-023-09609-6.
 [6] Y. Xie, Z. Xu, K. S. Meel, M. S. Kankanhalli, H. Soh, Embedding symbolic knowledge into
     deep networks, in: Advances in Neural Information Processing Systems 32: Annual Confer-
     ence on Neural Information Processing Systems 2019, NeurIPS 2019, December 8-14, 2019, Van-
     couver, BC, Canada, 2019, pp. 4235–4245. URL: https://proceedings.neurips.cc/paper/2019/hash/
     7b66b4fd401a271a1c7224027ce111bc-Abstract.html.
 [7] L. d. Raedt, S. Dumančić, R. Manhaeve, G. Marra, From statistical relational to neuro-symbolic
     artificial intelligence, in: C. Bessiere (Ed.), Proceedings of the Twenty-Ninth International Joint
     Conference on Artificial Intelligence, IJCAI-20, 2020, pp. 4943–4950. doi:10.24963/ijcai.2020/
     688.
 [8] G. Ciatto, F. Sabbatini, A. Agiollo, M. Magnini, A. Omicini, Symbolic knowledge extraction and
     injection with sub-symbolic predictors: A systematic literature review, ACM Computing Surveys
     56 (2024) 161:1–161:35. doi:10.1145/3645103.
 [9] D. Hendrycks, T. G. Dietterich, Benchmarking neural network robustness to common corruptions
     and perturbations, in: 7th International Conference on Learning Representations, ICLR 2019, New
     Orleans, LA, USA, May 6-9, 2019, OpenReview.net, 2019. URL: https://openreview.net/forum?id=
     HJz6tiCqYm.
[10] A. Agiollo, A. Omicini, Measuring trustworthiness in neuro-symbolic integration, in: M. Ganzha,
     L. Maciaszek, M. Paprzycki, D. Ślęzak (Eds.), Proceedings of the 18th Conference on Computer
     Science and Intelligence Systems, volume 35 of Annals of Computer Sciences and Information
     Systems, 2023, pp. 1–10. doi:10.15439/2023F6019.
[11] M. Magnini, G. Ciatto, A. Omicini, On the design of PSyKI: a platform for symbolic knowledge in-
     jection into sub-symbolic predictors, in: D. Calvaresi, A. Najjar, M. Winikoff, K. Främling (Eds.), Ex-
     plainable and Transparent AI and Multi-Agent Systems, volume 13283 of Lecture Notes in Computer
     Science, Springer, Cham, Switzerland, 2022, pp. 90–108. doi:10.1007/978-3-031-15565-9_6,
     4th International Workshop, EXTRAAMAS 2022, Virtual Event, May 9–10, 2022, Revised Selected
     Papers.
[12] X. Liu, M. Cheng, H. Zhang, C. Hsieh, Towards robust neural networks via random self-
     ensemble, in: V. Ferrari, M. Hebert, C. Sminchisescu, Y. Weiss (Eds.), Computer Vision -
     ECCV 2018 - 15th European Conference, Munich, Germany, September 8-14, 2018, Proceed-
     ings, Part VII, volume 11211 of Lecture Notes in Computer Science, Springer, 2018, pp. 381–397.
     doi:10.1007/978-3-030-01234-2_23.
[13] G. Montavon, G. B. Orr, K. Müller (Eds.), Neural Networks: Tricks of the Trade - Second
     Edition, volume 7700 of Lecture Notes in Computer Science, Springer, 2012. doi:10.1007/
     978-3-642-35289-8.
[14] C. Shorten, T. M. Khoshgoftaar, A survey on image data augmentation for deep learning, Journal
     of Big Data 6 (2019) 60:1–60:48. doi:10.1186/s40537-019-0197-0.
[15] A. Shrivastava, T. Pfister, O. Tuzel, J. Susskind, W. Wang, R. Webb, Learning from simulated and
     unsupervised images through adversarial training, in: 2017 IEEE Conference on Computer Vision
     and Pattern Recognition, CVPR 2017, Honolulu, HI, USA, July 21-26, 2017, IEEE Computer Society,
     2017, pp. 2242–2251. doi:10.1109/CVPR.2017.241.
[16] J. Franceschi, A. Fawzi, O. Fawzi, Robustness of classifiers to uniform 𝑙𝑝 and Gaussian noise,
     in: A. Storkey, F. Perez-Cruz (Eds.), International Conference on Artificial Intelligence and Stat-
     istics (AISTATS 2018), volume 84 of Proceedings of Machine Learning Research, ML Research
     Press, Playa Blanca, Lanzarote, Spain, 2018, pp. 1280–1288. URL: http://proceedings.mlr.press/v84/
     franceschi18a/franceschi18a.pdf.
[17] C. Szegedy, W. Zaremba, I. Sutskever, J. Bruna, D. Erhan, I. J. Goodfellow, R. Fergus, Intriguing
     properties of neural networks, in: Y. Bengio, Y. LeCun (Eds.), 2nd International Conference on
     Learning Representations (ICLR 2014), Conference Track Proceedings, Banff, AB, Canada, 2014.
     URL: https://openreview.net/forum?id=kklr_MTHMRQjG.
[18] European Commission, Ethics guidelines for trustworthy AI, https://ec.europa.eu/digital-single-
     market/en/news/ethics-guidelines-trustworthy-ai, 2019.
[19] T. R. Besold, A. S. d’Avila Garcez, S. Bader, H. Bowman, P. Domingos, P. Hitzler, K. Kühnberger, L. C.
     Lamb, P. M. V. Lima, L. de Penning, G. Pinkas, H. Poon, G. Zaverucha, Neural-symbolic learning
     and reasoning: A survey and interpretation, in: P. Hitzler, M. K. Sarker (Eds.), Neuro-Symbolic
     Artificial Intelligence: The State of the Art, volume 342 of Frontiers in Artificial Intelligence and
     Applications, IOS Press, 2021, pp. 1–51. doi:10.3233/FAIA210348.
[20] S. Verma, J. P. Dickerson, K. Hines, Counterfactual explanations for machine learning: A review,
     CoRR abs/2010.10596 (2020). URL: https://arxiv.org/abs/2010.10596. arXiv:2010.10596.
[21] J. M. Joyce, Kullback-Leibler divergence, in: M. Lovric (Ed.), International Encyclopedia of
     Statistical Science, Springer Berlin Heidelberg, Berlin, Heidelberg, 2011, pp. 720–722. doi:10.
     1007/978-3-642-04898-2_327.
[22] I. Goodfellow, Y. Bengio, A. Courville, Deep Learning, MIT Press, 2016. http://www.
     deeplearningbook.org.
[23] J. Xu, Z. Zhang, T. Friedman, Y. Liang, G. V. den Broeck, A semantic loss function for deep learning
     with symbolic knowledge, in: J. G. Dy, A. Krause (Eds.), Proceedings of the 35th International
     Conference on Machine Learning, ICML 2018, Stockholmsmässan, Stockholm, Sweden, July 10-15,
     2018, volume 80 of Proceedings of Machine Learning Research, PMLR, 2018, pp. 5498–5507. URL:
     http://proceedings.mlr.press/v80/xu18h.html.
[24] A. Yazdani, L. Lu, M. Raissi, G. E. Karniadakis, Systems biology informed deep learning for
     inferring parameters and hidden dynamics, PLoS Computational Biology 16 (2020). doi:10.1371/
     journal.pcbi.1007575.
[25] P. Kairouz, Z. Liu, T. Steinke, The distributed discrete gaussian mechanism for federated learning
     with secure aggregation, in: M. Meila, T. Zhang (Eds.), Proceedings of the 38th International
     Conference on Machine Learning, ICML 2021, 18-24 July 2021, Virtual Event, volume 139 of
     Proceedings of Machine Learning Research, PMLR, 2021, pp. 5201–5212. URL: http://proceedings.
     mlr.press/v139/kairouz21a.html.
[26] P. Körner, M. Beuschel, J. Barbosa, V. S. Costa, V. Dahl, M. V. Hermenegildo, J. F. Morales, J. Wiele-
     maker, D. Diaz, S. Abreu, G. Ciatto, Fifty years of Prolog and beyond, Theory and Practice of
     Logic Programming 22 (2022) 776–858. doi:10.1017/S1471068422000102.
[27] G. G. Towell, J. W. Shavlik, M. O. Noordewier, Refinement of approximate domain theories by
     knowledge-based neural networks, Proceedings of the 8th National Conference on Artificial
     Intelligence (1990) 861–866. URL: http://www.aaai.org/Library/AAAI/1990/aaai90-129.php.
[28] M. Magnini, G. Ciatto, A. Omicini, KINS: Knowledge injection via network structuring, in:
     R. Calegari, G. Ciatto, A. Omicini (Eds.), CILC 2022 – Italian Conference on Computational Logic,
     volume 3204 of CEUR Workshop Proceedings, 2022, pp. 254–267. URL: http://ceur-ws.org/Vol-3204/
     paper_25.pdf.
[29] M. Magnini, G. Ciatto, A. Omicini, A view to a KILL: Knowledge injection via lambda layer, in:
     A. Ferrando, V. Mascardi (Eds.), WOA 2022 – 23rd Workshop “From Objects to Agents”, volume
     3261 of CEUR Workshop Proceedings, Sun SITE Central Europe, RWTH Aachen University, 2022,
     pp. 61–76. URL: http://ceur-ws.org/Vol-3261/paper5.pdf.
[30] L. von Rüden, J. Garcke, C. Bauckhage, How does knowledge injection help in informed machine
     learning?, in: International Joint Conference on Neural Networks, IJCNN 2023, Gold Coast,
     Australia, June 18-23, 2023, IEEE, 2023, pp. 1–8. doi:10.1109/IJCNN54540.2023.10191994.
[31] P. E. McKnight, J. Najab, Mann-Whitney U Test, John Wiley and Sons, Ltd, 2010, pp. 1–1. doi:10.
     1002/9780470479216.corpsy0524.