<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Safety Properties of Inductive Logic Programming</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Gavin Leech</string-name>
          <email>g.leech@bristol.ac.uk</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Nandi Schoots</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Joar Skalse</string-name>
          <xref ref-type="aff" rid="aff3">3</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Equal contribution</institution>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>King's College London and Imperial College London</institution>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>University of Bristol</institution>
        </aff>
        <aff id="aff3">
          <label>3</label>
          <institution>University of Oxford</institution>
        </aff>
      </contrib-group>
      <abstract>
        <p>This paper investigates the safety properties of inductive logic programming (ILP), particularly as compared to deep learning systems. We consider the following properties: ease of model specification; robustness to input change; control over inductive bias; verification of specifications; post-hoc model editing; and interpretability. We find that ILP could satisfy many of these properties in its home domains. Lastly, we propose a hybrid system using ILP as a preprocessor to generate specifications for other ML systems.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Introduction</title>
      <p>
        Symbolic approaches to AI are sometimes considered safer
than neural approaches
        <xref ref-type="bibr" rid="ref1 ref7">(Condry 2016; Anderson et al.
2020)</xref>
        . We investigate this by analysing how one symbolic
approach, inductive logic programming (ILP), fares on
specific safety properties.
      </p>
      <p>
        ILP is a declarative subfield of ML for learning from
examples and encoded “background knowledge” (predicates
and constraints), using logic programs to represent both
these inputs and the output model
        <xref ref-type="bibr" rid="ref34">(Muggleton 1999)</xref>
        . (We
use ‘output model’ and ‘ILP hypothesis’ interchangeably.)
      </p>
      <p>
        We find ILP to have potential to satisfy an array of safety
properties. To arrive at this, we survey existing work in ILP
and deep learning in light of the safety properties defined
in the framework of
        <xref ref-type="bibr" rid="ref40">Ortega and Maini (2018)</xref>
        . We also
formalise robustness to input change and model editing. We
suggest a hybrid system, in which ILP is used as a
preprocessing step to generate specifications for other ML systems.
      </p>
      <p>
        To our knowledge, this is the first analysis of ILP’s safety
potential, and of ILP’s differences from deep learning.
Related work includes Cropper, Dumancˇic´, and Muggleton
(2020)’s recent survey of ILP, the interpretability work of
Muggleton et al. (2018b), and
        <xref ref-type="bibr" rid="ref47">Powell and The´venod-Fosse
(2002)</xref>
        ’s study of rule-based safety-critical systems.
      </p>
      <p>
        Consider a machine learning system ‘safe’ when the
system’s goals are specified correctly, when it acts robustly
according to those goals, when we are assured about these two
properties
        <xref ref-type="bibr" rid="ref14 ref17 ref2 ref29 ref40 ref59">(Ortega and Maini 2018)</xref>
        , such that the risk of
harm from deploying the system is greatly reduced. ILP may
be a natural fit for the assurance side of safety: often, not just
the output model, but also the learning process takes place
at a relatively high level (that is, at the level of symbolic
inference). Similarly, ILP plausibly satisfies multiple
important specification and robustness properties. We assess ILP
on: Specification properties (ease of model specification and
value loading; ease of adjusting the learned model to satisfy
specifications; and control over inductive bias); Robustness
properties (robustness to input change and to post-training
model edits); Assurance properties (interpretability and
explainability; verification of specifications; and control over
the inductive bias).
      </p>
      <p>Many safety properties await formalisation, preventing
quantitative comparisons. Where a formal metric is lacking,
we qualitatively compare ILP to deep learning (DL).</p>
      <p>
        In the following we refer to ‘ILP’ as if it was
monolithic, but ILP systems differ widely in search strategy,
exactness, completeness, target logic (e.g. Prolog, Datalog,
ASP), noise-handling, ability to invent predicates, and the
order of the output theory
        <xref ref-type="bibr" rid="ref5">(Boytcheva 2002)</xref>
        . This diversity
limits the general statements we can make, but some remain.
      </p>
    </sec>
    <sec id="sec-2">
      <title>Safety properties of ILP</title>
      <sec id="sec-2-1">
        <title>Model Specification</title>
        <p>The specification of an ML system serves to define its
purpose. When this purpose is successfully implemented in hard
constraints, we may obtain guarantees about the system’s
behaviour. A defining feature of ILP systems is user-specified
background knowledge. This provides a natural way to
impose specifications on ILP systems. An ILP problem
specification is a set of positive examples, negative examples, and
background knowledge.</p>
        <p>Consider two important properties of classical ILP. Given
a background B, an output model M , and positive examples
E+, the model M is
weakly consistent if: B ^ M 6j= False; and
strongly consistent if: B ^ M ^ E+ 6j= False.</p>
        <p>
          Weak consistency forbids the generation of models that
contradict any clause in B
          <xref ref-type="bibr" rid="ref34">(Muggleton 1999)</xref>
          . In general,
ILP algorithms must satisfy weak consistency
          <xref ref-type="bibr" rid="ref34">(Muggleton
1999)</xref>
          , though probabilistic systems allow its violation; see
below. Hence, to guarantee that the learned model M
satisfies some specification s, all we need to do is encode s in
first-order logic (FOL) and add it to the background B.
However, there are still some specification challenges for ILP.
        </p>
        <p>
          Not all systems respect strong consistency. Many
modern implementations of ILP are designed to handle noise in
the example set
          <xref ref-type="bibr" rid="ref14 ref36 ref38 ref40 ref56">(Srinivasan 2006; Muggleton et al. 2018a)</xref>
          .
For specifications encoded in the example set, noise
handling means that the system is only nudged in the direction
of the specification. Furthermore, probabilistic ILP systems
can specify the background as probabilistic facts
          <xref ref-type="bibr" rid="ref15">(De Raedt
et al. 2015)</xref>
          . This means that even weak consistency can be
violated. As such, these systems may not offer specification
guarantees.
        </p>
        <p>Incompleteness. Even though a model satisfying our
specification exists, an incomplete ILP algorithm might not find
it. Some leading implementations of ILP are incomplete, i.e.
a solution may exist even though the system does not find
one (Cropper and Tourret 2018)</p>
        <p>
          Specifications may be hard to encode as FOL formulae. In
computer vision, a long tradition of manually encoding
visual concepts was rapidly outperformed by learned
representations
          <xref ref-type="bibr" rid="ref20">(Goodfellow, Bengio, and Courville 2016)</xref>
          : it proved
possible to learn these improved concepts, but intractable to
hand-code them. Insofar as ILP backgrounds must at present
be manually encoded (as opposed to learned via predicate
invention), we infer that some specifications are not
practically possible to impose on ILP.
        </p>
        <p>
          Human values are hard to encode as FOL formulae. A
particularly interesting kind of specification are those that
concern norms or values, i.e. specifications that aim to
ensure that the output respects ethical considerations. There is
precedent for formalizing norms and moral obligations
using logic – deontic logic is an area of philosophical logic
that aims to formalise and deduce moral claims
          <xref ref-type="bibr" rid="ref31">(McNamara
2019)</xref>
          . This has been used to partially formalise some
ethical frameworks
          <xref ref-type="bibr" rid="ref26 ref43">(Kroy 1976; Peterson 2014)</xref>
          . However,
encoding general normative requirements in formal logic is an
open problem. Further, we do not have a complete
articulation of all such requirements in any formalism. It seems
unlikely that in the near future we will obtain a complete
encoding, owing to deep inconsistencies across people and
the contextual nature of value
          <xref ref-type="bibr" rid="ref67">(Yudkowsky 2011)</xref>
          .
Furthermore, it may be impossible to learn a representation of these
preferences, in the absence of a strong model of human error
          <xref ref-type="bibr" rid="ref14 ref17 ref2 ref29 ref40 ref59">(Armstrong and Mindermann 2018)</xref>
          .
        </p>
        <p>
          Model specification in DL. Methods exist for limited
model specification in DL
          <xref ref-type="bibr" rid="ref45">(Platt and Barr 1988)</xref>
          , many of
which focus on specific domains
          <xref ref-type="bibr" rid="ref24 ref69">(Kashinath, Marcus et al.
2019; Zhang et al. 2020)</xref>
          . However, if we interpret a
specification as a hard constraint on outputs, then most current
DL methods do not allow specification. Instead they
impose soft constraints, modifying the loss to discourage
outof-specification behaviour. Imposing hard constraints in DL
amounts to imposing a linear set of constraints on the output
of the model. Soft constraints in the form of subtle
alterations to the loss function or learning algorithm are harder
to specify than e.g. a linear set of hard constraints
          <xref ref-type="bibr" rid="ref41 ref63">(Pathak,
Kra¨henbu¨hl, and Darrell 2015)</xref>
          . Soft constraints are
pervasive due to the computational expense of hard constraints in
neural networks: since networks can have millions of
adaptive parameters, it is not practical to use ordinary constrained
optimisation methods to impose them
          <xref ref-type="bibr" rid="ref30">(Ma´rquez-Neila,
Salzmann, and Fua 2017)</xref>
          .
        </p>
      </sec>
      <sec id="sec-2-2">
        <title>Robustness to Input Change</title>
        <p>Robustness concerns smooth output change: If we change
the input slightly, will the output (of the learning algorithm
or of the learned model) change only slightly? To formalize
this, we define similarity of inputs and output hypotheses.</p>
        <p>
          In DL input datasets, the problem description is usually
very correlated with the semantics of the problem. For
example, Gaussian noise usually does not affect the
semantics of the problem. DL models are often insensitive to small
changes in the description of the input. However,
adversarial changes induce large changes in output, despite the input
changes being trivial to the human eye
          <xref ref-type="bibr" rid="ref60">(Szegedy et al. 2014)</xref>
          .
        </p>
        <p>
          For Horn clauses (a typical form in ILP output
hypotheses), one distance measure is the ‘rewrite distance’ (the
minimum number of syntactic edits that transform one clause
into another)
          <xref ref-type="bibr" rid="ref10 ref16 ref18 ref42">(Edelmann and Kuncˇak 2019)</xref>
          . For our
purposes, this is inappropriate, since it neglects the semantic
distances we are targeting: a negation of the whole clause
would count as a rewrite distance of 1, despite being
maximally semantically different.
        </p>
      </sec>
      <sec id="sec-2-3">
        <title>Definition 1 (Similarity of Datasets) Given two datasets</title>
        <p>D1, D2, let H1 and H2 be the sets of hypotheses
compatible with D1 and D2 respectively. Let the weight of a set of
hypotheses H be defined as a weighed sum of the
hypotheses in H, where more complex hypotheses are given lower
weight (so that hypothesis h has weight 0:52c(h), where c(h)
is the complexity of h). We then say that D1 and D2 are
similar if H1 \ H2 has a large weight.</p>
      </sec>
      <sec id="sec-2-4">
        <title>Definition 2 (Similarity of Hypotheses) We say that two</title>
        <p>hypotheses h1 and h2 are similar if the probability that they
will agree on an instance x sampled from the underlying
data distribution is high.</p>
        <sec id="sec-2-4-1">
          <title>Definition 3 (Robustness to Input Change) Let L: D !</title>
          <p>M be a learning algorithm. We say that L is robust to input
change if it is the case that L(D1) and L(D2) are similar
whenever D1 and D2 are similar. More specifically, we say
that L has robustness parameters rD, rM if: for any D1 and
D2 such that they have similarity rD or higher, the similarity
between L(D1) and L(D2) is at least rM .</p>
          <p>We note that, for this notion of similarity between
datasets, the distance between two ILP problems may be
very large even if their descriptions are almost the same. For
example, adding a negation somewhere in the description of
D1 may completely change its distance to D2.</p>
          <p>ILP is robust to syntactic input change. ILP is largely
invariant to how the input problem is represented (in the sense
of symbol renaming or syntactic substitutions, which do not
affect the semantics). Two semantically equivalent problems
have identical sets of compatible output hypotheses.</p>
          <p>Examples of trivial syntactic changes to a problem
include: renaming atoms or predicates; substituting a ground
term for a variable; or substituting in a different variable. An
ILP problem statement is parsed as an ordered set of
logical sentences, and substitutions within these sentences do
not affect the semantics of the individual examples. Absent
complicating implementation details, they thus do not affect
the semantics of the output. Another syntactic change to a
problem is adding or removing copies of examples; these
changes do not have any effect on what hypothesis is output.</p>
          <p>
            Changing the order of examples could (depending on
the search algorithm) change the chosen output hypothesis.
Even though the set of consistent output hypotheses does not
change when the order of examples changes, the hypothesis
that comes up first in the search may change. For example,
Metagol depends on the order
            <xref ref-type="bibr" rid="ref12 ref13">(Cropper and Morel 2020)</xref>
            .
This order dependence is a property of some clause-level
search methods
            <xref ref-type="bibr" rid="ref56">(Srinivasan 2006)</xref>
            .
          </p>
          <p>Robustness to semantic input change. Naturally, semantic
changes to the problem can completely change the output
hypothesis. For example, negating a single example can
preclude finding any appropriate hypothesis.</p>
          <p>Suppose D1 and D2 are two datasets, with corresponding
hypothesis spaces respectively H1 and H2. ILP has a fixed
order (which depends on the inductive biases) of traversing
the total set of potential hypotheses for a solution. Say ILP
outputs hypothesis h1 for problem D1 and hypothesis h2 for
D2. Even if H1 6= H2, h1 and h2 may be the same. When
h1 6= h2, we would like to assess their similarity.</p>
          <p>Given an output model. If we change one input example,
then we may be able to check whether this input example
is consistent with the output model. We may not be able to
completely visualise the coverage, but may be able to predict
whether the output model will be different.</p>
          <p>Empirically assessing robustness to input change.
Potentially, sampling can inform us about the robustness to input
change of ILP and deep learning. An experiment could work
as follows: Generate ILP problems such that we
(approximately) know the distance between the datasets. Then run
ILP on each problem and store their output hypotheses. We
then select a distance measure and assess the distance
between each of the output hypotheses. This allows us to
(approximately) evaluate the robustness to input change of ILP.
A similar sampling process can be used for other learning
algorithms to compare the robustness of different algorithms.</p>
        </sec>
      </sec>
      <sec id="sec-2-5">
        <title>Control over Inductive Bias</title>
        <p>
          The inductive bias of a learning algorithm is the set of
(often implicit) assumptions used to generalise a finite input
set to a complete output model
          <xref ref-type="bibr" rid="ref21 ref33">(Mitchell 1980; Hu¨llermeier,
Fober, and Mernberger 2013)</xref>
          . If several hypotheses fit the
training data, the inductive bias of the learning algorithm
determines which is selected. Correct behaviour is generally
under-determined by the training data, so selecting a model
with the right behaviour demands inductive bias. It is thus
desirable to adapt the training algorithm through fine
control over the inductive bias.
        </p>
        <p>
          Informally, a learning algorithm has a low
Vapnik–Chervonenkis (VC) dimension if it can only express
simple models. If a learning algorithm has a low
VCdimension then it can be shown that it is likely to generalise
well with small amounts of data, regardless of its
inductive bias
          <xref ref-type="bibr" rid="ref63">(Vapnik and Chervonenkis 2015)</xref>
          . However, with
a more expressive learning algorithm (such as DL or ILP)
this is insufficient to yield good generalisation, and hence
such learning algorithms need a good inductive bias to work
well. ILP’s strong bias allows it to perform well on small
datasets, even though hypotheses can also be highly
expressive
          <xref ref-type="bibr" rid="ref62">(Tausend 1994)</xref>
          .
        </p>
        <p>The two main components of inductive bias are
Limits to the hypothesis space: Restricting the hypothesis
space, i.e. the set of possible output models; and
Guiding the search: The search order for traversing
through the hypothesis space, as well as heuristics to
assess hypotheses.</p>
        <p>
          Inductive bias in DL. The hypothesis space in DL is
largely determined by the network architecture
          <xref ref-type="bibr" rid="ref14 ref17 ref2 ref29 ref40 ref59 ref61">(Szymanski,
McCane, and Albert 2018)</xref>
          , which we have control over. For
example, convolutional neural networks hard-code the
assumption that output classes are invariant to shift
transformation
          <xref ref-type="bibr" rid="ref20">(Goodfellow, Bengio, and Courville 2016)</xref>
          .
Traintime methods like dropout and learning rate decay also
regularise networks and so add inductive bias
          <xref ref-type="bibr" rid="ref58">(Srivastava et al.
2014)</xref>
          . In addition, neural networks have a broad bias
towards simplicity, although it is unclear how this bias works
          <xref ref-type="bibr" rid="ref13 ref46 ref68">(Zhang et al. 2017; Poggio, Liao, and Banburski 2020)</xref>
          . The
lack of theoretical understanding of DL’s search bias implies
little explicit control over it.
        </p>
        <p>
          Inductive bias in ILP. We can restrict the hypothesis space
in many ways. A critical design decision for an ILP system
is which fragment of FOL represents the examples,
background and output model. The classical choice restricts FOL
to definite Horn clauses
          <xref ref-type="bibr" rid="ref37">(Muggleton and de Raedt 1994)</xref>
          .
        </p>
        <p>
          In addition, a strong ILP language bias stems from
usersupplied constraints on the structure or type of the
hypothesis, e.g. mode declarations, meta-rules, or program
templates
          <xref ref-type="bibr" rid="ref10 ref16 ref18 ref42">(Payani and Fekri 2019)</xref>
          . In some sense these are
hyperparameters, as found in any ML system. However, these
constraints can be enormously informative, e.g. specifying:
which predicates to use in the head or body of the
output model; the quantifier of each argument in the predicate;
which arguments are to be treated as input and output; and
the types of these argument
          <xref ref-type="bibr" rid="ref14 ref17 ref2 ref29 ref40 ref59">(Evans and Grefenstette 2018)</xref>
          .
        </p>
        <p>
          User-supplied constraints can pertain to
          <xref ref-type="bibr" rid="ref37">(Muggleton and
de Raedt 1994)</xref>
          among other things
        </p>
        <p>Syntax, e.g. second-order schema or bounded term depth;
Semantics (on the level of terms), e.g. hard-coding
the types of predicate arguments, or using determinate
clauses; and
Bounds on the length of the output model.</p>
        <p>
          Two elementary ways to order an ILP search over the set
of possible output models are top-down (‘from general to
specific’) or bottom-up (‘from specific to general’). At each
step of ILP learning, we need a way to score multiple
competing hypotheses. This can be done via computing the
information gain of the change to the hypothesis
          <xref ref-type="bibr" rid="ref48">(Quinlan 1990)</xref>
          or through probabilistic scoring
          <xref ref-type="bibr" rid="ref37">(Muggleton and de Raedt
1994)</xref>
          . A further source of search bias involves specifying
the order in which we prune candidate hypotheses.
        </p>
        <p>
          Comparing ILP with DL. In Table 1 we compare control
over inductive bias in ILP and DL. We consider the
following, from
          <xref ref-type="bibr" rid="ref65">Witten et al. (2017)</xref>
          : language bias (hypothesis
space restriction), search bias (how the search through the
hypothesis space is ordered), and simplicity bias (how
overfitting is prevented).
        </p>
        <p>
          When is control over inductive bias actually hand-coding
solutions? The more inductive biases are customised, the
more the learning method resembles explicit programming
of a solution class. For example, when doing reinforcement
learning it is possible to include information about how the
task should be solved in the reward function. As more
information is included, designing the reward function resembles
specifying a solution
          <xref ref-type="bibr" rid="ref14 ref17 ref2 ref29 ref40 ref59">(Sutton and Barto 2018)</xref>
          . In ILP,
taskspecific language biases are often unavoidable for
performance reasons, but they risk pruning unexpected solutions,
involve a good deal of expert human labour, and can lead to
brittle systems which may not learn the problem structure so
much as they are passed it to begin with
          <xref ref-type="bibr" rid="ref10 ref16 ref18 ref42">(Payani and Fekri
2019)</xref>
          . This problem could be mitigated by progress in
automating inductive bias choice in ILP.
        </p>
      </sec>
      <sec id="sec-2-6">
        <title>Verification of Specifications</title>
        <p>In cases where model specification does not give hard
guarantees about model behaviour, post-hoc verification is
needed. That is, determining, given program M and
specification s, whether M ’s behaviour satisfies s.</p>
        <sec id="sec-2-6-1">
          <title>Definition 4 (Specification) Let L: D ! M be a learning</title>
          <p>algorithm. A specification s is a property such that for all
models M 2 M, the model satisfies the property or not.</p>
          <p>
            The problem of verifying whether a model satisfies a
specification is NP-hard
            <xref ref-type="bibr" rid="ref6">(Clarke and Emerson 1981)</xref>
            , both
for a neural network
            <xref ref-type="bibr" rid="ref25">(Katz et al. 2017)</xref>
            and for logic
programs
            <xref ref-type="bibr" rid="ref14 ref17 ref2 ref29 ref40 ref59">(Madelaine and Martin 2018)</xref>
            .
          </p>
          <p>
            Verification in ILP. In practice, verifying properties of an
output hypothesis is often easy. Suppose you have a
propositional theory and want to verify whether this satisfies the
specification False. This is equivalent to solving
satisfiability, and so is at least NP hard. We can verify whether an
ILP model M satisfies an arbitrary Datalog specification s
by running resolution on M [ f:sg to see if it derives False.
In fact, this can be done in some cases where s is not in
Datalog. For example, this could be done as long as s is in the
Bernays-Scho¨nfinkel fragment, albeit in double-exponential
time
            <xref ref-type="bibr" rid="ref44">(Piskac, de Moura, and Bjørner 2008)</xref>
            . The proof search
can be attempted with arbitrary Prolog specifications, but
may not terminate.
          </p>
          <p>
            Verification in DL. To quote
            <xref ref-type="bibr" rid="ref25">Katz et al. (2017)</xref>
            , “Deep
neural networks are large, non-linear, and non-convex, and
verifying even simple properties about them is an NP-complete
problem”. In practice, complete solvers can verify properties
of networks with thousands of nodes, but time out for larger
networks (Wang et al. 2018). Incomplete methods can verify
properties of networks with 100 000 ReLU nodes
            <xref ref-type="bibr" rid="ref4 ref54">(Singh
et al. 2019; Botoeva et al. 2020)</xref>
            . Note that the smallest
networks that achieve decent results on CIFAR10 have 50k
nodes. Networks can be trained such that they are easier to
verify (Xiao et al. 2019).
          </p>
        </sec>
      </sec>
      <sec id="sec-2-7">
        <title>Post-hoc Model Editing</title>
        <sec id="sec-2-7-1">
          <title>Definition 5 (Model Editing) Let L: D ! M be a learn</title>
          <p>ing algorithm. Let M 2 M be a learned model. Let s be a
specification. We apply model editing to M on specification
s, if we find a model M 0 2 M that has property s without
re-applying the learning algorithm L.</p>
          <p>Let d be a distance metric on M. We say that we
successfully edit M to fit specification s with respect to distance d
if we find a model M 0 2 M that has property s and out of
all models with property s has minimal distance from M .</p>
          <p>Model Editing in ILP. The symbolic representation could
make ILP models easier to manipulate than DL models. ILP
output models are very interpretable and it is relatively easy
for humans to write logical sentences, which should make it
in some cases possible to apply model editing.</p>
          <p>The output model of ILP is a conjunction of logical
clauses. The model can easily be edited, by removing or
adding individual clauses. If we simply add clause s to M ,
then we get a new model M 0 = M [ fsg, which satisfies s
and has minimal distance to M with respect to the ‘rewrite
distance’. When adding clauses, one needs to ensure the new
model is still consistent.</p>
          <p>
            A form of post-hoc model editing has been applied to
large neural networks, though only by automating the edits.
The OpenAI Five agent was trained across several
different architectures, with an automatic process for discovering
weights to copy
            <xref ref-type="bibr" rid="ref10 ref16 ref18 ref42 ref49">(Raiman, Zhang, and Dennison 2019)</xref>
            .
          </p>
          <p>
            Model Editing in DL. After training a large neural
network, we (practically speaking) obtain a black-box model.
This black-box is not easy to manipulate, owing to the
number of parameters and the distributed nature of its
representation. Through active learning or incremental learning
we can update the model - we could add a module that
deals with exceptions, or fine-tune on extra training data for
low-performing subgroups. However, these do not give us
much control over exactly how the black-box changes
            <xref ref-type="bibr" rid="ref52">(Settles 2009)</xref>
            .
          </p>
          <p>Because the black-box is difficult to interpret, we do not
fully comprehend what function the network has learned and
so are not able to enhance it. Researchers can override the
output with a different learned module, but there is no
lowlevel interactive combination of model and human insight.</p>
        </sec>
      </sec>
      <sec id="sec-2-8">
        <title>Transparency</title>
        <p>We consider the transparency of learned DL and ILP models,
and the transparency of the learning algorithms.</p>
        <p>
          Transparency of the learned model. In contrast to DL
models, ILP outputs are represented in an explicit high-level
language, making them more transparent. We distinguish
between
          <xref ref-type="bibr" rid="ref27">(Lipton 2018)</xref>
          : a globally transparent or ‘simulatable’
model; and a locally transparent or ‘decomposable’ model.
        </p>
        <p>
          Decomposability. A decomposable model is one in which
each part of the model - each input, parameter, and
computation - admits an intuitive explanation, independent of other
instances of the model part
          <xref ref-type="bibr" rid="ref27">(Lipton 2018)</xref>
          . The many
parameters of a neural network form a distributed representation
of a nonlinear function, and as such it is unhelpful to reason
about individual parameters.
        </p>
        <p>An ILP output model is a conjunction of predicates and
literals. When the background is human-specified, each
individual predicate will admit an intuitive explanation. When
predicates are invented by the ILP system, the results can be
counter-intuitive or long; however, they are still themselves
encoded as decomposable clauses of intuitive features.</p>
        <p>Simulatability. A user can simulate a model if they can
take input and work out what the model would output. More
precisely, a model M is simulatable in proportion to the
mean accuracy with which, after brief study of M , a
population of users can reproduce the output of M on new data.</p>
        <p>
          A small usability study (n=16) found that access to an
ILP program did allow users to simulate a concept they were
unable to infer themselves
          <xref ref-type="bibr" rid="ref29 ref36 ref38 ref61">(Muggleton et al. 2018b)</xref>
          . It also
found, as expected, that increasing the complexity reduced
simulatability.
        </p>
        <p>
          Explainability of the learned model. Since ILP models are
relatively transparent, explanations are redundant (except in
very large programs). DL explainability is a highly active
field of research
          <xref ref-type="bibr" rid="ref19">(Gilpin et al. 2018)</xref>
          and has produced many
post-hoc tools, making use of: visualization, text
explanations, feature relevance, explanations by example, model
simplification and local explanations (Arrieta et al. 2020).
        </p>
        <p>Transparency of the learning algorithm. In DL the
learning algorithm optimises weights in a model that already has
the same architecture as the output model, such that during
training we see many intermediate models. This implies that
many of the transparency properties of DL are relevant for
its accessibility as well. On the other hand, in ILP we have a
distinct training algorithm and output model.</p>
        <p>Individual model updates during learning. In DL,
backpropagation attributes changes in the loss to individual
weights. However, backpropagation can lead to local
minima, and so sometimes weights are updated in a direction
opposite from the ideal direction. This, along with their
(humanly) incomprehensible representation imply that
individual updates are not interpretable.</p>
        <p>This contrasts with ILP, where each step of the learning
algorithm occurs on a symbolic level (for instance,
generalising a candidate hypothesis through dropping one literal). In
principle a human user could step through ILP learning and
understand the concept represented at each step, the
complete effect of each change on the model coverage, and the
particular data points that constrain the change (though in
practice learning can involve many thousands of steps, and
so this can take an impractically long time).</p>
        <p>Attributing the solution to individual training inputs.
Given an ILP output model and an input example, a human
can usually assess whether they are consistent. So in ILP it
is relatively clear which example or background predicate is
causing the ILP algorithm to output a given model.</p>
        <p>In DL however, it would be very difficult to (for instance)
assess whether an image was a member of the training set of
a given model. That is, it is difficult to attribute aspects of
the output model to individual inputs.</p>
        <p>Inductive bias towards interpretability. User-supplied
program constraints and bounds on program length mean
that we only generate programs of a certain form, which can
be interpretable by construction. Moreover, control over
inductive bias itself can be seen as a form of accessibility.</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Discussion</title>
      <p>We have argued that ILP has a number of safety properties
that make it attractive compared to DL:
1. ILP is convenient for specification, insofar as it is intuitive
to encode examples and properties of correct behaviour;
2. ILP is robust to most syntactic changes in inputs;
3. The program templates, and bounds on program length
give control over the inductive bias in ILP;
4. We can verify whether an ILP model satisfies an arbitrary</p>
      <p>Datalog specification by running resolution;
5. We can edit an ILP model by adding or removing clauses;
6. ILP models are interpretable as they are quite transparent
and are reasonably accessible.</p>
      <p>
        Competitiveness of ILP. It is unlikely that the AI
community will adopt ILP if its performance is not competitive.
Consider chemistry applications
        <xref ref-type="bibr" rid="ref57">(Srinivasan et al. 1997)</xref>
        :
ILP continues to be applied
        <xref ref-type="bibr" rid="ref22">(Kaalia et al. 2016)</xref>
        , but DL
efforts are now more extensive
        <xref ref-type="bibr" rid="ref10 ref16 ref18 ref42">(Cova and Pais 2019)</xref>
        . One
benchmark is suggestive: ILP found success in the early
years of the Comparative Assessment of protein Structure
Prediction
        <xref ref-type="bibr" rid="ref9">(Cootes, Muggleton, and Sternberg 2003)</xref>
        , but
most submissions now use DL
        <xref ref-type="bibr" rid="ref51">(Senior et al. 2019)</xref>
        . These
results may not be indicative of ILP’s current potential as
far less research is being invested in ILP than in DL. As a
suggestive bound on the ratio of investment, compare the
130 researchers (worldwide) listed on the ILP community
hub
        <xref ref-type="bibr" rid="ref53">(Siebers 2019)</xref>
        , to the 420 researchers at a single DL
lab, Berkeley AI Research, or to the 13,000 attendees of a
single DL conference, NeurIPS. This relative neglect might
allow for performance gains from research into ILP and
hybrid ILP-DL systems.
      </p>
      <p>
        A deeper concern is the limited domains in which ILP
offers its benefits. ILP generates logic programs, where DL
approximates continuous functions. We have argued that logic
programs are more human interpretable, especially when the
predicates used in the program represent concepts we know
and use. Our discussion of ILP’s transparency only applies
to domains where the data is already available on a symbolic
level. Moreover, a major theme in recent AI, cognitive
science, and linguistics is that rule approaches are insufficient
to express or learn most human-level concepts, where
continuous features and similarity to exemplars appear
necessary
        <xref ref-type="bibr" rid="ref39 ref50 ref55">(Rouder and Ratcliff 2006; Spivey 2008; Norvig 2012)</xref>
        .
      </p>
      <p>
        Both of the above suggest a need to unify
connectionist and symbolic methods. Recent attempts implement
relational or logical reasoning on neural networks
        <xref ref-type="bibr" rid="ref10 ref14 ref16 ref17 ref18 ref2 ref29 ref40 ref42 ref59">(Garnelo
and Shanahan 2019; Evans and Grefenstette 2018)</xref>
        .
However, from a safety perspective, these unifications lose
desirable properties. We hope that future versions not only
increase in performance, but also retain their safety potential.
      </p>
      <p>ILP as specification module. The above suggests a
fruitful role for ILP: as a specification generator in a mixed AI
system. We may not be able to directly specify safety
properties, but may be able to give positive and negative
examples of safe behaviour. If it is natural to formulate these
examples in natural language or logic, then ILP can generate
hypotheses based on these partial specifications. Since ILP
output models are easy to interpret, we may be able to verify
whether they meet our preferences, and perhaps edit them to
account for noisy or missing features. In certain cases (e.g.
Datalog), it may be possible to formally verify the
hypothesis’ correctness. The specification can then be transferred
losslessly to any other learning system that can handle
logical expressions (e.g. graph neural networks).</p>
      <p>ILP’s differences with DL suggest solutions to DL’s safety
shortcomings. We are hopeful that hybrid systems can
provide safety guarantees.</p>
      <p>Acknowledgements Supported by UKRI studentships
EP/S022937/1 and EP/S023356/1 and the AI Safety
Research Programme. Thanks to Alex Jackson for suggesting
ILP has potential as a specification generator.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          <string-name>
            <surname>Anderson</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Verma</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Dillig</surname>
            ,
            <given-names>I.;</given-names>
          </string-name>
          and
          <string-name>
            <surname>Chaudhuri</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          <string-name>
            <surname>Armstrong</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ; and
          <string-name>
            <surname>Mindermann</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <year>2018</year>
          .
          <article-title>Occam's razor is insufficient to infer the preferences of irrational agents</article-title>
          .
          <source>In NeurIPS.</source>
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          2020.
          <article-title>Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI</article-title>
          .
          <source>Inf. Fusion</source>
          <volume>58</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          <string-name>
            <surname>Botoeva</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Kouvaros</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Kronqvist</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Lomuscio</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ; and Misener,
          <string-name>
            <surname>R.</surname>
          </string-name>
          <year>2020</year>
          .
          <article-title>Efficient Verification of ReLU-based Neural Networks via Dependency Analysis</article-title>
          .
          <source>In AAAI.</source>
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          <string-name>
            <surname>Boytcheva</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <year>2002</year>
          .
          <article-title>Overview of ILP Systems</article-title>
          . In Cybernetics and Information Technologies.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          <string-name>
            <surname>Clarke</surname>
            ,
            <given-names>E. M.</given-names>
          </string-name>
          ; and
          <string-name>
            <surname>Emerson</surname>
            ,
            <given-names>E. A.</given-names>
          </string-name>
          <year>1981</year>
          .
          <article-title>The design and synthesis of synchronization skeletons using temporal logic.</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          <string-name>
            <surname>Condry</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          <year>2016</year>
          .
          <article-title>Meaningful Models: Utilizing Conceptual Structure to Improve Machine Learning Interpretability</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          <source>arXiv:1607</source>
          .
          <fpage>00279</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          <string-name>
            <surname>Cootes</surname>
            ,
            <given-names>A. P.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Muggleton</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ; and Sternberg,
          <string-name>
            <surname>M. J.</surname>
          </string-name>
          <year>2003</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          <string-name>
            <surname>Cova</surname>
            ,
            <given-names>T. F.</given-names>
          </string-name>
          ; and
          <string-name>
            <surname>Pais</surname>
            ,
            <given-names>A. A.</given-names>
          </string-name>
          <year>2019</year>
          .
          <article-title>Deep Learning for Deep Chemistry: Optimizing the Prediction of Chemical Patterns</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          <source>Frontiers in Chemistry 7.</source>
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          <string-name>
            <surname>Cropper</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ; Dumancˇic´, S.; and
          <string-name>
            <surname>Muggleton</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <year>2020</year>
          .
          <article-title>Turning 30: New Ideas in Inductive Logic Programming</article-title>
          .
          <source>In IJCAI.</source>
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          <string-name>
            <surname>Cropper</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ; and Morel,
          <string-name>
            <surname>R.</surname>
          </string-name>
          <year>2020</year>
          .
          <article-title>Learning programs by learning from failures</article-title>
          .
          <source>arXiv</source>
          <year>2005</year>
          .
          <volume>02259</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          <string-name>
            <surname>Cropper</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ; and
          <string-name>
            <surname>Tourret</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <year>2018</year>
          .
          <article-title>Derivation reduction of metarules in meta-interpretive learning</article-title>
          .
          <source>In International Conference on Inductive Logic Programming.</source>
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          <string-name>
            <surname>De Raedt</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Dries</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Thon</surname>
            , I.; Van den Broeck, G.; and Verbeke,
            <given-names>M.</given-names>
          </string-name>
          <year>2015</year>
          .
          <article-title>Inducing probabilistic relational rules from probabilistic examples</article-title>
          .
          <source>In IJCAI.</source>
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          <string-name>
            <surname>Edelmann</surname>
            , R.; and Kuncˇak,
            <given-names>V.</given-names>
          </string-name>
          <year>2019</year>
          .
          <article-title>Neural-Network Guided Expression Transformation</article-title>
          . arXiv:
          <year>1902</year>
          .02194 .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          <string-name>
            <surname>Evans</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ; and
          <string-name>
            <surname>Grefenstette</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          <year>2018</year>
          .
          <article-title>Learning Explanatory Rules from Noisy Data</article-title>
          .
          <source>In JAIR.</source>
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          <string-name>
            <surname>Garnelo</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ; and Shanahan,
          <string-name>
            <surname>M.</surname>
          </string-name>
          <year>2019</year>
          .
          <article-title>Reconciling deep learning with symbolic artificial intelligence: representing objects and relations</article-title>
          .
          <source>In Current Opinion in Behavioral Sciences.</source>
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          <string-name>
            <surname>Gilpin</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Bau</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Yuan</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Bajwa</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Specter</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ; and
          <string-name>
            <surname>Kagal</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <year>2018</year>
          .
          <article-title>Explaining Explanations: An Overview of Interpretability of Machine Learning</article-title>
          .
          <source>In IEEE DSAA.</source>
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          <string-name>
            <surname>Goodfellow</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ; Bengio,
          <string-name>
            <given-names>Y.</given-names>
            ; and
            <surname>Courville</surname>
          </string-name>
          ,
          <string-name>
            <surname>A.</surname>
          </string-name>
          <year>2016</year>
          .
          <article-title>Deep Learning</article-title>
          . MIT Press. http://www.deeplearningbook.org.
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          <string-name>
            <surname>Hu</surname>
          </string-name>
          ¨llermeier, E.;
          <string-name>
            <surname>Fober</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ; and Mernberger,
          <string-name>
            <surname>M.</surname>
          </string-name>
          <year>2013</year>
          . Inductive Bias,
          <source>Encyclopedia of Systems Biology.</source>
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          <string-name>
            <surname>Kaalia</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Srinivasan</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Kumar</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ; and
          <string-name>
            <surname>Ghosh</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          <article-title>ILP-assisted de novo drug design</article-title>
          .
          <source>Machine Learning</source>
          <volume>103</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          <string-name>
            <surname>Kashinath</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Marcus</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ; et al.
          <year>2019</year>
          .
          <article-title>Enforcing Physical Constraints in Neural Neural Networks through Differentiable PDE Layer</article-title>
          .
          <source>In ICLR DeepDiffEq workshop.</source>
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          <string-name>
            <surname>Katz</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ; Barrett,
          <string-name>
            <given-names>C.</given-names>
            ;
            <surname>Dill</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            ;
            <surname>Julian</surname>
          </string-name>
          ,
          <string-name>
            <surname>K.</surname>
          </string-name>
          ; and Kochenderfer,
          <string-name>
            <surname>M.</surname>
          </string-name>
          <year>2017</year>
          .
          <article-title>Reluplex: An Efficient SMT Solver for Verifying Deep Neural Networks</article-title>
          .
          <source>Computer Aided Verification. Lecture Notes in Computer Science</source>
          <volume>10426</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          <string-name>
            <surname>Kroy</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <year>1976</year>
          .
          <article-title>A Partial Formalization of Kant's Categorical Imperative</article-title>
          .
          <article-title>An Application of Deontic Logic to Classical Moral Philosophy</article-title>
          .
          <source>Kant-Studien</source>
          <volume>67</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          <string-name>
            <surname>Lipton</surname>
            ,
            <given-names>Z. C.</given-names>
          </string-name>
          <year>2018</year>
          .
          <article-title>The mythos of model interpretability</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          <string-name>
            <surname>Queue</surname>
          </string-name>
          <year>16</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          <string-name>
            <surname>Madelaine</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ; and
          <string-name>
            <surname>Martin</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <year>2018</year>
          .
          <article-title>On the complexity of the model checking problem</article-title>
          .
          <source>SIAM J. Comput. 47.</source>
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          <article-title>Ma´rquez-</article-title>
          <string-name>
            <surname>Neila</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Salzmann</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ; and Fua,
          <string-name>
            <surname>P.</surname>
          </string-name>
          <year>2017</year>
          .
          <article-title>Imposing hard constraints on deep networks: Promises and limitations</article-title>
          .
          <source>arXiv:1706</source>
          .
          <year>02025</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref31">
        <mixed-citation>
          <string-name>
            <surname>McNamara</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <year>2019</year>
          . Deontic Logic. https://plato.stanford.
        </mixed-citation>
      </ref>
      <ref id="ref32">
        <mixed-citation>edu/archives/sum2019/entries/logic-deontic/.</mixed-citation>
      </ref>
      <ref id="ref33">
        <mixed-citation>
          <string-name>
            <surname>Mitchell</surname>
            ,
            <given-names>T. M.</given-names>
          </string-name>
          <year>1980</year>
          .
          <article-title>The need for biases in learning generalizations</article-title>
          . Department of Computer Science, Laboratory for Computer Science Research.
        </mixed-citation>
      </ref>
      <ref id="ref34">
        <mixed-citation>
          <string-name>
            <surname>Muggleton</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <year>1999</year>
          .
          <article-title>Inductive Logic Programming: Issues, results and the challenge of Learning Language in Logic.</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref35">
        <mixed-citation>
          <source>Artificial Intelligence</source>
          <volume>114</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref36">
        <mixed-citation>
          <string-name>
            <surname>Muggleton</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Dai</surname>
          </string-name>
          , W.-Z.;
          <string-name>
            <surname>Sammut</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>TamaddoniNezhad</surname>
          </string-name>
          , A.;
          <string-name>
            <surname>Wen</surname>
          </string-name>
          , J.; and
          <string-name>
            <surname>Zhou</surname>
            ,
            <given-names>Z.-H.</given-names>
          </string-name>
          <year>2018a</year>
          .
          <article-title>Metainterpretive learning from noisy images</article-title>
          .
          <source>Machine Learning</source>
          <volume>107</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref37">
        <mixed-citation>
          <string-name>
            <surname>Muggleton</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ; and
          <string-name>
            <surname>de Raedt</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <year>1994</year>
          .
          <article-title>Inductive Logic Programming: Theory and methods</article-title>
          .
          <source>The Journal of Logic Programming</source>
          <volume>19</volume>
          /
          <fpage>20</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref38">
        <mixed-citation>
          <string-name>
            <surname>Muggleton</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ; Schmid,
          <string-name>
            <given-names>U.</given-names>
            ;
            <surname>Zeller</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            ;
            <surname>Tamaddoni-Nezhad</surname>
          </string-name>
          ,
          <string-name>
            <surname>A.</surname>
          </string-name>
          ; and Besold,
          <string-name>
            <surname>T.</surname>
          </string-name>
          <year>2018b</year>
          .
          <article-title>Ultra-Strong Machine Learning: comprehensibility of programs learned with ILP</article-title>
          .
          <source>Machine Learning</source>
          <volume>107</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref39">
        <mixed-citation>
          <string-name>
            <surname>Norvig</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <year>2012</year>
          .
          <article-title>Chomsky and the two cultures of statistical learning</article-title>
          .
          <source>Significance 9.</source>
        </mixed-citation>
      </ref>
      <ref id="ref40">
        <mixed-citation>
          <string-name>
            <surname>Ortega</surname>
            ,
            <given-names>P. A.</given-names>
          </string-name>
          ; and
          <string-name>
            <surname>Maini</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          <year>2018</year>
          .
          <article-title>Building safe artificial intelligence: specification, robustness, and assurance</article-title>
          . https://medium.com/@deepmindsafetyresearch/ building-safe
          <source>-artificial-intelligence-52f5f75058f1.</source>
        </mixed-citation>
      </ref>
      <ref id="ref41">
        <mixed-citation>
          <string-name>
            <surname>Pathak</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ; Kra¨henbu¨hl, P.; and Darrell,
          <string-name>
            <surname>T.</surname>
          </string-name>
          <year>2015</year>
          .
          <article-title>Constrained Convolutional Neural Networks for Weakly Supervised Segmentation</article-title>
          . In IEEE ICCV.
        </mixed-citation>
      </ref>
      <ref id="ref42">
        <mixed-citation>
          <string-name>
            <surname>Payani</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ; and
          <string-name>
            <surname>Fekri</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <year>2019</year>
          .
          <article-title>Inductive Logic Programming via Differentiable Deep Neural Logic Networks</article-title>
          . CoRR abs/
          <year>1906</year>
          .03523.
        </mixed-citation>
      </ref>
      <ref id="ref43">
        <mixed-citation>
          <string-name>
            <surname>Peterson</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <year>2014</year>
          .
          <article-title>The categorical imperative: Category theory as a foundation for deontic logic</article-title>
          .
          <source>Applied Logic</source>
          <volume>12</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref44">
        <mixed-citation>
          <string-name>
            <surname>Piskac</surname>
          </string-name>
          , R.; de Moura, L.; and
          <string-name>
            <surname>Bjørner</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          <year>2008</year>
          .
          <article-title>Deciding effectively propositional logic with equality</article-title>
          .
          <source>Technical report, MSR-TR-2008-181</source>
          , Microsoft Research.
        </mixed-citation>
      </ref>
      <ref id="ref45">
        <mixed-citation>
          <string-name>
            <surname>Platt</surname>
            ,
            <given-names>J. C.</given-names>
          </string-name>
          ; and
          <string-name>
            <surname>Barr</surname>
            ,
            <given-names>A. H.</given-names>
          </string-name>
          <year>1988</year>
          .
          <article-title>Constrained differential optimization</article-title>
          .
          <source>In NeurIPS.</source>
        </mixed-citation>
      </ref>
      <ref id="ref46">
        <mixed-citation>
          <string-name>
            <surname>Poggio</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Liao</surname>
            ,
            <given-names>Q.</given-names>
          </string-name>
          ; and
          <string-name>
            <surname>Banburski</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <year>2020</year>
          .
          <article-title>Complexity control by gradient descent in deep networks</article-title>
          .
          <source>Nature Communications</source>
          <volume>11</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref47">
        <mixed-citation>
          <string-name>
            <surname>Powell</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <article-title>; and The´venod-</article-title>
          <string-name>
            <surname>Fosse</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <year>2002</year>
          .
          <article-title>Dependability issues in ai-based autonomous systems for space applications</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref48">
        <mixed-citation>
          <string-name>
            <surname>Quinlan</surname>
            ,
            <given-names>J. R.</given-names>
          </string-name>
          <year>1990</year>
          .
          <article-title>Learning Logical Definitions from Relations</article-title>
          .
          <source>Machine Learning 5.</source>
        </mixed-citation>
      </ref>
      <ref id="ref49">
        <mixed-citation>
          <string-name>
            <surname>Raiman</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ; Zhang, S.; and
          <string-name>
            <surname>Dennison</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <year>2019</year>
          .
          <article-title>Neural Network Surgery with Sets</article-title>
          .
          <source>arXiv:1607</source>
          .
          <fpage>00279</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref50">
        <mixed-citation>
          <string-name>
            <surname>Rouder</surname>
            ,
            <given-names>J. N.</given-names>
          </string-name>
          ; and Ratcliff,
          <string-name>
            <surname>R.</surname>
          </string-name>
          <year>2006</year>
          .
          <article-title>Comparing Exemplarand Rule-Based Theories of Categorization</article-title>
          .
          <source>Current Directions in Psychological Science</source>
          <volume>15</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref51">
        <mixed-citation>
          <string-name>
            <surname>Senior</surname>
            ,
            <given-names>A. W.</given-names>
          </string-name>
          ; Evans,
          <string-name>
            <surname>R.</surname>
          </string-name>
          ; Jumper,
          <string-name>
            <given-names>J.</given-names>
            ;
            <surname>Kirkpatrick</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            ;
            <surname>Sifre</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            ;
            <surname>Green</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            ;
            <surname>Qin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            ; Zˇ ´ıdek, A.;
            <surname>Nelson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. W. R.</given-names>
            ;
            <surname>Bridgland</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            ;
            <surname>Penedones</surname>
          </string-name>
          ,
          <string-name>
            <surname>H.</surname>
          </string-name>
          ; Petersen,
          <string-name>
            <given-names>S.</given-names>
            ;
            <surname>Simonyan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            ;
            <surname>Crossan</surname>
          </string-name>
          ,
          <string-name>
            <surname>S.</surname>
          </string-name>
          ; Kohli,
          <string-name>
            <given-names>P.</given-names>
            ;
            <surname>Jones</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. T.</given-names>
            ;
            <surname>Silver</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            ;
            <surname>Kavukcuoglu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            ; and
            <surname>Hassabis</surname>
          </string-name>
          ,
          <string-name>
            <surname>D.</surname>
          </string-name>
          <year>2019</year>
          .
          <article-title>Improved protein structure prediction using potentials from deep learning</article-title>
          .
          <source>Nature</source>
          <volume>577</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref52">
        <mixed-citation>
          <string-name>
            <surname>Settles</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <year>2009</year>
          .
          <article-title>Active learning literature survey</article-title>
          .
          <source>Technical report</source>
          , University of Wisconsin-Madison Department of Computer Sciences.
        </mixed-citation>
      </ref>
      <ref id="ref53">
        <mixed-citation>
          <string-name>
            <surname>Siebers</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <year>2019</year>
          .
          <article-title>inductive-programming community site</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref54">
        <mixed-citation>
          <string-name>
            <surname>Singh</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ; Gehr,
          <string-name>
            <surname>T.</surname>
          </string-name>
          ; Pu¨schel, M.; and Vechev,
          <string-name>
            <surname>M.</surname>
          </string-name>
          <year>2019</year>
          .
          <article-title>An Abstract Domain for Certifying Neural Networks</article-title>
          .
          <source>ACM Program. Lang. 3.</source>
        </mixed-citation>
      </ref>
      <ref id="ref55">
        <mixed-citation>
          <string-name>
            <surname>Spivey</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <year>2008</year>
          .
          <article-title>The continuity of mind</article-title>
          . Oxford University Press.
        </mixed-citation>
      </ref>
      <ref id="ref56">
        <mixed-citation>
          <string-name>
            <surname>Srinivasan</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <year>2006</year>
          .
          <article-title>The Aleph System</article-title>
          . https://www.cs.ox.
        </mixed-citation>
      </ref>
      <ref id="ref57">
        <mixed-citation>
          <string-name>
            <surname>Srinivasan</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>King</surname>
            ,
            <given-names>R. D.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Muggleton</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ; and Sternberg,
          <string-name>
            <surname>M. J. E.</surname>
          </string-name>
          <year>1997</year>
          .
          <article-title>Carcinogenesis Predictions Using ILP</article-title>
          .
          <source>In Inductive Logic Programming</source>
          , 7th International Workshop.
        </mixed-citation>
      </ref>
      <ref id="ref58">
        <mixed-citation>
          <string-name>
            <surname>Srivastava</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Hinton</surname>
          </string-name>
          , G.;
          <string-name>
            <surname>Krizhevsky</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Sutskever</surname>
            ,
            <given-names>I.;</given-names>
          </string-name>
          and Salakhutdinov,
          <string-name>
            <surname>R.</surname>
          </string-name>
          <year>2014</year>
          .
          <article-title>Dropout: A Simple Way to Prevent Neural Networks from Overfitting</article-title>
          .
          <source>JMLR 15.</source>
        </mixed-citation>
      </ref>
      <ref id="ref59">
        <mixed-citation>
          <string-name>
            <surname>Sutton</surname>
            ,
            <given-names>R. S.</given-names>
          </string-name>
          ; and
          <string-name>
            <surname>Barto</surname>
            ,
            <given-names>A. G.</given-names>
          </string-name>
          <year>2018</year>
          .
          <article-title>Reinforcement learning: An introduction</article-title>
          . MIT press.
        </mixed-citation>
      </ref>
      <ref id="ref60">
        <mixed-citation>
          <string-name>
            <surname>Szegedy</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Zaremba</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          ; Sutskever,
          <string-name>
            <surname>I.</surname>
          </string-name>
          ; Bruna,
          <string-name>
            <given-names>J.</given-names>
            ;
            <surname>Erhan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            ;
            <surname>Goodfellow</surname>
          </string-name>
          ,
          <string-name>
            <surname>I.;</surname>
          </string-name>
          and Fergus,
          <string-name>
            <surname>R.</surname>
          </string-name>
          <year>2014</year>
          .
          <article-title>Intriguing properties of neural networks</article-title>
          .
          <source>In ICLR.</source>
        </mixed-citation>
      </ref>
      <ref id="ref61">
        <mixed-citation>
          <string-name>
            <surname>Szymanski</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>McCane</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ; and
          <string-name>
            <surname>Albert</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <year>2018</year>
          .
          <article-title>The effect of the choice of neural network depth and breadth on the size of its hypothesis space</article-title>
          . arXiv:
          <year>1806</year>
          .02460 .
        </mixed-citation>
      </ref>
      <ref id="ref62">
        <mixed-citation>
          <string-name>
            <surname>Tausend</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <year>1994</year>
          .
          <article-title>Representing biases for inductive logic programming</article-title>
          .
          <source>In ECML.</source>
        </mixed-citation>
      </ref>
      <ref id="ref63">
        <mixed-citation>
          <string-name>
            <surname>Vapnik</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ; and
          <string-name>
            <surname>Chervonenkis</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <year>2015</year>
          .
          <article-title>On the Uniform Convergence of Relative Frequencies of Events to Their Probabilities</article-title>
          . Springer.
        </mixed-citation>
      </ref>
      <ref id="ref64">
        <mixed-citation>
          2018.
          <article-title>Efficient Formal Safety Analysis of Neural Networks</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref65">
        <mixed-citation>
          <string-name>
            <surname>Witten</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Frank</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ; Hall,
          <string-name>
            <given-names>M.</given-names>
            ; and
            <surname>Pal</surname>
          </string-name>
          ,
          <string-name>
            <surname>C.</surname>
          </string-name>
          <year>2017</year>
          .
          <article-title>Data Mining: Practical Machine Learning Tools</article-title>
          and Techniques.
        </mixed-citation>
      </ref>
      <ref id="ref66">
        <mixed-citation>
          2019.
          <article-title>Training for Faster Adversarial Robustness Verification via Inducing ReLU Stability</article-title>
          . In ICLR.
        </mixed-citation>
      </ref>
      <ref id="ref67">
        <mixed-citation>
          <string-name>
            <surname>Yudkowsky</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          <year>2011</year>
          .
          <article-title>Complex Value Systems are Required to Realize Valuable Futures</article-title>
          . The Singularity Institute, San Francisco, CA.
        </mixed-citation>
      </ref>
      <ref id="ref68">
        <mixed-citation>
          <string-name>
            <surname>Zhang</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Bengio</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ; Hardt,
          <string-name>
            <given-names>M.</given-names>
            ;
            <surname>Recht</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            ; and
            <surname>Vinyals</surname>
          </string-name>
          ,
          <string-name>
            <surname>O.</surname>
          </string-name>
          <year>2017</year>
          .
          <article-title>Understanding deep learning requires rethinking generalization</article-title>
          .
          <source>In ICLR.</source>
        </mixed-citation>
      </ref>
      <ref id="ref69">
        <mixed-citation>
          <string-name>
            <surname>Zhang</surname>
            ,
            <given-names>H.</given-names>
            ; Chen, H.
          </string-name>
          ; Xiao,
          <string-name>
            <given-names>C.</given-names>
            ;
            <surname>Gowal</surname>
          </string-name>
          ,
          <string-name>
            <surname>S.</surname>
          </string-name>
          ; Stanforth,
          <string-name>
            <given-names>R.</given-names>
            ;
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            ;
            <surname>Boning</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            ; and
            <surname>Hsieh</surname>
          </string-name>
          , C.-J.
          <year>2020</year>
          .
          <article-title>Towards Stable and Efficient Training of Verifiably Robust Neural Networks</article-title>
          .
          <source>In ICLR.</source>
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>