=Paper=
{{Paper
|id=Vol-2028/paper7
|storemode=property
|title=A Discussion of Analogical-Proportion Based Inference
|pdfUrl=https://ceur-ws.org/Vol-2028/paper7.pdf
|volume=Vol-2028
|authors=Henri Prade, Gilles Richard
|dblpUrl=https://dblp.org/rec/conf/iccbr/PradeR17a
}}
==A Discussion of Analogical-Proportion Based Inference==
73
A discussion of analogical-proportion based inference
Henri Prade1,2 and Gilles Richard1
1. IRIT, Toulouse University, France
2. QCIS, University of Technology, Sydney, Australia
{prade, richard}@irit.fr,
Abstract. The Boolean expression of an analogical proportion, i.e., a statement
of the form “a is to b as c is to d”, expresses that “a differs from b as c differs from
d, and vice-versa. This is the basis of an analogical inference principle, which is
shown to be a particular instance of the analogical “jump”: from P (s), P (t), and
Q(s), deduce Q(t). Roughly speaking, an analogical proportion sounds like a sort
of qualitative derivative. A counterpart of a first order Taylor-like formula indeed
exists for affine Boolean functions. Affine functions can be predicted without
error by means of analogical proportions. These affine functions are essentially
the constants, the projections, the xor-based functions, and their complements.
We discuss how one might take advantage of this state of fact for refining the
scope of application of the analogical-proportion based inference to subparts of a
Boolean function that may be assumed to be “locally” linear.
1 Introduction
Analogical proportions are statements of the form “a is to b as c is to d” that have
been introduced at the time of Aristotle by mimicking numerical proportions. Such
statements are appealing since they relate comparisons inside pair (a, b) to comparisons
inside pair (c, d), by suggesting that “a differs from b in the same way as c differs from
d”, and for symmetry reason that “b differs from a in the same way as d differs from c”.
Following a series of works aiming at formalizing the idea of analogical propor-
tion in different settings, a Boolean logic modeling has been proposed almost a decade
ago. This modeling formally acknowledges the above intuitive reading of an analogi-
cal proportion. The analogical-proportion based inference amounts to postulating that
if analogical proportions hold on a series of features used to describe four situations
a, b, c, d, such a proportion may also hold for other related attributes as well.
It turns out that such a view has been proved to be quite effective for classifica-
tion tasks in particular. A natural question is then to try to understand why and in what
respect. This question is not straightforward. An interesting clue has been recently ob-
tained when showing that if (and only if) the classification function is an affine Boolean
function, then the analogical-proportion based inference always predicts the right class
[5]. This confirms previous experimental observations.
It also echoes some informal remarks pointing out the fact that an analogical pro-
portion may be reminiscent of a qualitative notion of derivative. Indeed affine Boolean
functions satisfy a first order Taylor-like formula as recalled in this paper. Since any
Boolean function is piecewise linear (in terms of affine Boolean functions), one may
Copyright © 2017 for this paper by its authors. Copying permitted for private and
academic purpose. In Proceedings of the ICCBR 2017 Workshops. Trondheim, Norway
74
2
wonder if one cannot take advantage of these facts for a better adjustment of the scope
of the analogical-proportion based inference inside areas where the classification func-
tions would be presumably linear.
The Boolean logic modeling of analogical proportions provides a simple basis for
computing the result of an analogical inference. The paper mainly aims at pointing
out the interest of a functional view of Boolean expressions when discussing analog-
ical proportion-based inference. The paper first restates the notion of analogical pro-
portion and its Boolean logic modeling. It then discusses the nature of the analogical
proportion-based inference, first showing that it is a particular instance of a general
“analogical jump” pattern, then explaining its link with affine Boolean functions, be-
fore discussing how we might take advantage of this situation for a better focusing of
the analogical proportion-based inference.
2 Analogical proportion
Analogical proportions are statements of the form “a is to b as c is to d” like “Queen
is to King as Woman is to Man”, or “Paris is to France as Madrid is to Spain”. In
this paper, we assume that i) a, b, c, d are described in terms of Boolean features and
they can be represented by the vector of their values on these features, and that ii) the
relevant features are the same for a, b, c, and d. This second hypothesis is obviously
not satisfied in the second example, indeed a and c belong to a conceptual universe (the
one of cities) distinct from the one of b and d (the one of countries). Such more tricky
proportions are discussed in [14].
2.1 Boolean logic modeling
The analogical proportion “a is to b as c is to d”, denoted a : b :: c : d in the following,
intuitively suggests that a differs from b as c differs from d and b differs from a as
d differs from c”. In this subsection, a, b, c, and d are just Boolean variables, which
pertain to the same unique feature for four items. The analogical proportion is logically
expressed as [16] by the quaternary connective:
a : b :: c : d , ((a ∧ ¬b) ≡ (c ∧ ¬d)) ∧ ((¬a ∧ b) ≡ (¬c ∧ d)) (1)
Note that this logical expression of an analogical proportion put forward dissimilarity,
in agreement with the idea that analogy is as much a matter of dissimilarity as a matter
of similarity. Similarity appears in the logically equivalent expression
a : b :: c : d = ((a ∧ d) ≡ (b ∧ c)) ∧ ((¬a ∧ ¬d) ≡ (¬b ∧ ¬c)) (2)
This latter expression states that what a and d have in common (positively or nega-
tively), b and c have it also in common.
Table 1 gives the truth table of a : b :: c : d. We can see that a : b :: c : d is true for
6 patterns: 0000, 1111, 0011, 1100, 0101 and 1010 (in bold in Table 1 ) .
It is easy to see that the logical expression of a : b :: c : d satisfies the key properties
of an analogical proportion, namely
75
3
a b c d a : b :: c : d a b c d a : b :: c : d
0000 1 1000 0
0001 0 1001 0
0010 0 1010 1
0011 1 1011 0
0100 0 1100 1
0101 1 1101 0
0110 0 1110 0
0111 0 1111 1
Table 1. Boolean valuations for a : b :: c : d
– reflexivity: a : b : a : b .
– symmetry: a : b :: c : d ⇒ c : d :: a : b
– central permutation: a : b :: c : d ⇒ a : c :: b : d
Moreover, it is also worth noticing that the analogical proportion is independent with
respect to the positive or negative encoding of a considered feature: a : b :: c : d =
¬a : ¬b :: ¬c : ¬d. Besides, with this definition, the analogical proportion is transitive
in the following sense: (a : b :: c : d) ∧ (c : d :: e : f ) ⇒ a : b :: e : f .
A simple extension of the definition of analogical proportion to Boolean vectors in Bn
of the form a = (a1 , ..., an ) is as follows: a : b :: c : d iff ∀i ∈ [1, n], ai : bi :: ci : di .
2.2 Equation solving
It is an acknowledged property of analogy to be creative. In this modeling, this is related
to the following equation solving problem: find x such as a : b :: c : x holds true, a,
b, and c being given. It is easy to see that the equation has no solution in two cases:
1 : 0 :: 0 : x and 0 : 1 :: 1 : x. When it exists the solution is clearly unique. It was first
suggested by [11,12] that x can be computed as
x , c ≡ (a ≡ b)
where ≡ is the equivalence connective s ≡ t , (¬s ∨ t) ∧ (¬t ∨ s). Moreover, note that
s ≡ t = ¬((s ∧ ¬t) ∨ (¬s ∧ t)) = ¬(s ⊕ t) where ⊕ is the xor connective (exclusive or).
Thus it is clear that c ≡ (a ≡ b) can be rewritten as c ≡ (a ≡ b) = ¬(c ⊕ ¬(a ⊕ b)) =
a ⊕ b ⊕ c since ¬s = s ⊕ 1 and 1 ⊕ 1 = 0. Connectives ≡ and ⊕ are associative
operators. Thus, we can write x = a ⊕ b ⊕ c and Table 2 shows the values of x in the
6 cases where equation a : b :: c : x has a solution, as well as in the two remaining
cases where there is no analogical solution for a : b :: c : x. In these two latter cases
corresponding to patterns 0110 and 1001, we have a reverse analogy [19,20], where “b
is to a as c is to d” holds rather than “a is to b as c is to d”.
a010101 01
b010110 10
c011001 10
x011010 01
Table 2. Solving a : b :: c : x
Remark Interestingly enough, the eight patterns appearing in Table 2 with an even
.
number of 0 and of 1 are involved in the four homogeneous logical proportions (which
includes the analogical proportion and the reverse analogical proportion) [19]. The eight
remaining patterns among the 24 = 16 patterns of Table 1, which have an odd number of
76
4
0 and of 1 (and are at the basis of heterogeneous logical proportions [20]), appear in the
columns of Table 3. The computation of the fourth line in such a case from the first three
lines, in an equation now denoted a/b//c/x, is then given by x = a ⊕ b ⊕ c ⊕ 1. This
is an interesting operator that takes the majority value in a, b, c (the 6 first columns),
provided that it does not lead to unanimity (the last two columns).
a100101 01
b011001 01
c010110 01
x010101 10
Table 3. Solving a/b//c/x
2.3 Analogical proportions induced by the comparison of two objects
.
As soon as we have two distinct Boolean vectors a and d, it is possible to find two other
vectors b and c such that a : b :: c : d. Indeed, let Agr(a, d) be the set of indices where
a and d agree and Dis(a, d) the set of indices where the two vectors differ. Let us now
consider two new vectors b and c such that : ∀i ∈ Agr(a, d), ai = bi = ci = di (all equal
to 1 or all equal to 0) and ∀i ∈ Dis(a, d)(bi = ai and ci = di ) or (bi = ¬ai and ci = ¬di ).
For instance, a = 0110, d = 0011, Agr(a, d) = {1, 3} and Dis(a, d) = {2, 4}.
Then b = 0111 and c = 0010 make a : b :: c : d true. This may be viewed as instances
of the equation solving problem a : x :: x0 : d with two unknowns x and x0 . Obviously,
we have always a solution: x = a and x0 = d or x = d and x0 = a. But as soon
as Dif (a, d) contains at least two indices as in the above example, we have solutions
where the four vectors a, x, x0 , d are distinct, as shown in the example. The creation of
(b, c) from a and d is illustrated in [10] on images, using a non logical approach.
2.4 Non Boolean attributes
Real life datasets rarely involve Boolean features only. There may be a mix between
Boolean and nominal feature (like color), or real-valued features. In the case of nom-
inal attributes, it is quite common to binarise in the following way: for instance color
can take three values red, green, blue which will be coded as 100, 010, 001 (using fea-
tures as isRed, isGreen, isBlue). Then we are back to the Boolean case. The case
of real-valued features is more sophisticated and needs the tool of multi-valued logic
to be properly handled. We refer the interested reader to [21,7] for a comprehensive
development. Nevertheless, in this paper, we strictly stick to the Boolean case.
Another important issue is to get the relevant feature to code a given problem. We
do not focus on this issue here as we consider the vectors coming from existing datasets,
so the coding has already been done.
3 Analogical proportion-based inference
We have seen that we can obtain the solution x, when it exists, of an analogical pro-
portion equation a : b :: c : x as x = a ≡ b ≡ c = a ⊕ b ⊕ c. The analogical
proportion-based inference principle [23] can now be stated as follows:
∀i ∈ {1, ..., n}, ai : bi :: ci : di holds
∀j ∈ {n + 1, ..., p}, aj : bj :: cj : dj holds
77
5
This is a form of analogical reasoning where we transfer knowledge from some compo-
nents of our vectors to their remaining components, tacitly assuming that the values of
the n first components determine the values of the others. An important particular case
of this pattern is when p = n + 1, which corresponds to the situation in classification
where the (n + 1)th component corresponds to the class of the item described by the n
first features. Note also that this pattern is a tool for guessing missing values in a table,
a problem, which has been considered for a long time [1]. Let us now examine how this
inference pattern can be related to a more usual “analogical jump” pattern.
3.1 An instance of a general “analogical jump” pattern
In its simplest form, analogical reasoning, without any reference to the notion of pro-
portion, is usually viewed as a way to infer some new fact on the basis of a single
observation. Analogical reasoning has been mainly formalized in the setting of first or-
der logic [6,13], and in second order logic [9]. A basic pattern for analogical reasoning
is then to consider 2 terms s and t, to observe that they share a property P , and knowing
that another property Q also holds for s, to infer that it holds for t as well. This is known
as the “analogical jump” and can be described with the following simplified inference
pattern, leading (possibly) to a wrong conclusion:
P (s) P (t) Q(s)
(AJ)
Q(t)
Making such an inference pattern valid would require the implicit hypothesis that
P determines Q inasmuch as 6 ∃u P (u) ∧ ¬Q(u). This may be ensured if there exists
an underlying functional dependency, or more generally, if it is known for instance that
when something is true for an object of a certain type, then it is true for all objects of
that type. Otherwise, without such guarantees, the result of an analogical inference may
turn to be definitely wrong.
To link the above analogical pattern with the concept of analogical proportion, it
is tempting to write something like: P (s) : P (t) :: Q(s) : Q(t) since we have 4
terms which obey, at least from a syntactic viewpoint, the structure of an analogical
proportion. Indeed, it is sufficient to encode each piece of information in a binary way
according to the presence or the absence of P , Q, s, or t in the corresponding term, and
we get the encoding d of Q(t) via the equation solving process as in Table 4. In that
P Q s t
a 1 0 1 0 P (s)
b 1 0 0 1 P (t)
c 0 1 1 0 Q(s)
−−− −−− −−− −−− −−− −−−
d 0 1 0 1 Q(t)
Table 4. A syntactic view of analogical jump
case, a = P (s), b = P (t), c = Q(s), d = Q(t) are encoded as Boolean vectors where
the semantics carried by the predicate symbols P and Q is not considered.
78
6
In [26,2,15,3], the authors take a similar inspiration where, starting from Boolean
datasets and focusing on binary classification problem, they apply the following infer-
ence principle (and obtain competitive results on benchmark data sets):
a : b :: c : d
AP
cl(a) : cl(b) :: cl(c) : cl(d)
It means that if 4 Boolean vectors build a valid analogical proportion, then it should
be true that their classes build also a valid proportion. Starting from this viewpoint, in
the case where a, b, c are in a sample set, i.e., their classes are known, and d being the
object to be classified, if the equation cl(a) : cl(b) :: cl(c) : x = 1 is solvable (in that
case, we say that the triple (a, b, c) is class solvable), they allocate its solution to cl(d)
just by applying the previous principle. Experiments highlight the predictive power of
this principle. Let us understand why AP principle is just a particular instance of (AJ):
– Considering a and b as Boolean vectors in Bn , the vector k = a − b belongs
to {−1, 0, 1}n and summarizes the result of the comparison between a and b. So
given such a vector k, we define the predicate Pk (a, b) := (a − b = k).
– Then we can consider 3 predicate symbols Q1 , Q2 , Q3 defined as follows:
1. Q1 (a, b) := (cl(a) = cl(b))
2. Q2 (a, b) := (cl(a) = 0) ∧ (cl(b) = 1))
3. Q3 (a, b) := (cl(a) = 1) ∧ (cl(b) = 0))
Let us note that the Qi ’s are pairwise mutually exclusive predicates. Using these predi-
cate symbols, the following rule:
Pk (a, b) Pk (c, d) Qi (a, b)
Qi (c, d)
is just an instance of (AJ). Moreover, it states that when the difference a − b equals
to c − d, then the relation between cl(a) and cl(b) is the same as the relation between
cl(c) and cl(d). If we notice that a : b :: c : d is just equivalent to Pk (a, b) ∧ Pk (c, d)
(in the exact sense of the formal definition of the analogical proportion applied compo-
nentwise), and Qi (a, b) ∧ Qi (c, d) entails cl(a) : cl(b) :: cl(c) : cl(d), we obtain:
a : b :: c : d
cl(a) : cl(b) :: cl(c) : cl(d)
which is exactly the expression of AP used in [2].
In pattern AP , we transfer the identity of differences pertaining to pairs (a, b) and
(c, d) to the relation between their classes. It enables us to predict the missing informa-
tion about d, using AP as an extrapolation principle. This is clearly a form of reasoning
that is not sound, but which may be useful for trying to guess unknown values.
3.2 Link with affine Boolean functions
As already mentioned [4], an analogical proportion of the form “cl(b) is to cl(a) as b
is to a” sounds a bit like the expression of the qualitative derivative of a function cl un-
derlying the classification process, since the derivative of a function f in a is the limit
79
7
when x −→ a of the ratio f (x)−f x−a
(a)
, which is a matter of comparing two algebraic
1
differences . Moreover, considering the 6 patterns that make true the analogical propor-
tion, it can be also noticed that if there is a change from a to b, there should a change in
the same direction from c to d. Besides, it has been observed that once extended from
Boolean to graded scales, the analogical proportion-based inference provides a linear
interpolation mechanism [7]. This is due to the fact that in this case a : b :: c : d = 1
if and only if d − c = b − a (where a, b, c, d ∈ [0, 1]) and the solution of the equation
a : x :: x : d = 1 is x = a+d2 .
In fact, it has been recently formally proved [5] that the AP principle is sound as
soon as the labeling function is an affine Boolean function (this means in practice that
the function is a constant, a projection, an xor function, or an ≡ function, over some
subsets of the n attributes). Moreover it can be shown that there is no other Boolean
function for which this is true [5].
It is well-known that any Boolean function can be put in a polynomial form where
the sum is taken as ⊕ and the product as the min [22]. With a functional view of Boolean
expressions in mind, one can use the notion of qualitative derivative and define a Taylor-
like development of a Boolean function; see, e.g., [17,25]. For instance, consider the
linear function f (x, y) = x ⊕ y (indeed a polynomial of degree 1 and arity 2). Then it
can be checked (see Table 5), that we can write a Taylor-like development of the form:
f (x, y) = f (a, b) ⊕ ∂xf (a) ∧ (x a) ⊕ ∂yf (b) ∧ (y b) , Σ(x, y)
where s t = s ⊕ t since s t = x ⇔ s = t ⊕ x, and where all the partial derivatives
are equal to 1 here. One can also rewrite the above equality as f (x, y) f (a, b) =
(x a) ⊕ (y b), which is indeed the Boolean counterpart of what holds for affine
functions in Rn .
x y a b f (x, y) = x ⊕ y f (a, b) = a ⊕ b ∂xf (a) ∧ (x a) ∂yf (b) ∧ (y b) Σ(x, y)
0000 0 0 1 ∧ (0 0) = 0 1 ∧ (0 0) = 0 0
0001 0 1 1 ∧ (0 0) = 0 1 ∧ (0 1) = 1 0
0010 0 1 1 ∧ (0 1) = 1 1 ∧ (0 0) = 0 0
0011 0 0 1 ∧ (0 1) = 1 1 ∧ (0 1) = 1 0
0100 1 0 1 ∧ (0 0) = 0 1 ∧ (1 0) = 1 1
0101 1 1 1 ∧ (0 0) = 0 1 ∧ (1 1) = 0 1
0110 1 1 1 ∧ (0 1) = 1 1 ∧ (1 0) = 1 1
0111 1 0 1 ∧ (0 1) = 1 1 ∧ (1 1) = 0 1
1000 1 0 1 ∧ (1 0) = 1 1 ∧ (0 0) = 0 1
1001 1 1 1 ∧ (1 0) = 1 1 ∧ (0 1) = 1 1
1010 1 1 1 ∧ (1 1) = 0 1 ∧ (0 0) = 0 1
1011 1 0 1 ∧ (1 1) = 0 1 ∧ (0 1) = 1 1
1100 0 0 1 ∧ (1 0) = 1 1 ∧ (1 0) = 1 0
1101 0 1 1 ∧ (1 0) = 1 1 ∧ (1 1) = 0 0
1110 0 1 1 ∧ (1 1) = 0 1 ∧ (1 0) = 1 0
1111 0 0 1 ∧ (1 1) = 0 1 ∧ (1 1) = 0 0
Table 5. Taylor-like expression of the linear function f (x, y) = x ⊕ y
1
We may also remember that the idea of differential has inspired adaption methods in case-
.
based reasoning for solving numerical problems [8], even if case-based reasoning deals with cases
one by one rather than handling triples of cases.
80
8
3.3 Adjusting analogical proportion-based inference
Since analogical proportion-based inference works perfectly for predicting affine Boolean
functions, it is natural to wonder if given a training set T it would not be possible to
cover it with a piecewise affine Boolean function.
The answer is yes, ... and in many different ways! It is easy to see that we just need
two pieces. Indeed let f (x1 , · · · , xn ) be an affine function. In a binary classification
problem, any function (be affine or not) partitions T into two parts Tf and Tf ⊕1 on
which respectively f and f ⊕1 correctly predict the class (since f ⊕1 is just ¬f ).
If the classification in T obeys an affine Boolean function f then either Tf or Tf ⊕1
is empty. This also means that the application of the analogical-proportion principle
will amount to apply function f to a new item to be classified (or function f ⊕1 if Tf
is empty). So if the training set T is covered by an “almost” linear function, this means
that one of the two subsets of the partition of T is very large with respect to the other.
When the training set is not covered by a unique affine Boolean function, this means
that there exist triples that lead to false predictions when applying AP . Then we might
think that it happens more often when a, b, c do not all belong to the same subset in
some partition induced by a function f . So an idea for “almost” linear functions, would
be to look for the affine Boolean functions such as Tf (or Tf ⊕1 ) is the largest possible
subset, which would make easier the finding of triples such as all the three a, b, c are
in it. However another issue is to wonder if some partitions are more appropriate than
others for guessing the class of a particular new item.
Generally speaking, the issue is to find a way to identify those triples that are “sus-
pect”, i.e., likely to yield a faulty prediction, among a set of candidate triples that en-
ables you to apply AP . Indeed in case of multiple triples, which is the usual situation,
we apply a voting procedure among the predictions of the applicable triples, where
sometimes the faulty triples are the majority. How to restrict this voting procedure to
“good triples”? Another idea may come from a careful examination of the way triples
are built and of the meaning of pairs inside, as first suggested in [4].
In Table 6, we have reordered the vectors in a particular way. Indeed the table shows
that building the analogical proportion a : b :: c : d is a matter of pairing the pair
(a, b) with the pair (c, d). More precisely, on features or attributes A1 to Aj−1 , the
four vectors are equal; on attributes Aj to Ar−1 , a = b and c = d, but (a, b) 6= (c, d).
In other words, on attributes A1 to Ar−1 a and b agree and c and d agree as well.
This contrasts with attributes Ar to An , for which we can see that a differs from b as
c differs from d (and vice-versa). In columns we recognize the 6 patterns that makes
the analogical proportion true. There are two cases, either cl(a) = cl(b) (and then
cl(x) = cl(c)), or cl(a) 6= cl(b) (and then cl(x) = cl(b)). In the first case, it suggests
that the particular change observed between a and b on features from Ar to An does
not affect cl in the context defined by the values of the features from A1 to Ar−1 where
a and b are equal. Applying AP amounts to assuming that this absence of effect is true
in other contexts of values of features from A1 to Ar−1 . So the smaller the number
of features from Ar to An , the more cautious. A similar reasoning can be done when
cl(a) 6= cl(b) where the change on the features from Ar to An should be responsible
of the change of class in the context of the values of the other attributes. Observe also
that if we have two pairs (a, b) and (a0 , b0 ) such as a0 : b0 :: a : b, while a : b :: c : x,
81
9
then by transitivity we have a0 : b0 :: c : x. Thus transitivity agrees with the idea that if
a change has an effect (or no effect) in some context, then it may be the same elsewhere.
A1 ... Ai−1 Ai ... Aj−1 Aj ... Ak−1 Ak ... Ar−1 Ar ... As−1 As ... An cl
a 1 ... 1 0 ... 0 1 ... 1 0 ... 0 1 ... 1 0 ... 0 cl(a)
b 1 ... 1 0 ... 0 1 ... 1 0 ... 0 0 ... 0 1 ... 1 cl(b)
c 1 ... 1 0 ... 0 0 ... 0 1 ... 1 1 ... 1 0 ... 0 cl(c)
x 1 ... 1 0 ... 0 0 ... 0 1 ... 1 0 ... 0 1 ... 1 cl(x)?
Table 6. Pairing pairs (a, b) and (c, d)
A last idea would be to consider “continuous” analogical proportion and to solve
interpolative equation a : x :: x : b. In a Boolean setting, such an equation has no
solution, except in the trivial situation where a = b, then x = a. In the associated class
equation one has necessarily cl(a) = cl(x) = cl(b). Then one may relax a : x :: x : b
to a subset of features and makes sure that x is between a and b in the sense that
max(h(a, x), h(x, b)) ≤ h(a, b), where h is the Hamming distance. In such a case,
we have a variant of nearest neighbors methods.
We have emphasized the role played by affine Boolean functions, suggesting that
the training set in a classification problem might be restricted to subsets of examples
more relevant for a new item to be classified. These subsets of examples should be
covered by some affine Boolean function. Finding them remains an open question.
4 Conclusion
The paper has intended to provide an advanced discussion of the analogical proportion-
based inference principle in the Boolean case, in a classification perspective. As already
said, analogical proportion-based inference is also available for nominal and real valued
data. The application of analogical proportions to regression is an open problem; then
the agreement between a qualitative and a quantitative view of these proportions is cru-
cial (see, e.g., [24] on such issue in learning); in that respect the main gradual extension
[21,7] clearly distinguishes between situations where the changes from a to b and from
c to d are in the same direction, and where the changes are in opposite directions.
Generally speaking, some authors, e.g., [18], view qualitative reasoning as made of
components such as: comparison, categorization, identification of relations, and emer-
gence of a meaning. Analogical proportions seem to offer an interesting mixture of at
least two or three of these ingredients [14]; the proper understanding of their interrela-
tionships is still to be further explored.
References
1. M. Abraham, D. M. Gabbay, and U. J. Schild. Analysis of the talmudic argumentum a fortiori
inference rule (kal vachomer) using matrix abduction. Studia Logica, 92(3):281–364, 2009.
2. S. Bayoudh, L. Miclet, and A. Delhay. Learning by analogy: A classification rule for binary
and nominal data. Proc. Int. Joint Conf. on Artif. Intellig. (IJCAI’07), pages 678–683, 2007.
3. M. Bounhas, H. Prade, and G. Richard. Analogical classification: A new way to deal with
examples. In Proc. 21st Eur.Conf. Artif. Intell. (ECAI’14), Prague, IOS Press,135-140, 2014.
4. W. Correa, H. Prade, and G. Richard. Trying to understand how analogical classifiers work.
In E. Hüllermeier et al., editor, Proc. 6th Inter. Conf. on Scalable Uncertainty Management
(SUM’12), Marburg, Germany, LNAI 7520, pages 582–589. Springer Verlag, 2012.
82
10
5. M. Couceiro, N. Hug, H. Prade, and G. Richard. Analogy-preserving functions: A way to
ext-end Boolean samples. In Proc.26th Int. J. Conf. Artif. Intel. IJCAI’17, Melbourne, 2017.
6. T. R. Davies and Russell S. J. A logical approach to reasoning by analogy. In J. P. Mc-
Dermott, editor, Proc. of the 10th International Joint Conference on Artificial Intelligence
(IJCAI’87), Milan, pages 264–270. Morgan Kaufmann, 1987.
7. D. Dubois, H. Prade, and G. Richard. Multiple-valued extensions of analogical proportions.
Fuzzy Sets and Systems, 292:193–202, 2016.
8. B. Fuchs, J. Lieber, A. Mille, and A. Napoli.Differential adaptation: An operational approach
to adaptation for solving numerical problems with CBR. Know.-Bd. Sys., 68:103–114, 2014.
9. H. Gust, K. Kühnberger, and U. Schmid. Metaphors and heuristic-driven theory projection
(HDTP). Theoretical Computer Science, 354(1):98 – 117, 2006.
10. L. Jing, Y. Lu, H. Gang, and K. Sing Bing. Visual attribute transfer through deep image
analogy. In ACM Transactions on Graphics, (Proc. of Siggraph), volume 36, 2017.
11. S. Klein. Culture, mysticism & social structure and the calculation of behavior. In Proc. 5th
Europ. Conf. in Artificial Intelligence (ECAI’82), Orsay, France, pages 141–146, 1982.
12. S. Klein. Analogy and mysticism and the structure of culture (and Comments & Reply).
Current Anthropology, 24 (2):151–180, 1983.
13. E. Melis and M. Veloso. Analogy in problem solving. In Handbook of Practical Reasoning:
Computational and Theoretical Aspects. Oxford Univ. Press, 1998.
14. L. Miclet, N. Barbot, and H. Prade. From analogical proportions in lattices to proportional
analogies in formal concepts. In T. Schaub, G. Friedrich, and B. O’Sullivan, editors, Proc.
21st Europ. Conf. on Artificial Intelligence (ECAI’14), Prague, Aug. 18-22, volume 263 of
Frontiers in Artificial Intelligence and Applications, pages 627–632. IOS Press, 2014.
15. L. Miclet, S. Bayoudh, and A. Delhay. Analogical dissimilarity: Definition, algorithms and
two experiments in machine learning. J. Artif. Intell. Res. (JAIR), 32:793–824, 2008.
16. L. Miclet and H. Prade. Handling analogical proportions in classical logic and fuzzy logics
settings. In Proc. 10th Eur. Conf. on Symbolic and Quantitative Approaches to Reasoning
with Uncertainty (ECSQARU’09),Verona, pages 638–650. Springer, LNCS 5590, 2009.
17. E. Mizraji. Differential and integral calculus for logical operations. a matrix-vector approach.
J. Log. Comput., 25(3):613–638, 2015.
18. A. Mucchielli. Les processus intellectuels fondamentaux sous-jacents aux techniques et
mthodes qualitatives. RecherchesQualitatives, 3:1–27, 2006.
19. H. Prade and G. Richard. From analogical proportion to logical proportions. Logica Univer-
salis, 7(4):441–505, 2013.
20. H. Prade and G. Richard. Homogenous and heterogeneous logical proportions. IfCoLog J.
of Logics and their Applications, 1(1):1–51, 2014.
21. H. Prade and G. Richard. Multiple-valued logic interpretations of analogical, reverse analog-
ical, and paralogical proportions. In Proc. 40th IEEE Int. Symp. on Multiple-Valued Logic
(ISMVL’10), pages 258–263, Barcelona, 2010.
22. M. H. Stone. The theory of representation for Boolean algebras. Trans. of the American
Mathematical Society, 40(1):37111, 1936.
23. N. Stroppa and F. Yvon. Analogical learning and formal proportions: Definitions and
methodological issues. Technical Report D004, ENST-Paris, 2005.
24. Dorian Suc, Daniel Vladusic, and Ivan Bratko. Qualitatively faithful quantitative prediction.
Artificial Intelligence, 158(2):189 – 214, 2004.
25. S. N. Yanushkevich and V. P. Shmerko. Taylor expansion of logic functions: From conventio-
nal to nanoscale design. Int. TICSP Wksp Spectral Meth. & Multirate Signal Process.,2004.
26. F. Yvon and N. Stroppa. Formal models of analogical proportions. Technical Report D008,
Ecole Nationale Supérieure des Télécommunications, Paris, France, 2006.