=Paper= {{Paper |id=Vol-1339/paper2 |storemode=property |title=On Indicative Conditionals |pdfUrl=https://ceur-ws.org/Vol-1339/paper2.pdf |volume=Vol-1339 |dblpUrl=https://dblp.org/rec/conf/iwost/DietzHP15 }} ==On Indicative Conditionals== https://ceur-ws.org/Vol-1339/paper2.pdf
                     On Indicative Conditionals

     Emmanuelle Anna Dietz1 , Steffen Hölldobler1 , and Luı́s Moniz Pereira2?
        1
        International Center for Computational Logic, TU Dresden, Germany,
                            {dietz,sh}@iccl.tu-dresden.de
    2
      NOVA Laboratory for Computer Science and Informatics, Caparica, Portugal,
                                    lmp@fct.unl.pt



        Abstract. In this paper we present a new approach to evaluate indica-
        tive conditionals with respect to some background information specified
        by a logic program. Because the weak completion of a logic program ad-
        mits a least model under the three-valued Lukasiewicz semantics and this
        semantics has been successfully applied to other human reasoning tasks,
        conditionals are evaluated under these least L-models. If such a model
        maps the condition of a conditional to unknown, then abduction and
        revision are applied in order to satisfy the condition. Different strategies
        in applying abduction and revision might lead to different evaluations of
        a given conditional. Based on these findings we outline an experiment to
        better understand how humans handle those cases.


1     Indicative Conditionals
Conditionals are statements of the form if condition then consequence. In the
literature the condition is also called if part, if clause or protasis, whereas the
consequence is called then part, then clause or apodosis. Conditions as well as
consequences are assumed to be finite sets (or conjunctions) of ground literals.
    Indicative conditionals are conditionals whose condition may or may not be
true and, consequently, whose consequence also may or may not be true; however,
the consequence is asserted to be true if the condition is true. Examples for
indicative conditionals are the following:

                                                 If it is raining, then he is inside. (1)
    If Kennedy is dead and Oswald did not shoot him, then someone else did. (2)
                         If rifleman A did not shoot, then the prisoner is alive. (3)
                        If the prisoner is alive, then the captain did not signal. (4)
                               If rifleman A shot, then rifleman B shot as well. (5)
                If the captain gave no signal and rifleman A decides to shoot,
                      then the prisoner will die and rifleman B will not shoot. (6)

   Conditionals may or may not be true in a given scenario. For example, if we
are told that a particular person is living in a prison cell, then most people are
?
    The authors are mentioned in alphabetical order.
20




expected to consider (1) to be true, whereas if we are told that he is living in
the forest, then most people are expected to consider (1) to be false. Likewise,
most people consider (2) to be true.
    The question which we shall be discussing in this paper is how to automate
reasoning such that conditionals are evaluated by an automated deduction sys-
tem like humans do. This will be done in a context of logic programming (cf.
[11]), abduction [9], Stenning and van Lambalgen’s representation of condition-
als as well as their semantic operator [19] and three-valued Lukasiewicz logic
[12], which has been put together in [6,7,5,8,3] and has been applied to the sup-
pression [2] and the selection task [1], as well as to model the belief-bias effect
[15] and contextual abductive reasoning with side-effects [16].
    The methodology of the approach presented in this paper differs significantly
from methods and techniques applied in well-known approaches to evaluate
(mostly subjunctive) conditionals like Ramsey’s belief-retention approach [17],
Lewis’s maximal world-similarity one [10], Rescher’s systematic reconstruction
of the belief system using principles of saliency and prioritization [18], Ginsberg’s
possible worlds approach [4] and Pereira and Aparı́co’s improvements thereof by
requiring relevancy [14]. Our approach is inspired by Pearl’s do-calculus [13] in
that it allows revisions to satisfy conditions whose truth-value is unknown and
which cannot be explained by abduction, but which are amenable to hypothetical
intervention instead.

2    Preliminaries
We assume the reader to be familiar with logic and logic programming. A (logic)
program is a finite set of (program) clauses of the form A ← B1 ∧ . . . ∧ Bn where
A is an atom and Bi , 1 ≤ i ≤ n, are literals or of the form > and ⊥, denoting
truth- and falsehood, respectively. A is called head and B1 ∧ . . . ∧ Bn is called
body of the clause. We restrict terms to be constants and variables only, i.e., we
consider so-called data logic programs. Clauses of the form A ← > and A ← ⊥
are called positive and negative facts, respectively.
    In the this paper we assume for each program that the alphabet consists
precisely of the symbols mentioned in the program. When writing sets of literals
we will omit curly brackets if the set has only one element.
    Let P be a program. gP denotes the set of all ground instances of clauses
occurring in P. A ground atom A is defined in gP iff gP contains a clause whose
head is A; otherwise A is said to be undefined. Let S be a set of ground literals.
def (S, P) = {A ← body ∈ gP | A ∈ S ∨ ¬A ∈ S} is called definition of S.
    Let P be a program and consider the following transformation:
 1. For each defined atom A, replace all clauses of the form A ← body 1 , . . . ,
     A ← body m occurring in gP by A ← body 1 ∨ . . . ∨ body m .
 2. If a ground atom A is undefined in gP, then add A ← ⊥ to the program.
 3. Replace all occurrences of ← by ↔.
The ground program obtained by this transformation is called completion of P,
whereas the ground program obtained by applying only the steps 1. and 3. is
called weak completion of P or wcP.
                                                                                  21




    We consider the three-valued Lukasiewicz (or L-) semantics [12] and represent
each interpretation I by a pair hI > , I ⊥ i, where I > contains all atoms which are
mapped to true by I, I ⊥ contains all atoms which are mapped to false by I, and
I > ∩ I ⊥ = ∅. Atoms occurring neither in I > not in I ⊥ are mapped to unknown.
Let hI > , I ⊥ i and hJ > , J ⊥ i be two interpretations. We define

                 hI > , I ⊥ i ⊆ hJ > , J ⊥ i iff I > ⊆ J > and I ⊥ ⊆ J ⊥ .

Under L-semantics we find F ∧ > ≡ F ∨ ⊥ ≡ F for each formula F , where ≡
denotes logical equivalence. Hence, occurrences of the symbols > and ⊥ in the
bodies of clauses can be restricted to those occurring in facts.
   It has been shown in [6] that logic programs as well as their weak completions
admit a least model under L-semantics. Moreover, the least L-model of the weak
completion of P can be obtained as least fixed point of the following semantic
operator, which was introduced in [19]: ΦP (hI > , I ⊥ i) = hJ > , J ⊥ i, where

    J > = {A | A ← body ∈ gP and body is true under hI > , I ⊥ i},
    J ⊥ = {A | def (A, P) 6= ∅ and
               body is false under hI > , I ⊥ i for all A ← body ∈ def (A, P)}.

We define P |=Llmwc F iff formula F holds in the least L-model of wcP.
    As shown in [2], the L-semantics is related to the well-founded semantics
as follows: Let P be a program which does not contain a positive loop and let
P + = P \ {A ← ⊥ | A ← ⊥ ∈ P}. Let u be a new nullary relation symbol not
occurring in P and B be a ground atom in

               P ∗ = P + ∪ {B ← u | def (B, P) = ∅} ∪ {u ← ¬u}.

Then, the least L-model of wcP and the well-founded model for P ∗ coincide.
    An abductive framework consists of a logic program P, a set of abducibles
AP = {A ← > | A is undefined in gP} ∪ {A ← ⊥ | A is undefined in gP}, a set
of integrity constraints IC, i.e., expressions of the form ⊥ ← B1 ∧ . . . ∧ Bn , and
the entailment relation |=lmwc
                           L     , and is denoted by hP, AP , IC, |=Llmwc i.
    One should observe that each finite set of positive and negative ground facts
has an L-model. It can be obtained by mapping all heads occurring in this set
to true. Thus, in the following definition, explanations are always satisfiable.
    An observation O is a set of ground literals; it is explainable in the abductive
framework hP, AP , IC, |=lmwc
                           L    i iff there exists an E ⊆ AP called explanation
such that P ∪ E is satisfiable, the least L-model of the weak completion of P ∪ E
satisfies IC, and P ∪ E |=lmwc
                           L      L for each L ∈ O.


3   A Reduction System for Indicative Conditionals

When parsing conditionals we assume that information concerning the mood of
the conditionals has been extracted. In this paper we restrict our attention to
indicative mood. In the sequel let cond (T , A) be a conditional with condition T
22




and consequence A, both of which are assumed to be finite sets of literals not
containing a complementary pair of literals, i.e., a pair B and ¬B.
    Conditionals are evaluated wrt background information specified as a logic
program and a set of integrity constraints. More specifically, as the weak com-
pletion of each logic program always admits a least L-model, the conditionals
are evaluated under these least L-models. In the reminder of this section let P
be a program, IC be a finite set of integrity constraints, and MP be the least
L-model of wcP such that MP satisfies IC. A state is either an expression of
the form ic(P, IC, T , A) or true, false, unknown, or vacuous.


3.1    A Revision Operator

Let S be a finite set of ground literals not containing a complementary pair of
literals and let B be a ground atom in

      rev (P, S) = (P \ def (S, P)) ∪ {B ← > | B ∈ S} ∪ {B ← ⊥ | ¬B ∈ S}.

The revision operator ensures that all literals occurring in S are mapped to true
under the least L-model of wc rev (P, S).


3.2    The Abstract Reduction System

Let cond (T , A) be an indicative conditional which is to be evaluated in the
context of a logic program P and integrity constraints IC such that the least
L-model MP of wcP satisfies IC. The initial state is ic(P, IC, T , A).
   If the condition of the conditional is true, then the conditional holds if its
consequent is true as well; otherwise it is either false or unknown.

 ic(P, IC, T , A) −→it true    iff MP (T ) = true and MP (A) = true
 ic(P, IC, T , A) −→if false   iff MP (T ) = true and MP (A) = false
 ic(P, IC, T , A) −→iu unknown iff MP (T ) = true and MP (A) = unknown

   If the condition of the conditional is false, then the conditional is true under
L-semantics. However, we believe that humans might make a difference between
a conditional whose condition and consequence is true and a conditional whose
condition is false. Hence, for the time being we consider a conditional whose
condition is false as vacuous.

              ic(P, IC, T , A) −→iv vacuous iff MP (T ) = false

   If the condition of the conditional is unknown, then we could assign a truth-
value to the conditional in accordance with the L-semantics. However, we suggest
that in this case abduction and revision shall be applied in order to satisfy the
condition. We start with the abduction rule:

                  ic(P, IC, T , A) −→ia ic(P ∪ E, IC, T \ O, A)
                                                                                   23




iff MP (T ) = unknown and E explains O ⊆ T in the abductive framework
hP, AP , IC, |=Llmwc i and O 6= ∅. Please note that T may contain literals which
are mapped to true by MP . These literals can be removed from T by the rule
−→ia because the empty set explains them.
    Now we turn to the revision rule:
                  ic(P, IC, T , A) −→ir ic(rev (P, S), IC, T \ S, A)
iff MP (T ) = unknown, S ⊆ T , S =      6 ∅, for each L ∈ S we find MP (L) =
unknown, and the least L-model of wc rev (P, S) satisfies IC.
    Altogether we obtain the reduction system RIC operating on states and
consisting of the rules {−→it , −→if , −→iu , −→iv , −→ia , −→ir }.

4     Examples
4.1     Al in the Jailhouse
Rainy Day Suppose we are told that Al is imprisoned in a jailhouse on a rainy
day, i.e.., he is living in a cell inside the jailhouse and it is raining:
      P1 = {inside(X) ← imprisoned (X), imprisoned (al ) ← >, raining ← >}.
The least L-model of wcP1 is h{imprisoned (al ), inside(al ), raining}, ∅i. In order
to evaluate conditional (1) with respect to P1 we observe that this model maps
raining and inside to true. Hence,
                         ic(P1 , ∅, raining, inside) −→it true.

Sunny Day       Let us assume that Al is still imprisoned but that it is not raining:
      P2 = {inside(X) ← imprisoned (X), imprisoned (al ) ← >, raining ← ⊥}.
The least L-model of wcP2 is h{imprisoned (al ), inside(al )}, {raining}i. In order
to evaluate conditional (1) wrt P2 we observe that this model maps raining to
false. Hence,
                    ic(P2 , ∅, raining, inside) −→iv vacuous.

No Information about the Weather Suppose we are told that Al is imprisoned
in a jailhouse but we know nothing about the weather:
            P3 = {inside(X) ← imprisoned (X), imprisoned (al ) ← >}.
The least L-model of wcP1 is h{imprisoned (al ), inside(al )}, ∅i. In order to eval-
uate conditional (1) wrt P3 we observe that this model maps raining to un-
known. Hence, we view raining as an observation which needs to be explained.
The only possible explanation wrt hP3 , {raining ← >, raining ← ⊥}, ∅, |=lmwc L    i
is {raining ← >}. Altogether we obtain
            ic(P3 , ∅, raining, inside) −→ia ic(P1 , ∅, ∅, inside) −→it true.
Please note that P3 ∪ {raining ← >} = P1 = rev (P3 , raining). Hence, we could
replace −→ia by −→ir in the previous reduction sequence.
24




4.2    The Shooting of Kennedy
President Kennedy was killed. There was a lengthy investigation about who
actually shot the president and in the end it was determined that Oswald did it:
     P4 = {Kennedy dead ← os shot, Kennedy dead ← se shot, os shot ← >}.
The least L-model of wcP4 is h{os shot, Kennedy dead }, ∅i. Evaluating the in-
dicative conditional (2) under this model we find that its condition
T = {Kennedy dead , ¬os shot} is mapped to false. Hence,
           ic(P4 , ∅, {Kennedy dead , ¬os shot}, se shot) −→iv vacuous.
Now consider the case that we do not know that Oswald shot the president:
           P5 = {Kennedy dead ← os shot, Kennedy dead ← se shot}.
As least L-model of wcP5 we obtain h∅, ∅i and find that it maps T to un-
known. We may try to consider T as an observation and explain it wrt the
abductive framework hP5 , AP5 , ∅, |=Llmwc i, where AP5 consists of the positive
and negative facts for os shot and se shot. The only possible explanation is
E = {os shot ← ⊥, se shot ← >}. As least L-model of wc(P5 ∪ E) we obtain
h{Kennedy dead , se shot}, {os shot}i. As this model maps se shot to true we
find
                 ic(P5 , ∅, {Kennedy dead , ¬os shot}, se shot)
                 −→ia ic(P5 ∪ E, ∅, ∅, se shot) −→it true.
In this example we could also apply revision. Let
             P6 = rev (P5 , T ) = {Kennedy dead ← >, os shot ← ⊥}.
We obtain
                  ic(P5 , ∅, {Kennedy dead , ¬os shot}, se shot)
                  −→ir ic(P6 , ∅, ∅, se shot) −→iu unknown
because the least L-model of wcP6 is h{Kennedy dead }, {os shot}i and maps
se shot to unknown. However, as conditional (2) can be evaluated by abduction
and without revising the initial program, this derivation is not preferred.

4.3    The Firing Squad
This example is presented in [13]. If the court orders an execution, then the
captain will give the signal upon which riflemen A and B will shoot the prisoner.
Consequently, the prisoner will be dead. We assume that the court’s decision is
unknown, that both riflemen are accurate, alert and law-abiding, and that the
prisoner is unlikely to die from any other causes. Let
              P7 = { sig ← execution, rmA ← sig, rmB ← sig,
                     dead ← rmA, dead ← rmB , alive ← ¬dead }.
The least L-model of wcP7 is
                                      h∅, ∅i.                                 (7)
                                                                             25




Rifleman A did not Shoot To evaluate conditional (3) wrt this model we first
observe that the condition rmA is mapped to unknown by (7). Considering the
abductive framework

               hP7 , {execution ← >, execution ← ⊥}, ∅, |=lmwc
                                                          L    i,           (8)

¬rmA can be explained by

                                {execution ← ⊥}.                            (9)

Let P8 = P7 ∪ (9). The least L-model of wcP8 is

                  h{alive}, {execution, sig, rmA, rmB , dead }i.           (10)

Because alive is mapped to true under this model, we obtain

            ic(P7 , ∅, ¬rmA, alive) −→ia ic(P8 , ∅, ∅, alive) −→it true.

The Prisoner is Alive    Now consider conditional (4). Because (7) maps alive
to unknown we treat alive as an observation. Considering again the abductive
framework (8) this observation can be explained by (9). Hence, we evaluate the
consequence of (4) under (10) and find that the captain did not signal:

             ic(P7 , ∅, alive, ¬sig) −→ia ic(P8 , ∅, ∅, ¬sig) −→it true.

Rifleman A Shot     Let us turn the attention to conditional (5). Because (7)
maps rmA to unknown, we treat rmA as an observation. Considering the ab-
ductive framework (8) this observation can be explained by

                                {execution ← >}.                           (11)

Let P9 = P7 ∪ (11). The least L-model of wcP9 is

                  h{execution, sig, rmA, rmB , dead }, {alive}i.           (12)

Because rmB is mapped to true under this model, we obtain

             ic(P7 , ∅, rmA, rmB ) −→ia ic(P9 , ∅, ∅, rmB ) −→it true.

The Captain Gave no Signal Let us now consider conditional (6). Its condition
T = {¬sig, rmA} is mapped to unknown by (7). We can only explain ¬sig by (9)
and rmA by (11), but we cannot explain T because

          wc((9) ∪ (11)) = {execution ↔ > ∨ ⊥} ≡ {execution ↔ >}.

In order to evaluate this conditional we have to consider revisions.
26




1. A brute force method is to revise the program wrt all conditions. Let
             P10 = rev (P7 , {¬sig, rmA})
                 = (P7 \ def ({¬sig, rmA}, P7 )) ∪ {sig ← ⊥, rmA ← >}.
     The least L-model of wcP10 is
                           h{rmA, dead }, {sig, rmB , alive}i.               (13)
     This model maps dead to true and rmB to false and we obtain
                     ic(P7 , ∅, {¬sig, rmA}, {dead , ¬rmB })
                     −→ir ic(P10 , ∅, ∅, {dead , ¬rmB }) −→it true.
2. As we prefer minimal revisions let us consider
             P11 = rev (P7 , rmA) = (P7 \ def (rmA, P7 )) ∪ {rmA ← >}.
     The least L-model of wcP11 is h{dead , rmA}, {alive}i. Unfortunately, ¬sig
     is still mapped to unknown by this model, but it can be explained in the ab-
     ductive framework hP11 , {execution ← >, execution ← ⊥}, ∅, |=Llmwc i by (9).
     Let P12 = P11 ∪ (9). Because the least L-model of wcP12 is
                     h{dead , rmA}, {alive, execution, sig, rmB }i           (14)
     we obtain
                   ic(P7 , ∅, {¬sig, rmA}, {dead , ¬rmB })
                   −→ir ic(P11 , ∅, ¬sig, {dead , ¬rmB })
                   −→ia ic(P12 , ∅, ∅, {dead , ¬rmB }      −→it true.
   The revision leading to P11 is minimal in the sense that only the definition
   of rmA is revised and without this revision the condition of (6) cannot be
   explained. This is the only minimal revision as we will show in the sequel.
3. An alternative minimal revision could be the revision of P7 wrt to ¬sig:
              P13 = rev (P7 , ¬sig) = (P7 \ def (¬sig, P7 )) ∪ {sig ← ⊥}.
     The least L-model of wcP13 is
                           h{alive}, {sig, dead , rmA, rmB }i.               (15)
     Because this model maps rmA to false we obtain:
                 ic(P7 , ∅, {¬sig, rmA}, {dead , ¬rmB })
                 −→ir ic(P13 , ∅, rmA, {dead , ¬rmB }) −→iv vacuous.
4. So far the first step in evaluating the conditional was a revision step. Alter-
   natively, we could start with an abduction step. ¬sig can be explained in
   the abductive framework (8) by (9) leading to the program P8 and the least
   L-model (10). Because this model maps rmA to false we obtain:
                 ic(P7 , ∅, {¬sig, rmA}, {dead , ¬rmB })
                 −→ia ic(P8 , ∅, rmA, {dead , ¬rmB }) −→iv vacuous.
                                                                                 27




 5. Let us now reverse the order in which the conditions are treated and start
    by explaining rmA. This has already been done before and we obtain P9 and
    the least L-model (12). Because this model maps ¬sig to false we obtain:
                 ic(P7 , ∅, {¬sig, rmA}, {dead , ¬rmB })
                 −→ia ic(P9 , ∅, ¬sig, {dead , ¬rmB }) −→iv vacuous.
    In the last example we have discussed five different approaches to handle
the case that the truth value of the conditions of a conditional is unknown and
cannot be explained: maximal (parallel) revision (MaxRev), partial (sequen-
tial) revision as well as partial (sequential) explanation, where in the sequential
approaches the literals in the condition of the conditionals are treated in differ-
ent orders: left-to-right and right-to-left, where we consider sets to be ordered
(PRevLR,PRevRL,PExLR,PExRL). The results are summarized in Table 1,
where the conditional as well as the literals are evaluated wrt the final least
L-model computed in the different approaches.
    Which approach shall be preferred? Because rifleman A causally depends on
the captain’s signal but not vice-versa, plus given that in this example clauses
express causes, and effects come after causes, it would make sense to take the
cause ordering as the preferred one for abducing the conditions. Hence, PExLR
would be preferred. However, because rifleman A is an agent, the causes of his
actions can be internal to him, his decisions. Hence, when autonomous agents are
involved (or spontaneous phenomena like radioactivity), the ordering of abducing
the conditions is independent of causal dependency.


5   Properties
In this section, let P be a program, hI > , I ⊥ i the least L-model of wcP, IC a
set of integrity constraints, hP, AP , IC, |=lmwc
                                             L    i an abductive framework, and L
a literal.
Proposition 1. If O can be explained by E ⊆ AP and hJ > , J ⊥ i is the least
L-model of wc(P ∪ E), then hI > , I ⊥ i ⊆ hJ > , J ⊥ i.


                              MaxRev PRevRL PRevLR PExLR PExRL
               final program    P10    P12     P13     P8      P9
         final least L-model (13)      (14)   (15)    (10)    (12)
                           sig false  false   false   false   true
                         rmA true      true   false   false   true
                         dead   true   true   false   false   true
                         rmB false    false   false   false   true
                        alive false   false   true    true    false
                    execution unknown false unknown false     true
             conditional (6) true      true vacuous vacuous vacuous

             Table 1. Different approaches to evaluate conditional (6).
28




Proof. The least L-models hI > , I ⊥ i and hJ > , J ⊥ i are the least fixed points of
the semantic operators ΦP and ΦP∪E , respectively. Let hIn> , In⊥ i and hJn> , Jn⊥ i
be the interpretations obtained after applying ΦP and ΦP∪E n-times to h∅, ∅i,
respectively. We can show by induction on n that hIn> , In⊥ i ⊆ hJn> , Jn⊥ i. The
proposition follows immediately.
    Proposition 1 guarantees that whenever −→ia is applied, previously checked
conditions of a conditional need not to be re-checked. The following Proposition 3
gives the same guarantee whenever −→ir is applied.
Proposition 2. If the least L-model of wcP maps L to unknown and hJ > , J ⊥ i
is the least L-model of wc rev (P, L), then hI > , I ⊥ i ⊂ hJ > , J ⊥ i.
Proof. By induction on the number of applications of ΦP and Φrev (P,L) .
Proposition 3. RIC is terminating.
Proof. Each application of −→it , −→if , −→iu or −→iv leads to an irreducible
expression. Let cond (T , A) be the conditional to which RIC is applied. When-
ever −→ir is applied then the definition of at least one literal L occurring in T is
revised such that the least L-model of the weak completion of revised program
maps L to true. Because T does not contain a complementary pair of literals
this revised definition of L is never revised again. Hence, there cannot exist a
rewriting sequence with infinitely many occurrences of −→ir . Likewise, there
cannot exist a rewriting sequence with infinitely many occurrences of −→ia be-
cause each application of −→ia to a state ic(P, IC, T , A) reduces the number of
literals occurring in the T .
Proposition 4. RIC is not confluent.
Proof. This follows immediately from the examples presented in Section 4.

6    Open Questions and the Proposal of an Experiment
Open Questions The new approach gives rise to a number of questions. Which
of the approaches is preferable? This may be a question of pragmatics imputable
to the user. The default, because no pragmatic information has been added, is
maximals revision for skepticism and minimal revisions for credulity. Do humans
evaluate multiple conditions sequentially or in parallel? If multiple conditions
are evaluated sequentially, are they evaluated by some preferred order? Shall
explanations be computed skeptically or credulously? How can the approach be
extended to handle subjunctive conditionals?

The Proposal of an Experiment       Subjects are given the background informa-
tion specified in the program P9 . They are confronted with the conditionals
like (6) as well as variants with different consequences (e.g., execution instead
of {dead , ¬rmB } or conditionals where the order of two conditions are reversed.
We then ask the subjects to answer questions like: Does the conditional hold?
or Did the court order an execution? Depending on the answers we may learn
which approaches are preferred by humans.
                                                                                       29




Acknowledgements We thank Bob Kowalski for valuable comments on an ear-
lier draft of the paper.

References
 1. E.-A. Dietz, S. Hölldobler, and M. Ragni. A computational logic approach to
    the abstract and the social case of the selection task. In Proceedings Eleventh
    International Symposium on Logical Formalizations of Commonsense Reasoning,
    2013.
 2. E.-A. Dietz, S. Hölldobler, and M. Ragni. A computational logic approach to the
    suppression task. In N. Miyake, D. Peebles, and R. P. Cooper, editors, Proceedings
    of the 34th Annual Conference of the Cognitive Science Society, pages 1500–1505.
    Cognitive Science Society, 2012.
 3. E.-A. Dietz, S. Hölldobler, and C. Wernhard. Modelling the suppression task under
    weak completion and well-founded semantics. Journal of Applied Non-Classical
    Logics, 24:61–85, 2014.
 4. M. L. Ginsberg. Counterfactuals. Artificial Intelligence, 30(1):35–79, 1986.
 5. S. Hölldobler and C. D. P. Kencana Ramli. Contraction properties of a semantic
    operator for human reasoning. In Lei Li and K. K. Yen, editors, Proceedings of
    the Fifth International Conference on Information, pages 228–231. International
    Information Institute, 2009.
 6. S. Hölldobler and C. D. P. Kencana Ramli. Logic programs under three-valued
    Lukasiewicz’s semantics. In P. M. Hill and D. S. Warren, editors, Logic Pro-
    gramming, volume 5649 of Lecture Notes in Computer Science, pages 464–478.
    Springer-Verlag Berlin Heidelberg, 2009.
 7. S. Hölldobler and C. D. P. Kencana Ramli. Logics and networks for human rea-
    soning. In C. Alippi, Marios M. Polycarpou, Christos G. Panayiotou, and Georgios
    Ellinasetal, editors, Artificial Neural Networks – ICANN, volume 5769 of Lecture
    Notes in Computer Science, pages 85–94. Springer-Verlag Berlin Heidelberg, 2009.
 8. S. Hölldobler, T. Philipp, and C. Wernhard. An abductive model for hu-
    man reasoning.         In Proceedings Tenth International Symposium on Logi-
    cal Formalizations of Commonsense Reasoning, 2011.             commonsensereason-
    ing.org/2011/proceedings.html.
 9. A. C. Kakas, R. A. Kowalski, and F. Toni. Abductive Logic Programming. Journal
    of Logic and Computation, 2(6):719–770, 1993.
10. D. Lewis. Counterfactuals. Blackwell Publishers, Oxford, 1973.
11. J. W. Lloyd. Foundations of Logic Programming. Springer, Berlin, Heidelberg,
    1987.
12. J. Lukasiewicz. O logice trójwartościowej. Ruch Filozoficzny, 5:169–171, 1920.
    English translation: On Three-Valued Logic. In: Jan Lukasiewicz Selected Works.
    (L. Borkowski, ed.), North Holland, 87-88, 1990.
13. J. Pearl. Causality: Models, Reasoning, and Inference. Cambridge University Press,
    New York, USA, 2000.
14. L. M. Pereira and J. N. Aparı́cio. Relevant counterfactuals. In Proceedings 4th
    Portuguese Conference on Artificial Intelligence (EPIA), volume 390 of Lecture
    Notes in Computer Science, pages 107–118. Springer, 1989.
15. L. M. Pereira, E.-A. Dietz, and S. Hölldobler. An abductive reasoning approach to
    the belief-bias effect. In C. Baral, G. De Giacomo, and T. Eiter, editors, Principles
    of Knowledge Representation and Reasoning: Proceedings of the 14th International
    Conference, pages 653–656, Cambridge, MA, 2014. AAAI Press.
30




16. L. M. Pereira, E.-A. Dietz, and S. Hölldobler. Contextual abductive reasoning
    with side-effects. In I. Niemelä, editor, Theory and Practice of Logic Programming
    (TPLP), volume 14, pages 633–648, Cambridge, UK, 2014. Cambridge University
    Press.
17. F. Ramsey. The Foundations of Mathematics and Other Logical Essays. Harcourt,
    Brace and Company, 1931.
18. N. Rescher. Conditionals. MIT Press, Cambridge, MA, 2007.
19. K. Stenning and M. van Lambalgen. Human Reasoning and Cognitive Science.
    MIT Press, 2008.