=Paper=
{{Paper
|id=Vol-3209/7343
|storemode=property
|title=Argumentation-based Causal and Counterfactual Reasoning
|pdfUrl=https://ceur-ws.org/Vol-3209/7343.pdf
|volume=Vol-3209
|authors=Lars Bengel, Lydia Blümel, Tjitze Rienstra, Matthias Thimm
|dblpUrl=https://dblp.org/rec/conf/comma/BengelBRT22
}}
==Argumentation-based Causal and Counterfactual Reasoning==
Argumentation-based Causal and Counterfactual Reasoning Lars Bengel1 , Lydia Blümel1 , Tjitze Rienstra2 and Matthias Thimm1 1 Artificial Intelligence Group, FernUniversität in Hagen, Germany 2 Department of Advanced Computing Sciences, Maastricht University, The Netherlands Abstract In this paper we present a model for argumentative causal and counterfactual reasoning in a logical setting. Causal knowledge is represented in this system using Pearl’s causal model of a set of structural equations and a set of assumptions expressed in propositional logic. Queries concerning observations or actions can be answered by constructing an argumentation framework and determining its extensions. For counterfactual queries we propose an argumentation-based implementation of the twin network method and analyse its expressiveness. Keywords Causality, Counterfactuals, Abstract Argumentation 1. Introduction The area of explainable AI (XAI) has received increasing attention in recent years and many different approaches have been proposed (for an overview see e. g., [1, 2]). A central aspect of XAI is to provide human-understandable explanations for the output of some AI model. To provide these kinds of explanations, it seems only natural to utilise the techniques that humans employ to explain events and actions to each other. One such technique is formal argumentation [3]. In the past, many works have already suggested argumentation-based approaches for the purpose of providing explanations [4, 5]. More recently, we have seen different argumentative approaches to explain the output of various machine learning and other AI methods [6, 7]. Another field of research that is of increasing relevance is causality [8]. Especially interesting is how humans use causal connections for their explanations [9]. In the context of XAI, there are many advocates in the literature for using causal explanations [10, 11, 12]. Causality is furthermore closely linked with the notion of counterfactuals, which are statements that express what would have been true if something that is the case had been different. Counterfactual explanations have received much attention in recent research on XAI [13]. Relevant is also [14], in which the authors provide a logical representation of Pearl’s causal models in the framework of the causal calculus [15]. They show that the non-monotonic semantics of this causal calculus 1st International Workshop on Argumentation for eXplainable AI (ArgXAI, co-located with COMMA ’22), September 12, 2022, Cardiff, UK $ lars.bengel@fernuni-hagen.de (L. Bengel); lydia.bluemel@fernuni-hagen.de (L. Blümel); t.rienstra@maastrichtuniversity.nl (T. Rienstra); matthias.thimm@fernuni-hagen.de (M. Thimm) © 2022 Copyright for this paper by its authors. Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0). CEUR Workshop Proceedings http://ceur-ws.org ISSN 1613-0073 CEUR Workshop Proceedings (CEUR-WS.org) 1 Lars Bengel et al. CEUR Workshop Proceedings 1–12 corresponds directly to the solutions of causal models. In addition to that, this representation is also adequate for answering interventional queries for the causal model. In this paper we contribute to these lines of research by investigating an explainable, argumentation-based account of reasoning with causal models. The starting point in our approach is a propositional knowledge base that encodes a causal model in the sense of Pearl [8] restricted to boolean variables and default assumptions about background variables. The ap- proach makes it possible to answer interventional and counterfactual queries by constructing an argumentation framework whose stable extensions provide the answer to the queries. Our method of answering counterfactual queries is a reformulation of the dual approach due to Pearl, where a counterfactual query can be answered equivalently using two different methods: a three-step abduction-action-prediction procedure, and the so called twin network method [8]. We prove that, in our setting, the two methods are also equivalent. Our work can be understood as an initial investigation into the possibility of causal reasoning within Dung’s model of argumentation. To summarise, the contributions of this paper are the following. 1. We introduce the notion of causal knowledge bases (Section 3). 2. We provide a novel approach for argumentation-based explanations for counterfactual reasoning implementing the twin network method (Section 4). 3. We prove the standard three-step procedure and the twin network method for evaluating counterfactuals by Pearl are equivalent under our notion (Section 4). 4. We discuss our approach and related work (Section 5). In Section 2 we present the necessary background on knowledge bases and formal argumen- tation and Section 6 concludes the paper. 2. Preliminaries The causal reasoning framework we develop builds on a simple and well-known form of default reasoning based on maximal consistent subsets [16]. Let ℒ be a set of propositional formulas generated by a finite set of atoms. We assume there is a set 𝐾 ⊆ ℒ of facts and a set 𝐴 ⊆ ℒ of assumptions. We call a pair Δ = (𝐾, 𝐴) a knowledge base. Facts are true and thus we assume that 𝐾 is consistent. Assumptions are statements that we are willing to assume true unless we have evidence to the contrary. The consequences of a knowledge base Δ = (𝐾, 𝐴) are determined by 𝐾 together with the maximal 𝐾-consistent subsets of 𝐴. Definition 1. Let Δ = (𝐾, 𝐴) be a knowledge base and 𝜑, 𝜓 ∈ ℒ. A set Σ ⊆ 𝐴 is a maximal 𝐾-consistent subset of 𝐴 whenever Σ ∪ 𝐾 is consistent and Σ′ ∪ 𝐾 is inconsistent for all Σ′ ⊆ 𝐴 such that Σ ⊂ Σ′ . We say that: • Δ entails 𝜓 (written Δ |∼ 𝜓) whenever Σ ∪ 𝐾 ⊢ 𝜓 for every maximal 𝐾-consistent subset of 𝐴. • 𝜑 Δ-entails 𝜓 (written 𝜑 |∼Δ 𝜓) whenever (𝐾 ∪ {𝜑}, 𝐴) entails 𝜓. 2 Lars Bengel et al. CEUR Workshop Proceedings 1–12 In the next section we define a special class of knowledge bases used to represent causal knowledge. Our goal is to provide an argumentative method for causal and counterfactual reasoning based on such knowledge bases. The argumentative aspect of this method relies on a characterisation of reasoning with maximal consistent subsets based on the notion of argumentation framework or AF, for short [3]. Definition 2. An argumentation framework is a pair 𝐹 = (A, ⇒) where 𝐴 is a set whose elements are called arguments and where ⇒ ⊆ 𝐴 × 𝐴 is called the attack relation. We mostly use infix notation for attacks and write 𝑎 ⇒ 𝑏 for (𝑎, 𝑏) ∈ ⇒ and 𝑎 ̸ ⇒𝑏 for (𝑎, 𝑏) ̸∈ ⇒. If 𝑎 ⇒ 𝑏 we also say that 𝑎 attacks 𝑏. Given an AF, a semantics determines sets of jointly acceptable arguments called extensions. The admissible, preferred and stable semantics are defined as follows [3]. Definition 3. Let 𝐹 = (A, ⇒) be an AF. A set 𝐸 ⊆ A is: • conflict-free if for all 𝑎, 𝑏 ∈ 𝐸 we have 𝑎 ̸⇒ 𝑏. • self-defending if for all 𝑎 ∈ 𝐸 and 𝑏 ∈ A ∖ 𝐸 such that 𝑏 ⇒ 𝑎, there is a 𝑐 ∈ 𝐸 such that 𝑐 ⇒ 𝑏. • admissible if 𝐸 is conflict-free and self-defending. • preferred if 𝐸 is admissible and there is no admissible set 𝐸 ′ ⊆ A such that 𝐸 ⊂ 𝐸 ′ . • stable if 𝐸 is conflict-free and for every 𝑎 ∈ A ∖ 𝐸 there is a 𝑏 ∈ 𝐸 such that 𝑏 ⇒ 𝑎. Following Cayrol [17] we define an argument induced by a knowledge base Δ = (𝐾, 𝐴) to be a pair (Φ, 𝜓) where Φ ⊆ 𝐴 is a minimal set of assumptions (called the premises of the argument) that, together with 𝐾, consistently entails 𝜓 (called the conclusion of the argument). One argument undercuts another if the conclusion of the former is the negation of a premise of the latter. The set of all arguments induced by Δ, together with undercut as the attack relation, forms the AF induced by Δ 1 . Formally: Definition 4. Let Δ = (𝐾, 𝐴) be a knowledge base. • An argument induced by Δ is a pair (Φ, 𝜓) such that – Φ ⊆ 𝐴, – Φ ∪ 𝐾 ⊬ ⊥, – Φ ∪ 𝐾 ⊢ 𝜓, and if Ψ ⊂ Φ then Ψ ∪ 𝐾 ⊬ 𝜓. • An argument (Φ, 𝜓) undercuts an argument (Φ′ , 𝜓 ′ ) if for some 𝜑′ ∈ Φ′ we have 𝜑′ ≡ ¬𝜓. • The AF induced by Δ is the AF (A, ⇒) where A consists of all arguments induced by Δ and 𝑎 ⇒ 𝑏 holds whenever 𝑎 undercuts 𝑏. We denote the AF induced by Δ by 𝐹 (Δ). Cayrol et al. [17] showed that there is a one-to-one correspondence between maximal 𝐾- consistent subsets of a knowledge base and the stable extensions of the induced AF. Thus, given a knowledge base Δ = (𝐾, 𝐴), the question of whether 𝜑 Δ-entails 𝜓 can be determined by constructing the AF induced by (𝐾 ∪ {𝜑}, 𝐴) and checking whether every stable extension contains an argument with conclusion 𝜓. 1 Note that this construction method allows arguments to have syntactically different, but semantically equivalent conclusions. 3 Lars Bengel et al. CEUR Workshop Proceedings 1–12 Proposition 1. Let Δ = (𝐾, 𝐴) be a knowledge base. Then 𝜑 |∼Δ 𝜓 if and only if every stable extension 𝐸 of 𝐹 (𝐾 ∪ {𝜑}, 𝐴) contains an argument with conclusion 𝜓. We conclude this section by noting that the form of argumentation we consider here can also be captured by structured argumentation formalisms such as ASPIC and ABA [18]. 3. Causal Knowledge Bases Our definition of a causal model is essentially that of Pearl [8] except that we restrict our attention to Boolean-valued variables. A causal model consists of a set 𝑈 of background atoms, a set 𝑉 of explainable atoms, and a set 𝐾 of formulas which we call Boolean structural equations (this terminology was adopted from [14]). Background atoms represent variables that are determined outside of the model. They are typically unobservable and uncontrollable. Every explainable atom 𝑣 is functionally dependent on other atoms of the model. This dependency is specified by a Boolean structural equation of the form 𝑣 ↔ 𝜑. Intuitively, this equation represents the causal mechanism by which 𝑣 is determined by the other atoms in the model. We use bi-implication because the represented causal mechanism determines not only when 𝑣 is true, but also when 𝑣 is false. Definition 5. A causal model is a triple (𝑈, 𝑉, 𝐾) where 𝑈 and 𝑉 partition the set of atoms into, respectively, a set of background and explainable atoms. 𝐾 consists of a set of Boolean structural equations, one for each atom 𝑣 ∈ 𝑉 . A Boolean structural equation for 𝑣 is a formula of the form 𝑣↔𝜑 where 𝜑 is a propositional formula that does not contain 𝑣. Given a causal model (𝑈, 𝑉, 𝐾) we can reconstruct 𝑈 and 𝑉 from 𝐾 as follows; (1) for every formula 𝑣 ↔ 𝜑 in 𝐾, add 𝑣 to 𝑉 , and (2) add all the remaining atoms that appear in 𝐾 but not in 𝑉 to 𝑈 . Therefore we can refer to a causal model as 𝐾. For every Boolean structural equation 𝑣 ↔ 𝜑, we call an atom appearing in 𝜑 a parent (background or explainable) of 𝑣. Like in the standard case, a causal model induces a causal graph 𝐺 whose vertices are the explainable atoms of the model [8]. This graph contains an edge from atom 𝑣 to atom 𝑣 ′ whenever 𝑣 is an explainable parent of 𝑣 ′ . Example 1. Let 𝑈 = {corona, influenza, at-risk} and 𝑉 = {covid, flu, short-of-breath, fever, chills}. The set 𝐾 consisting of the following equations represents a causal model for a patient diagnosis scenario. covid ↔ corona flu ↔ influenza fever ↔ covid ∨ flu chills ↔ fever short-of-breath ↔ covid ∧ at-risk 4 Lars Bengel et al. CEUR Workshop Proceedings 1–12 corona influenza at-risk covid flu short-of-breath fever chills Figure 1: Causal graph for Example 1. In words: corona virus causes covid, influenza virus causes flu, covid and flu cause fever, fever causes chills, and covid causes short-of-breath, but only if the patient is at risk for this condi- tion (represented by the background atom at-risk). Note that corona, influenza and at-risk are background atoms and thus assumed unobservable. Figure 1 depicts the causal graph for this model. This figure also includes the background atoms of the model, drawn using dotted lines. We define a causal knowledge base to be a knowledge base where the set of facts is a causal model and where the set of assumptions is restricted to contain assumptions about background atoms. Here we depart from the probabilistic approach to causal modelling, where belief about background atoms is represented using probability distributions over their values [8]. Definition 6. A causal knowledge base is a knowledge base Δ = (𝐾, 𝐴) where 𝐾 is a causal model and where 𝐴 is a set of background assumptions, at least one for each background atom. A background assumption for an atom 𝑢 is a literal 𝑙 ∈ {𝑢, ¬𝑢}. By restricting background assumptions to be literals we prevent the ability to express depen- dencies among background atoms. There are three possible attitudes towards a background atom 𝑢, since we can assume just 𝑢, just ¬𝑢, or both. Assuming just 𝑢 or ¬𝑢 amounts to assuming that 𝑢 is true or false, unless we have evidence to the contrary. Assuming both 𝑢 and ¬𝑢 represents a state of uncertainty where 𝑢 may, depending on the evidence, be true as well as false. If Δ = (𝐾, 𝐴) is a causal knowledge base then Δ-entailment represents a relation between observations and predictions. These predictions include causes as well as effects of the observation in accordance with the causal model 𝐾 and background assumptions 𝐴. Example 2. Consider the causal knowledge base Δ = (𝐾, 𝐴) where 𝐾 is the causal model defined in Example 1 and where 𝐴 = {at-risk, ¬at-risk, ¬corona, ¬influenza}. In words, these assumptions mean that we are uncertain whether the patient is at risk to develop shortness of breath, and a corona or influenza infection is assumed false unless there is evidence to the contrary. Now consider the question of whether observing fever entails shortness of breath, i.e., whether fever |∼Δ short-of-breath. (1) 5 Lars Bengel et al. CEUR Workshop Proceedings 1–12 Flu and covid cause fever. Thus, fever is evidence for flu or covid. Of these two possible causes, covid may cause shortness of breath, but only if the patient is at risk, which may or may not be true. Hence, fever may or may not be evidence for shortness of breath. This reasoning is depicted by the AF 𝐹 = 𝐹 ((𝐾 ∪ {fever}, 𝐴)) shown in Figure 2 (we only depict arguments relevant to the conclusion of short-of-breath). This AF has four stable extensions {𝑎2 , 𝑎4 }, {𝑎3 , 𝑎4 }, {𝑎2 , 𝑎5 }, and {𝑎3 , 𝑎5 , 𝑎1 }. The argument 𝑎1 with conclusion short-of-breath is included in some but not all of these extensions. We thus have that (1) is false, but note that fever |∼Δ ¬short-of-breath is also false. Thus, given fever, shortness of breath is possible but not necessary. 𝑎1 : ({at-risk, ¬influenza}, short-of-breath) 𝑎2 : ({¬at-risk}, ¬at-risk) 𝑎4 : ({¬corona}, influenza) 𝑎3 : ({at-risk}, at-risk) 𝑎5 : ({¬influenza}, corona) Figure 2: The AF 𝐹 (𝐾 ∪ {fever}, 𝐴). 4. Counterfactuals In this section we show how causal knowledge bases provide a means to evaluate counterfactuals. Pearl introduces two methods for evaluating counterfactuals in causal models [8], a three- step procedure and the twin network method. We now demonstrate these two processes are equivalent under our notion of causal knowledge bases. Before dealing with counterfactual statements we start with interventional statements of the form if 𝑣 would be 𝑥 then 𝜓 would be true (2) where 𝑣 is a variable we intervene on and 𝑥 a truth value (we use ⊤ for true and ⊥ for false). The antecedent of this statement represents the action of setting 𝑣 to 𝑥. In our running example, a patient might be given Ibuprofen to treat fever. This is an action that amounts to setting fever to ⊥. Note that this is different from observing ¬fever. While the observation ¬fever would, in our model, imply ¬covid and ¬flu, the action of setting fever to ⊥ amounts to overriding the causal mechanism that determines fever, which does not affect covid or flu. Like Pearl, given a causal model 𝐾, we denote by 𝐾[𝑣=𝑥] the new causal model where the equation 𝑣 ↔ 𝜑 is replaced with 𝑣 ↔ 𝑥. Definition 7. Let 𝐾 be a causal model, let 𝑣 be an explainable atom, and let 𝑥 ∈ {⊤, ⊥}. We denote by 𝐾[𝑣=𝑥] the causal model defined by 𝐾[𝑣=𝑥] = {(𝑣 ′ ↔ 𝜑) ∈ 𝐾 | 𝑣 ′ ̸= 𝑣} ∪ {(𝑣 ↔ 𝑥)}. 6 Lars Bengel et al. CEUR Workshop Proceedings 1–12 Example 3. Let Δ = (𝐾, 𝐴) be the causal knowledge base from Example 2. Since short-of-breath is caused by covid, covid causes fever, and fever causes chills, we have that shortness of breath leads to the prediction of chills short-of-breath |∼Δ chills. (3) But what if we observe short-of-breath in a patient that has been given Ibuprofen? The answer is given by the causal knowledge base Δ′ = (𝐾[fever=⊥] , 𝐴). Here, the action of setting fever to ⊥ blocks the effect of covid on chills. We now have that shortness of breath leads to the prediction of no chills short-of-breath |∼Δ′ ¬chills. (4) We now turn to counterfactual statements. These are statements of the form given 𝜑, if 𝑣 had been 𝑥 then 𝜓 would be true (5) where 𝜑 is an observed piece of evidence and 𝜓 is the counterfactual conclusion. An example of a counterfactual in our running example would be: given that the patient has fever, would the patient have had fever if we had administered a covid vaccine? This amounts to the statement given fever, if covid had been ⊥ then fever would be true. The standard causal modelling approach of evaluating a counterfactual statement of the form (5) is based on a three-step procedure [8] that we adapt for our setting as follows. Definition 8. Given a causal knowledge base Δ = (𝐾, 𝐴), the truth of a counterfactual state- ment (5) is determined by: • Step 1 (abduction) Determine the maximal 𝐾 ∪ {𝜑}-consistent subsets Σ1 , . . . , Σ𝑛 of 𝐴. These sets represent the possible configurations of the background atoms consistent with evidence 𝜑. • Step 2 (action) Set 𝑣 to 𝑥 by updating the causal model 𝐾 to 𝐾[𝑣=𝑥] . • Step 3 (prediction) For each Σ𝑖 obtained in step 1, determine whether Σ𝑖 ∪ 𝐾[𝑣=𝑥] ⊢ 𝜓. If the answer is yes for all 𝑖 ∈ 0, . . . 𝑛, then the counterfactual statement (5) is true. The difficulty with this procedure is that it requires the computation of the maximal 𝐾 ∪ {𝜑}- consistent subsets of 𝐴. The same difficulty arises in probabilistic causal models, where the abduction step requires the computation of a probability distribution over configurations of the background atoms. The twin network method overcomes this difficulty [8]. It is based on constructing a so called twin model whose causal network consists of two parts, one to represent the actual world, and one to represent the counterfactual world. While the two networks share the same background variables, the counterfactual part consists of a “counterfactual copy” of all the explainable variables. In what follows we use 𝑣 * to denote a unique new atom representing the counterfactual copy of 𝑣, and use 𝜑* to denote the formula 𝜑 with each occurrence of an explainable atom 𝑣 replaced with 𝑣 * . Definition 9. The twin model for a causal model 𝐾 is the causal model 𝐾 * defined by 𝐾 * = 𝐾 ∪ {(𝑣 * ↔ 𝜑* ) | (𝑣 ↔ 𝜑) ∈ 𝐾} . 7 Lars Bengel et al. CEUR Workshop Proceedings 1–12 While the abduction step takes place in the actual network, the action and prediction steps take place in the counterfactual network. Let us now prove this intuition of separated areas in the twin model holds up formally during the evaluation, i. e., the three step procedure is equivalent to the evaluation of the twin model. Proposition 2. Let Δ = (𝐾, 𝐴) be a causal knowledge base. The counterfactual statement given 𝜑, if 𝑣 had been 𝑥 then 𝜓 would be true is true in Δ if and only if 𝜑 |∼Δ* * 𝜓*. [𝑣 =𝑥] Proof. We prove that the twin model method is equivalent to the three-step procedure for evaluating counterfactuals from Pearl as given in Definition 8. Let Δ = (𝐾, 𝐴) be a causal knowledge base. First, we recall that 𝜑 |∼Δ 𝜓 if and only if (𝐾 ∪ {𝜑}, 𝐴) entails 𝜓. (𝐾 ∪ {𝜑}, 𝐴) entails 𝜓 whenever for every maximal K-consistent subset Σ of 𝐴, we have that Σ ∪ 𝐾 ⊢ 𝜓. (⇐) Now, for the given counterfactual statement, consider Δ* = (𝐾 * , 𝐴) with 𝐾 * as given in Definition 9. Furthermore, we modify Δ* by setting 𝑣 * = 𝑥 in 𝐾 * , i.e., we have Δ*[𝑣* =𝑥] . That means, 𝐾 * fully contains 𝐾 as well as an exact copy of 𝐾 (each copy of a formula 𝜑 of 𝐾 is denoted with 𝜑* ). In this copy, we also have set 𝑣 * = 𝑥. Therefore, we can essentially also consider 𝐾 * as the union of 𝐾 and 𝐾[𝑣=𝑥] . With the twin model method, we have that 𝜑 |∼Δ* * 𝜓 * . That means, according to [𝑣 =𝑥] Definition 1 we determine the 𝐾 * ∪{𝜑}-consistent subsets Σ*1 , ..., Σ*𝑛 of 𝐴. Now, since 𝐾 ⊂ 𝐾 * , it follows that Σ*1 ∩ 𝐾, ..., Σ*𝑛 ∩ 𝐾 are maximal consistent with 𝐾 ∪ {𝜑} (as required by step 1 abduction). The action step of Definition 8 is also executed in the twin model by setting 𝑣 * = 𝑥 in 𝐾 * . In the prediction step, we have to determine whether Σ𝑖 ∪ 𝐾[𝑣=𝑥] ⊢ 𝜓 for all 𝐾-consistent Σ𝑖 . In the twin model method, we consider the 𝐾 * ∪ {𝜑}-consistent subsets. As already explained above, 𝐾[𝑣=𝑥] is fully contained in 𝐾 * and therefore, similarly to above, if we have that Σ*𝑖 ∪ 𝐾 * ⊢ 𝜓 * it follows that Σ*𝑖 ∪ 𝐾[𝑣=𝑥] ⊢ 𝜓. Thus, the twin model method is equivalent to the three-step procedure for evaluating counterfactual statements. (⇒) Suppose the counterfactual statement is true according to the three step procedure. In the twin model it holds that 𝐾 and {(𝑣 * ↔ 𝜑* ) | (𝑣 ↔ 𝜑) ∈ 𝐾} have no variables in common. Thus, for any Σ maximal consistent wrt. 𝐾 * ∪{𝜑} it holds that Σ′ = Σ∩{(𝑣 * ↔ 𝜑* ) | (𝑣 ↔ 𝜑) ∈ 𝐾} is maximal consistent wrt. 𝐾[𝑣=𝑥] ∪ {𝜑}. By the assumption we have Σ′ ∪ 𝐾[𝑣=𝑥] ∪ {𝜑} ⊢ 𝜓 for any such Σ′ . By monotony of classic entailment it follows that Σ ∪ 𝐾[𝑣 * * =𝑥] ∪ {𝜑} ⊢ 𝜓. We have now proven that the standard three-step procedure can be reduced to a single evaluation step on the twin model. To conclude, let us give some examples of counterfactual reasoning with the AFs build from a twin model. Example 4. Let Δ = (𝐾, 𝐴) be the causal knowledge base defined in Example 2. Suppose we have evidence that the patient has fever. Would the patient have had fever if we had administered a covid vaccine (i.e., if covid had been false)? This depends on whether the fever is caused by covid or flu. If fever was caused by covid then the vaccine would indeed have prevented fever. If, on the 8 Lars Bengel et al. CEUR Workshop Proceedings 1–12 corona influenza at-risk covid flu covid* flu* short-of-breath fever short-of-breath* fever* chills chills* Figure 3: Twin causal diagram for Example 4. other hand, fever was caused by flu, then the vaccine would not have had an effect, and the patient would still have had fever. To evaluate this counterfactual statement we check whether fever |∼Δ* fever* . (6) [covid* =⊥] Figure 3 shows the causal graph of the twin model 𝐾covid * * =⊥ and Figure 4 shows the AF 𝐹 = * (𝐾[covid* =⊥] ∪ {fever}, 𝐴). This AF has two stable extensions {𝑎1 , 𝑎3 } and {𝑎2 }. While the first entails fever, the second does not. Thus, (6) is false, but note that fever |∼Δ* ¬fever* (7) [covid* =⊥] is also false. Thus, given that the patient has fever, the patient may or may not have had fever, had we administered a covid vaccine. 𝑎1 : ({¬corona}, fever* ) 𝑎2 : ({¬influenza}, corona) 𝑎3 : ({¬corona}, influenza) * Figure 4: The AF 𝐹 ((𝐾[covid* =⊥] ∪ {fever}, 𝐴)). Example 5. Suppose that, in addition to having evidence that the patient has fever, we also have evidence of shortness of breath. Again, we want to know if the patient would have had fever if we had administered a covid vaccine. This time, the evidence of shortness of breath implies that the fever must have been caused by covid. Hence, the vaccine would indeed have prevented fever. We check this formally by checking whether fever ∧ short-of-breath |∼Δ* fever* . (8) [covid* =⊥] * Figure 5 shows the AF 𝐹 = (𝐾covid * =⊥ ∪ {fever ∧ short-of-breath}, 𝐴). This AF is the same as the one shown in Figure 4 except that there is an extra argument (∅, corona) which is based on the evidence short-of-breath. This AF has one stable extension {𝑎4 , 𝑎2 } which does not entail fever* . Hence, (8) is true, i.e., given that the patient has fever and shortness of breath, the patient would not have had fever, had we administered a covid vaccine. 9 Lars Bengel et al. CEUR Workshop Proceedings 1–12 𝑎1 : ({¬corona}, fever* ) 𝑎2 : ({¬influenza}, corona) 𝑎4 : (∅, corona) 𝑎3 : ({¬corona}, influenza) * Figure 5: The AF 𝐹 (𝐾covid* =⊥ ∪ {fever ∧ short-of-breath}, 𝐴). 5. Discussion Our work can be understood as an initial investigation into the possibility of reasoning with causal models using tools of abstract information science and logic. By doing this we contribute to explaining technically advanced causal reasoning procedures with concepts closer to human reasoning. Our account of Pearl’s theory was motivated by the non-monotonic interpretation of his causal model in [14]. The need to further explain causal reasoning, e. g., using Bayesian networks, has also been pointed out in [19]. Their contribution towards this is making d- separation explicit in so called support graphs, which eliminate circular causal structures and help to explain interdependent causes. Implementing a similar mechanism within our framework has proven nontrivial. In [20] they illustrate the power of their framework by applying it to legal reasoning. As the authors point out under future work, the support graph is more of a visualization tool while the actual evaluation of arguments is still done in the original Bayesian network formalism. A sophisticated approach towards a framework capable of intrinsic evaluations of causal statements is the method for generating bipolar argumentation frameworks from causal models by Rago et al. [21], which is also based on Pearl’s notion of a causal model. For that they create so called explanation moulds, that reinterpret desirable properties of semantics of argumentation frameworks. Rago et. al. point out that, compared to [22], they make better use of the AF representation in their causal reasoning by basing their semantic evaluation on the support and attack relations themselves instead of relying on an additional ontological structure. This criticism does not apply to our framework, we only rely on the causal model. Once the AF is constructed, we do not need the inference apparatus of the underlying knowledge base formalism for causal reasoning in our framework. There are several major design differences between the two approaches. First Rago et al. use bipolar AFs in contrast to our Dung-style approach. Second, our approach is based on structured argumentation, while theirs uses causal atoms as arguments. On the evaluation level, they apply gradual semantics while we make use of classic stable semantics. For both frameworks an interesting future work direction would be to apply the resp. other type of semantics. An important difference lies in the resp. attack notions. Rago et al. let causes contribute negatively or positively towards the status of an argument via the attack and support relation, respectively. Our focus is on the explanation aspect. Attacks are used to exclude certain premises as causes. This is done by undercut-attacks on the corresponding argument. A current shortcoming of our approach is the restriction to propositional logic in the attack notion, which would have difficulties handling uncertain causal relations. A promising future work direction would therefore be to investigate ways for representing uncertain causal relations in some type of probabilistic argumentation framework, e. g. by adding a probability distribution 10 Lars Bengel et al. CEUR Workshop Proceedings 1–12 over possible causal structures [23]. Another option could be to utilize a form of probabilistic instantiation like PABA [24]. 6. Conclusion We have proposed a natural way to model causal reasoning with Dung-style AFs by means of causal knowledge bases. The framework introduced integrates given causal knowledge in form of structural equations into an attack relation and a set of structured arguments constructed from sets of assumptions resp. observations from the real world. By doing this we have laid the groundwork for an implementation of the twin network method for counterfactual reasoning. It offers an alternative for computing the truth value of counterfactual statements which, as we proved, is equivalent to the standard approach in our framework. The results we obtain demonstrate that causal reasoning can be done as an instantiation of Dung’s model of argumentation. This can be used to provide argumentative explanations for causal and counterfactual queries. That is, the argumentation framework contains not only the arguments used to arrive at a prediction, but also possible counterarguments that lead to alternative predictions. Acknowledgements The research reported here was partially supported by the Deutsche Forschungsgemeinschaft (grant 375588274). References [1] A. Adadi, M. Berrada, Peeking inside the black-box: a survey on explainable artificial intelligence (XAI), IEEE access 6 (2018) 52138–52160. [2] R. Guidotti, A. Monreale, S. Ruggieri, F. Turini, F. Giannotti, D. Pedreschi, A survey of methods for explaining black box models, ACM computing surveys (CSUR) 51 (2018) 1–42. [3] P. M. Dung, On the acceptability of arguments and its fundamental role in nonmonotonic reasoning, logic programming and n-person games, Artificial Intelligence 77 (1995) 321– 358. [4] C. Antaki, I. Leudar, Explaining in conversation: Towards an argument model, European Journal of Social Psychology 22 (1992) 181–194. [5] B. Moulin, H. Irandoust, M. Bélanger, G. Desbordes, Explanation and argumentation capabilities: Towards the creation of more persuasive agents, Artificial Intelligence Review 17 (2002) 169–222. [6] K. Čyras, A. Rago, E. Albini, P. Baroni, F. Toni, Argumentative xai: a survey, arXiv preprint arXiv:2105.11266 (2021). [7] A. Vassiliades, N. Bassiliades, T. Patkos, Argumentation and explainable artificial intelli- gence: a survey, The Knowledge Engineering Review 36 (2021) e5. [8] J. Pearl, Causality: models, reasoning and inference, volume 29, Cambridge University Press, 2000. [9] M. M. De Graaf, B. F. Malle, How people explain action (and autonomous intelligent systems should too), in: 2017 AAAI Fall Symposium Series, 2017, pp. 19–26. 11 Lars Bengel et al. CEUR Workshop Proceedings 1–12 [10] T. Miller, Explanation in artificial intelligence: Insights from the social sciences, Artificial intelligence 267 (2019) 1–38. [11] P. Schwab, W. Karlen, Cxplain: Causal explanations for model interpretation under uncertainty, Advances in Neural Information Processing Systems 32 (2019) 10188–10198. [12] D. Alvarez-Melis, T. S. Jaakkola, A causal framework for explaining the predictions of black-box sequence-to-sequence models, arXiv preprint arXiv:1707.01943 (2017). [13] R. M. J. Byrne, Counterfactuals in explainable artificial intelligence (XAI): evidence from human reasoning, in: S. Kraus (Ed.), Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence, IJCAI 2019, Macao, China, August 10-16, 2019, ijcai.org, 2019, pp. 6276–6282. doi:10.24963/ijcai.2019/876. [14] A. Bochman, V. Lifschitz, Pearl’s causality in a logical setting, in: B. Bonet, S. Koenig (Eds.), Proceedings of the Twenty-Ninth AAAI Conference on Artificial Intelligence, January 25-30, 2015, Austin, Texas, USA, AAAI Press, 2015, pp. 1446–1452. URL: http://www.aaai. org/ocs/index.php/AAAI/AAAI15/paper/view/9686. [15] N. McCain, H. Turner, et al., Causal theories of action and change, in: AAAI/IAAI, Citeseer, 1997, pp. 460–465. [16] N. R. R. Manor, N. Rescher, On inference from inconsistent premises, Theory and Decision 1 (1970) 179–219. [17] C. Cayrol, On the relation between argumentation and non-monotonic coherence-based entailment, in: Proceedings of the Fourteenth International Joint Conference on Artificial Intelligence, IJCAI 95, Montréal Québec, Canada, August 20-25 1995, 2 Volumes, Morgan Kaufmann, 1995, pp. 1443–1448. [18] O. Arieli, A. Borg, J. Heyninck, C. Straßer, Logic-based approaches to formal argumentation, FLAP 8 (2021) 1793–1898. URL: https://collegepublications.co.uk/ifcolog/?00048. [19] S. T. Timmer, J.-J. C. Meyer, H. Prakken, S. Renooij, B. Verheij, Explaining bayesian networks using argumentation, in: European Conference on Symbolic and Quantitative Approaches to Reasoning and Uncertainty, Springer, 2015, pp. 83–92. [20] S. T. Timmer, J.-J. C. Meyer, H. Prakken, S. Renooij, B. Verheij, A structure-guided approach to capturing bayesian reasoning about legal evidence in argumentation, in: Proceedings of the 15th International Conference on Artificial Intelligence and Law, 2015, pp. 109–118. [21] A. Rago, F. Russo, E. Albini, P. Baroni, F. Toni, Forging argumentative explanations from causal models, in: M. D’Agostino, F. A. D’Asaro, C. Larese (Eds.), Proceedings of the 5th Workshop on Advances in Argumentation in Artificial Intelligence 2021 co-located with the 20th International Conference of the Italian Association for Artificial Intelligence (AIxIA 2021), Milan, Italy, November 29th, 2021, volume 3086 of CEUR Workshop Proceedings, CEUR-WS.org, 2021. URL: http://ceur-ws.org/Vol-3086/paper3.pdf. [22] P. Besnard, M.-O. Cordier, Y. Moinard, Arguments using ontological and causal knowledge, in: International Symposium on Foundations of Information and Knowledge Systems, Springer, 2014, pp. 79–96. [23] A. Hunter, S. Polberg, N. Potyka, T. Rienstra, M. Thimm, Probabilistic argumentation: A survey, Handbook of Formal Argumentation 2 (2021) 397–441. [24] K. Čyras, Q. Heinrich, F. Toni, Computational complexity of flat and generic assumption- based argumentation, with and without probabilities, Artificial Intelligence 293 (2021) 103449. 12