=Paper=
{{Paper
|id=Vol-2742/short3
|storemode=property
|title=Considerations for Applying Logical Reasoning to Explain Neural Network Outputs
|pdfUrl=https://ceur-ws.org/Vol-2742/short3.pdf
|volume=Vol-2742
|authors=Federico Maria Cau,Lucio Davide Spano,Nava Tintarev
|dblpUrl=https://dblp.org/rec/conf/aiia/CauST20
}}
==Considerations for Applying Logical Reasoning to Explain Neural Network Outputs==
Considerations for Applying Logical Reasoning to Explain Neural Network Outputs Federico Maria Cau1 , Lucio Davide Spano1 , and Nava Tintarev2 1 University of Cagliari, Department of Mathematics and Computer Science, Cagliari, Italy federicom.cau@unica.it, davide.spano@unica.it 2 Maastricht University, DKE, Maastricht, the Netherlands n.tintarev@maastrichtuniversity.nl Abstract. We discuss the impact of presenting explanations to people for Artificial Intelligence (AI) decisions powered by Neural Networks, according to three types of logical reasoning (inductive, deductive, and abductive). We start from examples in the existing literature on ex- plaining artificial neural networks. We see that abductive reasoning is (unintentionally) the most commonly used as default in user testing for comparing the quality of explanation techniques. We discuss whether this may be because this reasoning type balances the technical challenges of generating the explanations, and the effectiveness of the explanations. Also, by illustrating how the original (abductive) explanation can be converted into the remaining two reasoning types we are able to identify considerations needed to support these kinds of transformations. Keywords: Explainable User Interfaces · XAI · Reasoning. 1 Introduction In the last decade, eXplainable AI (XAI) research has made great advances, introducing new explanation techniques like Grad-CAM [18], SHAP [13], LRP [1], LIME [17], LORE [10], RETAIN [6], DeepLIFT [19], a.o. However, much of this previous work does not consider which kind of logical reasoning is presented to users, or how this interacts with characteristics of task, much less individual differences between users. Given the possible implication of the reasoning type in the effectiveness of the XAI system, we inspect the transformation from one kind of reasoning to another. We do this intending to draw possible considerations for selecting between the types of reasoning, starting with neural networks models. While the categorization of the reasoning types applies to explanations for other probabilistic models, the available explanation techniques differ. The paper is organised as follows: in Section 2 we provide some background on XAI taxonomies, evaluation tasks and reasoning types, in Section 3 we inves- tigate the reasoning types, in Section 4 we establish the guidelines for reasoning transformation, in Section 5 we conclude the paper and we discuss possible ideas for future work. Copyright c 2020 for this paper by its authors. Use permitted under Creative Com- mons License Attribution 4.0 International (CC BY 4.0). 2 Related Work In this section, we explore three topics: the first concerns the existing XAI tax- onomies, to catalogue existing state-of-the-art techniques that explain neural networks. After that, we will talk about the task types with which explanations techniques are proposed and their importance for evaluation by humans. Finally, we define the reasoning types and their relationship with XAI explanations. 2.1 Explanation Taxonomies There have been articles that have categorized explanation methods for neural networks. Among them, [24, 3, 12, 20] were very useful to lay the foundations of our research: they describe a comprehensive taxonomy of interpretability meth- ods regarding Deep Neural Networks (DNN), including goals, properties and ar- chitecture, together with guiding principles for their safety and trustworthiness. Furthermore, other surveys go beyond the analysis of neural network models, and help us to expand the knowledge on methods of explanation and models used [2, 11]. However, there are two surveys that also focus on the impact of ex- planations on users [14, 22]. The former supplies a categorization between design goals for interpretable algorithms considering different XAI user groups. The latter introduces a conceptual framework that explains how human reasoning processes informs XAI techniques, which we will deepen in the next sections. 2.2 Types of Tasks In addition to determining which XAI which techniques to use, another key step is to identify what type of task the user will accomplish. We started studying tasks types from articles [7] and [4], following the distinction present in the latter and taking into account two types of tasks, proxy and real. In studies that use proxy tasks, the user mainly evaluates how well he perceives the AI’s explana- tions and what it has learned, focusing on the AI and on the actual goals the users have in interacting with the system. [16, 25, 21]. Conversely, studies that use real tasks evaluate the cooperation between users and AI: the user has a primary role regarding the decision to make and can decide or not to use the AI advice to complete the task [8, 4, 23]. The paper in [4] also criticises the current evaluation methodology of XAI based on proxy tasks, demonstrating that their conclusions may not reflect the usage of the system on real tasks. Given this discovery, we consider real tasks in the transformations explained in Section 4. 2.3 Types of Reasoning During the evaluation phase, where the interaction between user and AI takes place, one fundamental factor comes into play: the reasoning type. We started analyzing this subject from the article [22], cited above. The authors highlight that the AI’s role is to facilitate the user connection with its decisions, starting from the reasoning expressed through the AI’s explanations. Accordingly, a rea- sonable choice is to deeply explore a subset of the reasoning types, for instance, the logical ones: inductive, deductive and abductive. Here, we consider Peirce’s syllogistic theory [9], and note that we can translate between these by exchanging the conclusion (or result), the major premise (the rule) and the minor premise (the cause). We will investigate these reasoning types in the next section. 3 Investigating the Reasoning Types in Explainable Intelligent Interfaces In this section, we investigate the reasoning types previously mentioned borrow- ing some examples from the literature. Article [4] briefly discusses inductive and deductive reasoning, explaining how to integrate them in a user evaluation con- text but without an in-depth exploration. Furthermore, abductive reasoning is often (unintentionally) used to compare novel explanation techniques with state- of-the-art techniques where the user role is to identify what is the best-generated explanation during the evaluation. Before starting with the definitions, we in- troduce the three components identified in all types of reasoning (c.f., Peirce’s syllogistic theory [9]): one (or more) Cause (or Case/ Explanation/ Reason), Effect (or Observation/ Result) and Rule (or Generalization/ Theory). Table 1. Illustrative articles that explain neural network models, divided according to the type of reasoning, network and task. Year Article Reasoning of task Type/s of network Type/s of task 2018 [15] deductive MLP proxy 2019 [5] inductive RNN real 2017 [21] abductive LSTM, CNN proxy When applying this theory to XAI interfaces, it is important to identify whether their representation is implicit or explicit. For example, a rule or a cause is implicit when it comes from the user’s mental model and not from the AI. Instead, it is explicit when we consider the AI’s mental model, i.e. what it has learned in the training process and its explanations on the output’s predic- tion. The reference paper for the concepts we are going to describe is that of [9] (see Table 1). The order of the components described is unimportant, except for the last one: we will use the latter to highlight what is the reasoning to elicit from the user. Deduction: given a cause and a rule, deduce an effect. This type of reasoning starts with general rules and examines the possibilities to reach a specific, logical conclusion. Deductive reasoning is alternatively referred to as “top-down” logic because it usually starts with a general statement and ends with a narrower, specific conclusion. The article [15] contains an example of this reasoning, as depicted in Figure 1. Cause : The AI’s words in red that identify a negative or positive sentiment. Rule : Certain words contribute to the sentiment of text (implicit). Effect : The sentiment prediction. Fig. 1. The AI’s words in red that identify a negative or positive sentiment are the Cause ; the Rule is implicit (related to the mental model of the user regarding his semantic knowledge and not given by the AI); the user choice about the sentiment prediction (whether positive or negative), is the Effect [15]. Induction: given a cause and an effect, induce a rule. This type of reasoning involves drawing a general conclusion from a set of specific observations. It is al- ternatively referred to as “bottom-up” logic because it involves widening specific premises out into broader generalizations. Article [5] is an example of inductive reasoning, shown in Figure 2. Cause : The AI’s example-based explanations. Effect : The AI was unable to recognize the user’s sketch. Rule : Certain properties of sketches represent an object (implicit). Abduction: given an effect and a rule, abduct a cause. This type of reasoning typically begins with an incomplete set of observation/s and proceeds to the likeliest possible explanation. An example of abductive reasoning is [21], as de- picted in Figure 3. Effect : is given by the AI’s sentiment prediction. Rule : The chart in the explanation boxes give the user the intuition of the weights the AI uses for computing the valence of the sentence (implicit). Cause : The user selects the weights he considers the best (proxy task). Fig. 2. The example-based explanations shown by the AI identify the Cause ; the Effect is that the AI did not recognize the sketch the user draw; so, the user needs to understand the Rule from the AI’s examples. [5]. Fig. 3. The AI’s sentiment prediction represents the Effect ; the AI’s chart and the sentiment highlight that identify a negative or positive sentiment represents the Rule ; so the user has to find the best Cause , like: “I choose explanation B because it highlights better the sentiment meaning” [21]. 4 Transforming Explanations to different Reasoning Types The goal of transforming the reasoning task is to analyze possible preferences or performance differences on the user during the evaluation phase. Moreover, obtaining all three reasoning types allows us to find underrepresented reasoning types in the literature and to study if they work better with users than the orig- inal task’s reasoning. Abductive reasoning explanations (Fig. 3) are quite easy to generate and have a good understandability power: this compromise could be the reason why they are the most used reasoning for comparing the quality of explanation techniques. As for inductive explanations (Fig. 2), we can easily generate examples based on data, but the understandability is bounded by the selected examples. Deductive explanations (Fig. 1) are more challenging to gen- erate when we have to create explicit rules, but they are very understandable to the user. As mentioned in Section 2, the resulting transformation’s task will be a real one, for avoiding the mistake highlighted by article [4]. This can be achieved by revisiting the task’s question from the AI to the user perspective. Now, let us depict some ideas to formulate the transformation with the three reasoning types described previously, considering that we often have an explicit or implicit rule or cause, and nearly always an effect given by the AI (suggestion). Deductive to Inductive and Abductive. To adapt to inductive reasoning, we need to replace the cause giving similar or dissimilar examples concerning the data present in the task and based on the output of the AI, thus generating example-based explanations. So, the users grasp the Rule (that becomes im- plicit) that brought the AI to that result, and draw their conclusion Effect about the given task data. To switch to abductive reasoning, the AI should provide a Cause based on the task data. After that, we need to make the Rule implicit to not confuse the reasoning with the deductive one. Inductive to Deductive and Abductive. To switch to the deductive case, the AI needs to explicitly define a Rule , that also includes a Cause . We can ac- complish this by leveraging the properties of the AI model or adding a comple- mentary one to obtain a rule. Additionally, sometimes the user may take the decision without obtaining the Effect explicitly from the AI, but letting the user deduce it from the rules and causes. Passing instead from inductive to ab- ductive reasoning, we use the common traits in the inductive examples to create a Cause . Abductive to Inductive and Deductive. Starting from this reasoning type, we hypothesize to already have an Effect given by the AI. To translate to the inductive case, we need to replace the Cause given by the task’s data with that of example-based explanations and transform the Rule to implicit. To move to deductive reasoning, we need to explicitly define the AI’s Rule , that may change the original Cause , and if we want, hide the Effect . 5 Conclusion and Future Work In sum, we investigated the considerations that arise when transferring between different types of logical reasoning, considering real tasks as the resulting trans- formation’s tasks. We identified the importance of differentiating between im- plicit and explicit rule representation. We also consider whether the choice of reasoning type balances the technical challenges of generating the explanations, and the effectiveness of the explanations for humans. As future work, we plan to validate these ideas in user evaluations for different reasoning types. Also, we plan to create a taxonomy considering reasoning and task types, in addition to other useful metrics related to the XAI explanation, and further explore logical reasoning on other black-box models beyond neural networks. References 1. Bach, S., Binder, A., Montavon, G., Klauschen, F., Müller, K.R., Samek, W.: On pixel-wise explanations for non-linear clas- sifier decisions by layer-wise relevance propagation. PLOS ONE 10(7), 1–46 (07 2015). https://doi.org/10.1371/journal.pone.0130140, https://doi.org/10.1371/journal.pone.0130140 2. Barredo Arrieta, A., Diaz Rodriguez, N., Del Ser, J., Bennetot, A., Tabik, S., Bar- bado González, A., Garcia, S., Gil-Lopez, S., Molina, D., Benjamins, V.R., Chatila, R., Herrera, F.: Explainable artificial intelligence (xai): Concepts, taxonomies, op- portunities and challenges toward responsible ai. Information Fusion (12 2019). https://doi.org/10.1016/j.inffus.2019.12.012 3. Buhrmester, V., Münch, D., Arens, M.: Analysis of explainers of black box deep neural networks for computer vision: A survey. ArXiv abs/1911.12116 (2019) 4. Buçinca, Z., Lin, P., Gajos, K.Z., Glassman, E.L.: Proxy tasks and sub- jective measures can be misleading in evaluating explainable ai sys- tems. Proceedings of the 25th International Conference on Intelligent User Interfaces (Mar 2020). https://doi.org/10.1145/3377325.3377498, http://dx.doi.org/10.1145/3377325.3377498 5. Cai, C.J., Jongejan, J., Holbrook, J.: The effects of example-based explanations in a machine learning interface. In: Proceedings of the 24th International Conference on Intelligent User Interfaces. p. 258–262. IUI ’19, Association for Computing Ma- chinery, New York, NY, USA (2019). https://doi.org/10.1145/3301275.3302289, https://doi.org/10.1145/3301275.3302289 6. Choi, E., Bahadori, T., Schuetz, A., Stewart, W., Sun, J.: Retain: Interpretable predictive model in healthcare using reverse time attention mechanism (08 2016) 7. Doshi-Velez, F., Kim, B.: Towards a rigorous science of interpretable machine learn- ing (2017) 8. Fang, H., Gupta, S., Iandola, F., Srivastava, R.K., Deng, L., Dollar, P., Gao, J., He, X., Mitchell, M., Platt, J.C., et al.: From captions to visual concepts and back. 2015 IEEE Conference on Computer Vision and Pat- tern Recognition (CVPR) (Jun 2015). https://doi.org/10.1109/cvpr.2015.7298754, http://dx.doi.org/10.1109/CVPR.2015.7298754 9. Flach, P., Kakas, A.: Abductive and inductive reasoning: Background and issues (01 2000). https://doi.org/10.1007/978-94-017-0606-3-1 10. Guidotti, R., Monreale, A., Ruggieri, S., Pedreschi, D., Turini, F., Gian- notti, F.: Local rule-based explanations of black box decision systems. ArXiv abs/1805.10820 (2018) 11. Guidotti, R., Monreale, A., Ruggieri, S., Turini, F., Giannotti, F., Pe- dreschi, D.: A survey of methods for explaining black box models. ACM Computing Surveys 51(5), 1–42 (Jan 2019). https://doi.org/10.1145/3236009, http://dx.doi.org/10.1145/3236009 12. Huang, X., Kroening, D., Ruan, W., Sharp, J., Sun, Y., Thamo, E., Wu, M., Yi, X.: A survey of safety and trustworthiness of deep neural networks: Verifica- tion, testing, adversarial attack and defence, and interpretability. Computer Sci- ence Review 37, 100270 (Aug 2020). https://doi.org/10.1016/j.cosrev.2020.100270, http://dx.doi.org/10.1016/j.cosrev.2020.100270 13. Lundberg, S., Lee, S.I.: A unified approach to interpreting model predictions (12 2017) 14. Mohseni, S., Zarei, N., Ragan, E.D.: A multidisciplinary survey and framework for design and evaluation of explainable ai systems (2018) 15. Nguyen, D.: Comparing automatic and human evaluation of local explanations for text classification. In: Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Tech- nologies, Volume 1 (Long Papers). pp. 1069–1078. Association for Computational Linguistics, New Orleans, Louisiana (Jun 2018). https://doi.org/10.18653/v1/N18- 1097, https://www.aclweb.org/anthology/N18-1097 16. Rajani, N.F., Mooney, R.J.: Ensembling visual explanations for vqa. In: Pro- ceedings of the NIPS 2017 workshop on Visually-Grounded Interaction and Language (ViGIL) (December 2017), http://www.cs.utexas.edu/users/ai-labpub- view.php?PubID=127684 17. Ribeiro, M., Singh, S., Guestrin, C.: “why should i trust you?”: Explaining the pre- dictions of any classifier. pp. 97–101 (02 2016). https://doi.org/10.18653/v1/N16- 3020 18. Selvaraju, R.R., Cogswell, M., Das, A., Vedantam, R., Parikh, D., Batra, D.: Grad-cam: Visual explanations from deep networks via gradient-based localiza- tion. International Journal of Computer Vision 128(2), 336–359 (Oct 2019). https://doi.org/10.1007/s11263-019-01228-7, http://dx.doi.org/10.1007/s11263- 019-01228-7 19. Shrikumar, A., Greenside, P., Kundaje, A.: Learning important features through propagating activation differences (04 2017) 20. Tjoa, E., Guan, C.: A survey on explainable artificial intelligence (xai): Towards medical xai (2019) 21. Tsang, M., Sun, Y., Ren, D., Liu, Y.: Can i trust you more? model-agnostic hier- archical explanations (2018) 22. Wang, D., Yang, Q., Abdul, A., Lim, B.: Designing theory-driven user-centric ex- plainable ai (05 2019). https://doi.org/10.1145/3290605.3300831 23. Yin, M., Vaughan, J., Wallach, H.: Understanding the effect of ac- curacy on trust in machine learning models. pp. 1–12 (04 2019). https://doi.org/10.1145/3290605.3300509 24. Yu, R., Shi, L.: A user-based taxonomy for deep learning visualization. Visual Informatics 2 (09 2018). https://doi.org/10.1016/j.visinf.2018.09.001 25. Zhou, B., Sun, Y., Bau, D., Torralba, A.: Interpretable basis decomposition for vi- sual explanation. In: Proceedings of the European Conference on Computer Vision (ECCV) (September 2018)