=Paper=
{{Paper
|id=Vol-2071/keynote1
|storemode=property
|title=Explain to whom? Putting the User in the Center of Explainable AI
|pdfUrl=https://ceur-ws.org/Vol-2071/CExAIIA_2017_keynote_1.pdf
|volume=Vol-2071
|authors=Alexandra Kirsch
|dblpUrl=https://dblp.org/rec/conf/aiia/Kirsch17
}}
==Explain to whom? Putting the User in the Center of Explainable AI==
Explain to whom? Putting the User in the Center of Explainable AI Alexandra Kirsch [orcid.org/0000−0002−5663−1798] Universität Tübingen alexandra.kirsch@uni-tuebingen.de Abstract. The ability to explain actions and decisions is often regarded as a basic ingredient of cognitive systems. But when researchers propose methods for making AI systems understandable, users are usually not involved or even mentioned. However, the purpose is to make people willing to accept the decision of a machine or to be better able to inter- act with it. Therefore, I argue that the evaluation of explanations must involve some form of user testing. 1 Reasons for Explanations I regularly present the following definition by Brachman [1] in my AI lecture: A truly cognitive system would be able to learn from its experience — as well as by being instructed — and perform better on day two than it did on day one. It would be able to explain what it was doing and why it was doing it. It would be reflective enough to know when it was heading down a blind alley or when it needed to ask for information that it simply couldn’t get to by further reasoning. [. . . ] Once a student raised his hand and asked whether the ability to explain one’s own actions is a necessary feature of cognition. After all, he cannot explain all his actions (and if he does, it is doubtful whether such an explanation reflects the psychological processes that really led to the action) and still would be considered as a cognitive system. The ability of an agent to explain its own decisions is often regarded as a necessary feature without a clear understanding of why agents should have this ability. Connected to this is the question of what an explanation is. With the current hype of statistical machine learning, the general feeling is that these methods need some way to make their decisions plausible to people (like the DARPA initiative Explainable Artificial Intelligence1 ). But when researchers talk about explanations, they hardly ever consider the end user. In this paper I argue that comprehensibility and explainability must always be regarded in the context of a specific use case and that its adequacy can only be determined by interaction with users. 1 https://www.darpa.mil/program/explainable-artificial-intelligence Copyright © 2018 for this paper by its authors. Copying permitted for private and academic purposes. 2 Usability Principles The idea to build comprehensible technical devices is much older than the cur- rent trend in AI. Many usability guidelines from the fields of design and human- computer interaction include implicit or explicit explanation to users. Norman [2] proposes four basic design principles, all of them contain elements of explain- ability: conceptual model: users should understand the underlying mechanisms of the system; visibility: all functionality should be visible, a kind of explanation of what the system is capable of; mapping: the visible elements should intuitively map to functionality; feedback: the user should always be informed about the system’s state. Thus, any device should be designed in a way that users are aware of the system’s capabilities and current state. This can be interpreted as a kind of explanation, one can also say that such a system does not need any additional explanation. For AI researchers this means that 1) comprehensible systems can often be built by adhering to well-known usability principles, 2) the design and usability of the overall system is often more important than AI features to result in an overall positive user experience. 3 Legibility Embodied and virtual agents differ from typical human-computer interaction in that they can perform more pronounced actions and often have additional sensing capabilities. What is an explanation in this context? Of course, such agents could generate explanations of their actions, possibly in natural language. But in many cases a constant stream of explanations may annoy users and its effectiveness is questionable. When we look at human-human interaction, we hardly ever need to explain our actions to others. So why should a machine do so? A more natural way of interaction could be to implicitly communicate goals and the necessity of actions. Lichtenthäler and Kirsch [3] define the term legibility as follows: Robot behavior is legible if: (Factor 1) a human observer or interactor is able to understand its intentions, and (Factor 2) the behavior met the expectations of the human observer or interactor. This definition centers on the human observer, implying that legibility can only be determined by experiments involving users. Such tests can again be inspired by standard usability testing, but the quality criteria and setup must often be adapted to the task and embodied situation of the agent. In addition, if physical robots are involved, there are additional questions of how to ensure the safety of participants. As an example, consider the task of legible robot navigation [4]. Robot nav- igation has traditionally been treated as a purely technical problem, but with humans sharing the space of the robot or directly interacting with it, legibility becomes an issue. But then, the quality of a navigation act cannot just be char- acterized by its success and possibly the time needed to reach the goal point, one must also determine whether human observers can indeed infer the goal point and that the human expectations are met [5]. Performing user tests is time-consuming. Therefore, it would be nice to have a general set of measures that determine legibility, but that can be measured without direct user involvement. An experiment of Papenmeier et al. [6] identifies two factors that influence the legibility of robot navigation: 1) the movement direction: a robot looking towards the direction it is moving to (or towards the goal point, both were identical in the experiment) is perceived as more autonomous than one with a different orientation while moving; 2) the velocity profile: deceleration of the robot causes surprise in human observers (as measured by a viewing time paradigm), while acceleration and constant velocities do not. 4 Explainability and Interaction Legibility as used above is a criterion for passive interaction: even though a person has no direct contact with the agent, the behavior should be legible. But it is also a prerequisite for direct interaction, for example a person trying to initiate an interaction would have to be able to predict the robot’s movements. But direct interaction may require more than just a basic, implicit understanding of the other’s intentions, it may at times require a more explicit explanation. Decision methods can be designed in a way that resembles human decision making. The Heuristic Problem Solver [7] and the FORR architecture [8] follow closely the way that humans make decisions: they explicitly generate alternatives and evaluate them. In this way, an explanation is generated along with the solution: the alternatives that were considered as well as the reasons for choosing one can directly be communicated to users [9]. In addition, such an explicit decision-making paragidm enables an interactive process of humans and machines jointly determining a solution. It is practically impossible to model all the knowledge a person has with respect to a task. But if people can interact with the decision process of the machine, they can use this additional knowledge without having to formalize all of it. Thus, a person could propose or delete alternatives, or change the evaluations of alternatives. Here again, the question is whether the explanation is adequate. Ideally, the machine should provide enough information for a person to be satisfied with or even be willing to take responsibility for the decision. Or in the case of interaction, a possible measurement is whether the joint decision-making is more efficient and effective than a purely computational or purely human one. 5 Conclusion The ability to explain one’s decisions and actions is often unquestioningly re- quired from cognitive systems. But the frequency, form and implementation of explanations depend on the specific application and context. In many cases, standard feedback mechanisms known from human-computer interaction are all that is needed. Other applications may demand that users are able to identify the source of errors. Even others may need an explicit dialogue between human and machine that allows the machine to request help if necessary. Explanations can take different forms, from standard interaction methods of flashing lights, sounds and graphical displays, to language or the display of legible actions. In the last section I have mentioned systems that generate an explanation along with the solution. An alternative is the post-hoc generation of explanations for black box algorithms, which is currently popular with the rise of statistical machine learning. In all respects, the users are the leveling rule. An explanation is not a math- ematical construct, an explanation is good if people find it helpful in the specific context. Therefore, the AI community should expand its evaluation metrics be- yond optimization criteria to user-centered measures and evaluation procedures. . References 1. Brachman, R.: Systems that know what they’re doing. IEEE Intelligent Systems (November/December 2002) 67 – 71 2. Norman, D.A.: The Design of Everyday Things. Basic Books, Inc., New York, NY, USA (2002) 3. Lichtenthäler, C., Kirsch, A.: Legibility of Robot Behavior : A Literature Review. preprint at https://hal.archives-ouvertes.fr/hal-01306977 (April 2016) 4. Kruse, T., Pandey, A.K., Alami, R., Kirsch, A.: Human-aware robot navigation: A survey. Robotics and Autonomous Systems 61(12) (2013) 1726–1743 5. Lichtenthäler, C., Lorenz, T., Karg, M., Kirsch, A.: Increasing perceived value between human and robots — measuring legibility in human aware navigation. In: IEEE Workshop on Advanced Robotics and its Social Impacts (ARSO). (2012) 89– 94 6. Papenmeier, F., Uhrig, M., Kirsch, A.: Human understanding of robot motion: The role of velocity and orientation. preprint at https://osf.io/tphnu/ (2017) 7. Kirsch, A.: Heuristic decision-making for human-aware navigation in domestic en- vironments. In: 2nd Global Conference on Artificial Intelligence (GCAI). (2016) 8. Epstein, S., Aroor, A., Evanusa, M., Sklar, E., Simon, S.: Navigation with learned spatial affordances. (2015) 9. Korpan, R., Epstein, S.L., Aroor, A., Dekel, G.: WHY: Natural explanations from a robot navigator. In: AAAI 2017 Fall Symposium on Natural Communication for Human-Robot Collaboration. (2017)