=Paper= {{Paper |id=Vol-1893/Paper2 |storemode=property |title=Explanation Systems for Influence Maximization Algorithms |pdfUrl=https://ceur-ws.org/Vol-1893/Paper2.pdf |volume=Vol-1893 |authors=Amulya Yadav,Aida Rahmattalabi,Ece Kamar,Phebe Vayanos,Milind Tambe,Venil Loyd Noronha |dblpUrl=https://dblp.org/rec/conf/ijcai/YadavRKVTN17 }} ==Explanation Systems for Influence Maximization Algorithms== https://ceur-ws.org/Vol-1893/Paper2.pdf
Proceedings of the 3rd International Workshop on Social Influence Analysis (SocInf 2017)
August 19th, 2017 - Melbourne, Australia




                        Explanation Systems for Influence Maximization
                                         Algorithms

                Amulya Yadav1 , Aida Rahmattalabi1 , Ece Kamar2 , Phebe Vayanos1 , Milind Tambe1 ,
                                             Venil Loyd Noronha1
               1
                   Center for Artificial Intelligence in Society, University of Southern California, LA, CA, 90089
                                              2
                                                Microsoft Research, Redmond, WA 98052
                               1
                                 {amulyaya, rahmatta, phebe.vayanos, tambe, vnoronha}@usc.edu
                                                     2
                                                       {eckamar}@microsoft.com



                         Abstract. The field of influence maximization (IM) has made rapid advances,
                         resulting in many sophisticated algorithms for identifying “influential” members
                         in social networks. However, in order to engender trust in IM algorithms, the
                         rationale behind their choice of “influential” nodes needs to be explained to its
                         users. This is a challenging open problem that needs to be solved before these al-
                         gorithms can be deployed on a large scale. This paper attempts to tackle this open
                         problem via four major contributions: (i) we propose a general paradigm for de-
                         signing explanation systems for IM algorithms by exploiting the tradeoff between
                         explanation accuracy and interpretability; our paradigm treats IM algorithms as
                         black boxes, and is flexible enough to be used with any algorithm; (ii) we utilize
                         this paradigm to build XplainIM, a suite of explanation systems; (iii) we illus-
                         trate the usability of XplainIM by explaining solutions of HEALER (a recent IM
                         algorithm) among ∼200 human subjects on Amazon Mechanical Turk (AMT);
                         and (iv) we provide extensive evaluation of our AMT results, which shows the
                         effectiveness of XplainIM.


               1      Introduction

               The field of influence maximization (IM) deals with finding a set of “influential” nodes
               in a social network, which can optimally spread influence throughout the network. Sig-
               nificant progress in this field over the last few years has led to a variety of sophisticated
               algorithms [1] for finding influential nodes. Recently, [10] evidenced the real-world
               usability of IM algorithms by deploying three such algorithms in the real world.
                   As with any technology, evidence of real-world usability of IM algorithms brings
               to the fore other important questions that need to be tackled before their widespread
               adoption. One of the most important open questions for IM algorithms (which are used
               as decision support tools by domain experts such as viral marketers, social workers,
               etc.) is how to engender trust by explaining their recommendations to domain experts
               [7]. This is important because explaining the rationale behind these algorithms decisions
               helps the domain experts decision making process: they can then choose to either use the
                    Copyright c 2017 for the individual papers by the papers’ authors. Copying permitted for
                    private and academic purposes. This volume is published and copyrighted by its editors.




                                                                8
Proceedings of the 3rd International Workshop on Social Influence Analysis (SocInf 2017)
August 19th, 2017 - Melbourne, Australia




               algorithm’s recommendation or discard it. Indeed, [9] noted the challenge of explaining
               the solutions of HEALER (an IM algorithm) to homeless shelter officials, who wanted
               explanations for why HEALER’s solutions were better than their preferred solutions.
               This raises the following important question: How to explain solutions generated by IM
               algorithms to its intended users? To the best of our knowledge, this problem has not
               received any attention in the literature so far.
                   Designing such an explanation system for IM algorithms is a challenging task for
               two reasons. First, instead of building different explanation systems for different IM
               algorithms, we want to build a single explanation system which is generalizable, i.e.,
               it should be flexible enough to explain the solutions of every possible IM algorithm.
               This makes our task challenging because we can no longer rely on “explaining the
               algorithm to explain the solution” (since there is no fixed algorithm). Second, most
               human end-users have cognitive limitations and therefore, in order to avoid overloading
               them with information, we need to ensure that our system’s explanations are relatively
               interpretable. However, this involves trading off explanation accuracy (i.e., correctness
               of explanation) with its interpretability (as we explain later).
                   This paper attempts to address this open problem via four major contributions. First,
               we propose a machine learning based paradigm for designing generalizable explanation
               systems for IM algorithms. Our paradigm achieves generalizability by using an algo-
               rithm independent classification dataset to build a machine learning classifier, which
               generates natural language explanations (in terms of this dataset’s features) for IM so-
               lutions. In order to aid interpretability of the generated explanations, our classification
               dataset does not contain features for un-interpretable complicated terms (e.g., marginal
               gain). This transforms the problem of explaining IM algorithm solutions into a problem
               of explaining the predictions made by that machine learning classifier (as we explain
               later). Further, our paradigm exploits a fundamental tradeoff between explanation accu-
               racy and interpretability (i.e., more accurate explanations are less interpretable, and vice
               versa) to generate a Pareto frontier of machine learning classifiers, which is then used
               to generate natural language explanations of varying accuracy and interpretability. Sec-
               ond, we utilize our paradigm to build XplainIM, a suite of explanation systems. Third,
               we illustrate the usability of XplainIM by explaining the solutions of HEALER (a recent
               IM algorithm) to ∼200 human subjects on Amazon Mechanical Turk (AMT). Fourth,
               we provide extensive evaluation of our AMT results, which shows the effectivenes of
               XplainIM.
                    Motivating IM Algorithm Our proposed explanation system is flexible enough to
               explain the solutions of any IM algorithm. However, we motivate our discussion in
               the rest of the paper by focusing on HEALER [9], a recent IM algorithm. We choose
               HEALER primarily because its paper [9] was the first to pose the question of explain-
               ing the solutions of an IM algorithm. HEALER was designed to select multiple sets
               of “influential” nodes sequentially from a homeless youth social network, who could
               optimally spread awareness about HIV in their social network. HEALER solves this
               problem by modeling it as a Partially Observable Markov Decision Process (POMDP)
               [5] and uses hierarchical ensembling techniques to scale up to real-world network sizes
               [9].




                                                           9
Proceedings of the 3rd International Workshop on Social Influence Analysis (SocInf 2017)
August 19th, 2017 - Melbourne, Australia




                   Thus, HEALER is a specialized algorithm for solving a more general problem than
               standard influence maximization (defined below). However, in order to ensure the gen-
               eralizability of our explanation system to other IM algorithms, we ignore the POMDP
               style sequential planning aspects of HEALER, and use it as an algorithm for solving the
               standard single-shot influence maximization problem: Given a social network graph G,
               select a set of K “influential” nodes which optimally spread influence in the network
               after T time steps.


               2   Related Work

               The problem of explaining solutions of IM algorithms was first posed by [9] in the con-
               text of HEALER, but their main focus was on justifying the need for such explanations,
               as opposed to providing any practical solutions to this problem. While explanation sys-
               tems for POMDPs have been proposed [8], they cannot be applied to explain solutions
               of every IM algorithm. To the best of our knowledge, no other previous work has looked
               at this problem so far.


               3   Machine Learning based Paradigm for Explanation Systems

               We now define an explanation system formally. Given a social network graph G as
               input, an IM algorithm outputs an optimal set of K nodes OG . Moreover, the end-user
               of the IM algorithm has a preferred set of K nodes UG , which he/she wrongly thinks is
               the optimal solution. Then, an explanation system is defined as follows:

               Definition 1. An explanation system for IM algorithms is a function E : G × OG ×
               UG → S which takes a graph G, an IM algorithm’s solution on G (i.e., OG ) , and an
               end-user’s preferred solution on G (i.e., UG ) as input, and produces as output a natural
               language string S, which explains why OG is a better solution than UG on graph G.

                   We believe that Definition 1 accounts for most practical use cases of explanation
               systems. For example, [9] noted that homeless shelter officials (i.e., end-users) wanted
               explanations for why HEALER’s solutions were better than their preferred solutions.
               Next, we explain our novel paradigm for designing explanation systems (for IM algo-
               rithms) in two stages.
                   First, we transform our original problem of explaining IM solutions into an alternate
               problem: “how to explain predictions made by ML classifiers?” Second, having made
               this transformation, we provide novel solutions for explaining predictions made by ML
               classifiers.
                   Step 1: Transformation Consider the following two problems for any given IM
               algorithm ∆:

               Problem 1. Given as input an IM algorithm ∆, output an explanation system E ∆ which
                                             ∆
               explains why ∆’s solutions (OG  ) are better than UG (for all possible G and UG ).




                                                         10
Proceedings of the 3rd International Workshop on Social Influence Analysis (SocInf 2017)
August 19th, 2017 - Melbourne, Australia




               Problem 2. Given as input (i) an IM algorithm ∆, and (ii) a binary classifier M , which
                                                                                              ∆
               perfectly (i.e., without any misclassifications) differentiates ∆’s solutions OG from all
               other possible solutions UG (for all possible UG and G), output an explanation system
                                                                        ∆
               E M which explains how classifier M differentiates OG       from UG .

                   We claim that solving Problem 2 is sufficient to solve Problem 1, as formally stated
               below (proof in appendix1 ):

               Theorem 1. Problem 2 solution =⇒ Problem 1 solution

                   As a result, we transform the problem of explaining IM solutions (Problem 1) into
               the alternate problem: “how to explain predictions made by ML classifiers?” (Problem
               2). This completes the transformation step of our approach. Next, we provide a novel
               solution method for solving Problem 2.
                   Step 2: Solution Method First, we specify the input to Problem 2: an ML classifier
               M which can differentiate between OG and UG . Second, we need to use this classifier
               M to generate output for Problem 2: the process of explaining the predictions made by
               classifier M needs to be specified. We focus on the input specification in this section
               (output generation is explained in Section 5).
                   To train classifier M , each possible IM solution (i.e., set of K nodes) is mapped to
               a set of network-centric feature values (e.g., degree, mutual connections, etc.), and is
               labeled as either being “optimal” (if it is the IM algorithm’s solution) or “sub-optimal”
               (for any other solution). These network-centric features generalize across all IM algo-
               rithms, as they only depend on the network based characteristics of any possible IM
               solution. This creates a labeled dataset (details in Section 6) which is then used to train
               the ML classifier M which differentiates an optimal solution from a sub-optimal solu-
               tion.
                   We use decision trees [6] as our ML classifier of choice. The key advantage of
               decision trees over other classifiers is their interpretability, as their predictions can be
               explained to end-users as a conjunction of decision rules, where each rule involves a
               single feature in the dataset. In many applications, including our own, this interpretabil-
               ity is preferred over other methods that may have higher accuracy but are relatively
               un-interpretable [9].
                   Unfortunately, while decision trees are more interpretable than other ML classifiers
               (e.g., neural nets), the explanations for their predictions (i.e., a conjunction of decision
               rules) can overload human users with too much information to process (as we show
               in our experiments). Thus, there is a need to ensure interpretability of the generated
               explanations.
                   We address this challenge by introducing “interpretable decision trees”, which al-
               lows us to trade off the prediction accuracy of a decision tree with its interpretability
               in a quantifiable manner. Further, we exploit the competing objectives of prediction ac-
               curacy and interpretability to generate a Pareto frontier of decision trees. This Pareto
               frontier allows us to correctly differentiate between OG and UG , while maintaining
               different levels of interpretability. We then utilize trees from this Pareto frontier to gen-
               erate explanations for the IM algorithm’s solutions (see Section 4). The novelty of our
                1
                    See http://bit.ly/2kXDhE4 for proofs




                                                           11
Proceedings of the 3rd International Workshop on Social Influence Analysis (SocInf 2017)
August 19th, 2017 - Melbourne, Australia




               work lies in the versatility of our explanation system, which is able to provide explana-
               tions of varying interpretability upon demand. Specifically, our explanation system is
               flexible enough to dynamically increase the interpretability of the generated explanation
               (by using more “interpretable” decision trees to generate the explanation), if the end-
               user is unconvinced about previously supplied explanations. We now formally define
               interpretable decision trees.


               4     Interpretable Decision Trees
               Decision trees take an input a labeled dataset with N datapoints (X i , Y i ) ∀i ∈ {1, N },
               where X i is a p-dimensional feature vector for the ith datapoint, and Y i is the label
               associated with the datapoint. In our domain, X i represents the vector of network-
               centric feature values corresponding to the ith IM solution (i.e., set of K nodes) in
               our dataset. Similarly, Y i = 1 if the ith solution in our dataset is the IM algorithm’s
               solution, and it is 0 otherwise.
                   While a decision tree’s prediction for X i can be explained as a conjunction of all
               the feature splits that X i traverses on its way to a leaf node, this explanation is un-
               interpretable, as it overloads end-users with too much information (we validate this in
               our experiments). Thus, in order to ensure interpretability of the generated explanations,
               we introduce the notion of interpretable decision trees, which allows us to tradeoff the
               explanation accuracy with its interpretability in a quantifiable manner. Next, we define
               the interpretability score, which quantifies the interpretability of a decision tree.
                   Interpretability Score We associate with each feature t ∈ {1, p} an interpretabil-
               ity weight wt ∈ [0, 1] which quantifies the interpretability of feature t. We estimate
               wt values ∀t ∈ {1, p} by surveying AMT workers (see Section 6). Further, let xD       t ∀t ∈
               {1, p} denote a binary variable which indicates whether a split with feature t is used
               at any branch node in the decision tree D or not. Then, the interpretability score of
                                                      p              p
               decision tree D is defined as: ID = ( xD                xD
                                                      P             P
                                                           t wt )/(      t ).
                                                      t=1          t=1
                   Thus, trees which use fewer number of features to split (the denominator in ID ’s
               definition) have higher ID scores (as it is easier to explain the tree’s predictions with
               fewer features). Moreover, trees which use features having more interpretable weights
               have higher ID scores. Next, we define the A&I score of a decision tree, which captures
               the tradeoff between its prediction accuracy and interpretability.
                   A&I Score For decision trees (and most other classification algorithms), inter-
               pretability and prediction accuracy (measured by ID and F1 score2 , respectively) are
               two competing objectives. Increasing the prediction accuracy of a decision tree leads to
               a loss of its interpretability, and vice versa.
                   The A&I score of decision tree D captures the tradeoff between these competing
               objectives and is defined as follows: A&ID (α) = F1 D + αID . Here, α is a parameter
               which controls the tradeoff between the F1 and ID scores. Higher α values lead to
               highly interpretable decision trees with low prediction accuracy, and vice versa. Next,
               we explain an algorithm which finds tree Dα which maximizes the A&I(α) score (i.e.,
               it optimally trades off between F1 and ID scores for the tradeoff level specified by α).
                2
                    F1 score = 2T P/(2T P + F P + F N )




                                                            12
Proceedings of the 3rd International Workshop on Social Influence Analysis (SocInf 2017)
August 19th, 2017 - Melbourne, Australia




                 Algorithm 1: Maximize A&I score
                   Input: α, Dataset (X i , Y i ) ∀i ∈ {1, N }
                   Output: Dα , the A&I score maximizing decision tree
                 1 (T rain, V alidation, T est) = Split(Dataset);
                 2 T ree = CART (T rain);
                 3 for level ∈ P runingLevels(T ree) do
                 4      P runedT ree = P rune(T ree, level);
                 5      rlevel = A&I(α, P runedT ree, V alidation);
                 6 OptLevel = argmaxj rj ;
                 7 Dα = P rune(T ree, OptLevel);
                 8 return Dα ;


                   Maximizing A&I score For each α value, we use Algorithm 1 to build a decision
               tree (on the dataset from Section 3) which maximizes A&I(α). First, we randomly
               split our dataset into a training, validation, and test set in Step 1(in our experiments,
               we use a 80:10:10 split). Next, we use CART, a state-of-the-art decision tree algorithm
               [4], to learn a decision tree on our training set in Step 2. Normally, CART outputs
               deep decision trees, which have high prediction accuracies on the training set. However,
               deeper trees have lower ID scores, as they have more number of features. As a result,
               we use our validation set to find the optimal pruning level of CART’s decision tree,
               which maximizes A&I(α). Specifically, for each possible pruning level , we prune
               CART’s decision tree to that level in Step 4. We use this pruned decision tree’s F1 score
               on the validation set to calculate the pruned tree’s A&I(α) score on the validation set
               (Step 5). Finally, we select the optimal pruning level which maximizes A&I(α) on the
               validation set (Step 6), and prune CART’s decision tree to this level (Step 7). To account
               for randomization in splitting of the dataset into the training, validation and test set, we
               average over 50 such splits.
                   Decision Tree Pareto Frontier Instead of generating a single decision tree (using
               Algorithm 1), we exploit the accuracy-interpretability tradeoff by varying the value of
               α to generate multiple decision trees. We do this because of two reasons. First, end-
               users may find explanations generated from one decision tree hard to understand, and
               thus, we may need a more interpretable tree (measured by ID ) to generate the explana-
               tions. Second, explanations can only be generated using decision trees which correctly
               differentiate between the IM algorithm’s optimal solution (OG ) and the end-user’s sub-
               optimal solution (UG ) (else we don’t have a valid classifier M in Problem 2). Thus, if
               more interpretable (less accurate) trees fail to differentiate between OG and UG , we
               need more accurate (less interpretable) trees to generate explanations.
                    Therefore, for each α value, we find the optimal Dα tree using Algorithm 1. Thus,
               we get a set of trees Θ, each of which has different F1 and ID scores. Next, we remove
               all trees D ∈ Θ which are dominated by other trees D0 in terms of both F1 and ID
                                                                            0
               scores (i.e., there exist trees D0 ∈ Θ for which either {F1D > F1D ∧ ID0 > ID } or
                   D0      D
               {F1 > F1 ∧ ID0 > ID } holds). This gives us our Pareto frontier of trees.
                   Our strategy is to search this Pareto-frontier for a decision tree D∗ , which can cor-
               rectly differentiate between OG and UG , and then use D∗ to generate explanations.




                                                           13
Proceedings of the 3rd International Workshop on Social Influence Analysis (SocInf 2017)
August 19th, 2017 - Melbourne, Australia




               Next, we complete the discussion of our explanation system by proposing XplainIM,
               a solution method for Problem 2 which utilizes our Pareto frontier (recall that solving
               Problem 2 is sufficient to create an explanation system for IM algorithms).


               5   Solution for Problem 2: XplainIM
               Problem 2 takes as input a machine learning classifier M (which can differentiate be-
               tween OG and UG ), and produces as output an explanation system which explains how
               classifier M differentiates OG from UG . We propose XplainIM as a solver for Problem
               2. We begin by discussing XplainIM’s design.
                    XplainIM builds an explanation system for a specific IM algorithm in two stages. In
               the offline stage, it takes as input a classification dataset for that algorithm, and uses it
               to store a Pareto-frontier of decision trees. This input dataset is created using a fixed set
               of network-centric features (Figure 1), irrespective of the specific IM algorithm under
               consideration. In the online stage, it takes as input a network G and user solution UG
               and uses them to choose classifier M . We now explain how XplainIM chooses this
               classifier M .
                    Choosing Classifier Given a choice of OG and UG , XplainIM finds the set F of
               “feasible trees” in its Pareto frontier which can successfully classify OG as optimal, and
               UG as sub-optimal. Each tree in F is a valid classifier for Problem 2, as it can differen-
               tiate between OG and UG . To choose M , XplainIM simply uses the most interpretable
               tree in F, i.e., M = argmaxD∈F ID .
                    Note that it is possible that for some choice of OG and UG , no tree in our Pareto
               frontier is feasible. Then, solving Problem 2 is impossible, as we wouldn’t have an
               appropriate classifier M . However, this case rarely arises in practice (see Section 6).
                    Explanation System Next, having chosen a feasible decision tree as the classifier
               M , we now discuss XplainIM’s explanation system, which generates explanations for
               “how M differentiates OG from UG ”. Xplain relies on a natural language template
               (described below) to generate explanations, which is populated at runtime (depending
               on OG and UG ).

               Template 1 Your solution UG had a ft feature value of UG (ft ). On the other hand,
               the optimal solution OG had a ft feature value of OG (ft ). For solutions to be optimal,
               the value of ft feature should be {6, >}bt .

                    Given OG and UG , we propose three different methods for populating Template 1.
               Each of these methods relies on the observation that the paths taken by OG and UG
               (to leaf nodes) in tree M diverge at some branch node t∗ ∈ TB (because tree M is
               feasible). We use features ft and corresponding thresholds bt (recall Section 4) starting
               from this point of divergence t∗ to populate Template 1 to generate explanations for the
               end-user. Our three methods are explained below:
                    1. XplainIM-A Template 1 is populated using only ft∗ and bt∗ for branch node t∗ ,
               which is the first point of divergence between paths of OG and UG in tree M . This
               populated template is presented as explanation to the end-user.
                    2. XplainIM-B Template 1 is populated using ft∗ and bt∗ for branch node t∗ (simi-
               lar to XplainIM-A). However, the choice of classifier M used to generate explanations




                                                           14
Proceedings of the 3rd International Workshop on Social Influence Analysis (SocInf 2017)
August 19th, 2017 - Melbourne, Australia




               is different. Instead of using the feasible tree with the highest ID score as our classifier
               M , we use the feasible tree in which the feature ft∗ (present at the branch node t∗ )
               has the highest interpretability weight wft∗ . This is because only feature ft∗ is used to
               populate Template 1, and a high value of wft∗ would ensure the interpretability of the
               generated explanation.
                   3. XplainIM-Full Instead of providing explanations in terms of a single feature,
               XplainIM-Full uses multiple features in its explanation. Starting from branch node t∗ ,
               XplainIM-Full populates Template 1 for all features along the two divergent paths of
               OG and UG . This method produces explanations which are identical in structure to
               ones generated by decision sets [2].


               6   Evaluation

               We evaluated XplainIM by using it to explain HEALER’s solutions to human work-
               ers on AMT. We provide results from these experiments in three parts. First, we de-
               scribe the HEALER specific dataset (which XplainIM uses to build its Pareto-frontier
               of trees). Second, we provide results from AMT surveys which were used to estimate
               interpretability weights wt for features in our dataset. Third, we provide AMT game re-
               sults to evaluate XplainIM’s effectiveness in explaining HEALER’s solutions to AMT
               workers.
                   Dataset We generate 100 different Watts-Strogatz (WS) networks (p = 0.25, k =
               7), each of size 20 nodes. The influence probabilities for network edges were ran-
               domly sampled from a uniform distribution U [0, 1]. We set K = 2, i.e., every set of
               two network nodes is a possible IM solution. For each of these 100 networks, we add
               HEALER’s solution, along with 20 randomly chosen IM solutions to our dataset with
               the labels “optimal” and “suboptimal”, respectively. Next, we map the ith IM solution
               in our dataset to a unique feature vector X i containing the network-centric feature val-
               ues (see Figure 1) for that IM solution. This gives us a labeled dataset (with 20 features
               and 2100 datapoints), which serves as XplainIM’s input in our experiments.


                                               Feature Name            wt
                                               Betweenness Centrality 0.0
                                               Clustering Coefficient 0.25
                                               Pagerank               0.28
                                               Closeness Centrality   0.71
                                               Degree                  1.0


                                     Fig. 1: Interpretability Weights for Features


                  Interpretability Weights We deployed a survey among 35 AMT workers to esti-
               mate interpretability weights wt for all 20 features in our dataset. Each worker is (i)
               provided the definition of each feature; and (ii) given an example of how that feature’s




                                                           15
Proceedings of the 3rd International Workshop on Social Influence Analysis (SocInf 2017)
August 19th, 2017 - Melbourne, Australia




               value is calculated for a sample IM solution. Then, the worker is asked to calculate each
               feature’s value for a new IM solution on an unseen network.
                   For each feature t, we compare the user response distribution Mt (i.e., histogram
               of feature values calculated by the 21 workers who passed the validation game) with
               the actual value distribution Pt in order to calculate wt . We assume that a feature’s
               interpretability correlates heavily with ease of calculating that feature’s value. Thus,
               if Mt is very similar to Pt , it implies that AMT workers find feature t to be highly
               interpretable. However, significant differences between Mt and Pt implies that fea-
               ture t is uninterpretable. Specifically, we set wt = 1 − EarthMover(Mt , Pt ), where
               EarthMover(Mt , Pt ) measures the distance between the distributions Mt and Pt [3].
                   Due to a lack of space, Figure 1 shows normalized wt values for 5 out of the 20
               features. This figure shows that “Degree” and “Betweenness Centrality” are the most
               and least interpretable features, respectively.
                   AMT Game Results We now present results from two different game types that
               we deployed on AMT to evaluate XplainIM’s effectiveness at explaining HEALER’s
               solutions. Both games begin by asking AMT workers to pick their preferred IM solution
               UG on a randomly chosen WS network (p = 0.25, k = 7, n = 20). We begin by
               explaining the design and motivations of both games.
                   Game 1 This game has three different versions, one for each variant of XplainIM.
               This game presents explanations generated by XplainIM to AMT workers, and evalu-
               ates the effect of these explanations in improving the worker’s IM solutions UG . The
               game proceeds as follows. After the worker picks solution UG , he/she is given an expla-
               nation for why HEALER’s solution OG is better than UG (this explanation is generated
               by XplainIM-{A, B or Full} depending on the game version). Then, the worker is asked
               to improve his solution UG based on the explanation shown.
                   Game 2 This game presents explanations generated by all three XplainIM variants
               to the same AMT worker, and compares the self-assessed interpretabilities of these
               three explanation schemes. The game proceeds as follows. After the worker picks so-
               lution UG , he/she is given three different explanations (generated using XplainIM-{A,
               B & Full}) for why HEALER’s solution OG is better than UG . Then, the worker is
               asked to rate these three explanations (each on a scale of 1-10) in terms of their in-
               terpretability. Further, we also deploy a “No Explanation” game (as a baseline), where
               workers are asked to improve their suboptimal solutions UG , without being given any
               explanation for UG ’s suboptimality. Each of these game versions was deployed with 50
               unique AMT workers (to avoid learning bias), each of whom plays on five different WS
               networks sequentially. We now discuss evaluation metrics used in this section.
                   Evaluation Metrics For Game 1, we compare the interpretability of XplainIM-A,
               B & Full’s explanations by measuring the effect of these explanations in improving the
               worker’s IM solutions UG . Specifically, we measure the percentage of workers who (in
               response to the provided explanation) modified UG to either (i) match OG (Optimal
               group); OR (i) correctly followed the provided explanation (Correct group); OR (iii)
               moved in the right direction (Close group).
                   For example, assume that “Degree” (Figure 1) of UG is 4, and the provided ex-
               planation is “For solutions to be optimal, the Degree should be greater than 9”. If the
               worker modifies UG to match OG , he/she is counted in the Optimal group. Otherwise,




                                                         16
Proceedings of the 3rd International Workshop on Social Influence Analysis (SocInf 2017)
August 19th, 2017 - Melbourne, Australia




                          0.71                                                             0.71

                          0.61                                                             0.61
               F1 Score




                                                                                F1 Score
                          0.51                                                             0.51

                          0.41                                                             0.41

                          0.31                                                             0.31
                              0.95          1.05          1.15          1.25                   0.95        1.05          1.15         1.25
                                               Interpretability Score                                        Interpretability Score

                                     (a) Set of all decision trees                                (b) Trees forming Pareto frontier

                                              Fig. 2: Plotting the Pareto Frontier of Decision Trees
               if his/her modified solution’s Degree is greater than 9, he/she is counted in the Correct
               group. Finally, if his/her modified solution’s Degree is greater than 4 (UG ’s original
               Degree), but is less than 9, then he/she moved UG in the right direction and is counted
               in the Close group.
                   Also, we use the failure rate, which is defined as the fraction of times an explanation
               system fails to produce an explanation. This happens when there are no feasible decision
               trees for a given choice of OG and UG . We calculate an explanation system’s failure
               rate on network G by measuring the fraction of possible user solutions UG for which no
               tree is feasible in the Pareto frontier. We now present results comparing XplainIM-A, B
               & Full with the “No Explanation” treatment. All our comparison results are statistically
               significant under t-test.
                   Pareto Frontier First, we illustrate the accuracy-interpretability tradeoff exploited
               by all XplainIM variants to find Pareto-optimal trees. Figure 2a shows all decision trees
               (depicted by individual points) obtained by varying α values. The X and Y axes show
               the ID and F1 scores, respectively. We prune this set of decision trees to find the set
               of Pareto-optimal trees (shown in Figure 2b). These figures clearly depict the inverse
               correlation between the two competing objectives of prediction accuracy (F1 ) and in-
               terpretability (ID ).
                   Next, we validate our decision of maintaining this Pareto frontier of trees inside
               XplainIM (instead of maintaining just a single Pareto-optimal tree). Figure 4a compares
               the failure rate of XplainIM-A, B & Full to that of single Pareto-optimal trees. The X-
               axis shows trees of increasing prediction accuracy. The Y-axis shows the average failure
               rate (across 100 WS nets). This figure shows that our Pareto-frontier based explanation
               systems (XplainIM-A, B & Full) have an average failure rate of 0.01, i.e., XplainIM
               generate valid explanations in ∼99% of situations. However, the minimum failure rate
               achieved by a single Pareto-optimal tree is 0.27. This shows that using single trees as our
               classifier for Problem 2 (instead of maintaining a Pareto frontier) results in failures 27%
               of the times. This signifies the importance of the Pareto frontier in making XplainIM
               work.
                   XplainIM Results Figure 3a shows Game 2 results. The X-axis shows the three
               XplainIM variants. The Y-axis shows average ratings (scale of 1-10) given by workers to
               different explanation schemes. This figure shows that XplainIM-A & B’s explanations
               average rating was 6.04, as compared to 4.84 for XplainIM-Full (p=0.034). This shows
               that workers found XplainIM-A & B’s explanations to be more interpretable than those
               generated by XplainIM-Full.




                                                                               17
Proceedings of the 3rd International Workshop on Social Influence Analysis (SocInf 2017)
August 19th, 2017 - Melbourne, Australia




                                           7                                                                               60

               Average Subjective Rating
                                           6                                                                               50     No Explanation




                                                                                                     Percentage of users
                                                                                                                                  XplainIM-Full
                                           5
                                                                                                                           40
                                           4
                                                                                                                           30
                                           3
                                                                                                                           20
                                           2
                                           1                                                                               10
                                           0                                                                                0
                                                    XplainIM-A       XplainIM-B      XplainIM-Full                               Optimal            Correct          Close
                                                            Different Explanation Systems                                                  Different Groups of Users

                                                    (a) Subjective rating comparison                                            (b) Effect on modified solutions

                                                                  Fig. 3: Evaluating interpretability of XplainIM-Full

                                           0.6                                                                             70
                                                                      Single Decision Tree                                 60    No Explanation
                                           0.5
                 Average Failure Rate




                                                                                                     Percentage of Users
                                                                      XplainIM                                                   XplainIM-A
                                                                                                                           50
                                           0.4                                                                                   XplainIM-B
                                                                                                                           40
                                           0.3
                                                                                                                           30
                                           0.2                                                                             20
                                           0.1                                                                             10
                                               0                                                                           0
                                                    0.31 0.35 0.41 0.45 0.51 0.55 0.61 0.65 0.71                                 Optimal            Correct          Close
                                                                      F1 score                                                             Different Groups of Users

                                                        (a) Failure rate comparison                                             (b) Effect on modified solutions

                                                               Fig. 4: Frontier Validation & Evaluating XplainIM-A&B
                     Figure 3b shows Game 1 results for XplainIM-Full and the “No Explanation” treat-
                 ment. The X-axis shows the Optimal, Correct, and Close groups. The Y-axis shows the
                 percentage of workers belonging to these groups. This figure shows that only ∼4.6%
                 workers are in the Correct group after seeing XplainIM- Fulls explanations. Surpris-
                 ingly, even without any explanations, 24.6% workers are in the Correct group (p=3e−9 ),
                 which shows that users perform better without any explanations (compared to XplainIM-
                 Full). This shows that XplainIM-Full’s longer explanations (which are based on [2]) are
                 completely uninterpretable to AMT workers (possibly because they overload workers
                 with too much information). As a result, we now focus our attention on XplainIM-A &
                 B.
                     Figure 4b shows Game 1 results for XplainIM-A & B and the “No Explanation”
                 treatment. This figure shows that with XplainIM-A & B’s explanations, ∼8.5% users
                 are in the Optimal group (∼4% with no explanation), 36% are in the Correct group
                 (24% with no explanation), and ∼60% are in the Close group (46% with no explana-
                 tion) (p=0.05). This establishes that XplainIM-A & B’s explanations are far more inter-
                 pretable, as significantly more AMT workers follow these explanations to improve their
                 solutions. Further, this shows the importance of exploiting the accuracy-interpretability
                 tradeoff (used by XplainIM) in generating explanations to solutions of IM algorithms.


                 7                                 Conclusion

                 This paper solves the open problem of explaining solutions of IM algorithms to its
                 users. This paper shows that the problem of designing an explanation system can be




                                                                                                 18
Proceedings of the 3rd International Workshop on Social Influence Analysis (SocInf 2017)
August 19th, 2017 - Melbourne, Australia




               equivalently viewed as a problem of explaining an ML classifier’s predictions. We pro-
               pose XplainIM, a suite of ML based explanation systems which exploit the accuracy-
               interpretability tradeoff to generate natural language explanations. Extensive human
               subject experiments show the effectiveness of XplainIM in explaining IM algorithm
               solutions to human users.


               8    Acknowledgements

               This research was supported by MURI Grant W911NF-11-1-0332.


               References
                1. Borgs, C., Brautbar, M., Chayes, J., Lucier, B.: Maximizing Social Influence in Nearly Opti-
                   mal Time. In: Proceedings of the Twenty-Fifth Annual ACM-SIAM Symposium on Discrete
                   Algorithms. pp. 946–957. SODA ’14, SIAM (2014)
                2. Lakkaraju, H., Bach, S.H., Leskovec, J.: Interpretable decision sets: A joint framework for
                   description and prediction. In: Proceedings of the 22nd ACM SIGKDD International Con-
                   ference on Knowledge Discovery and Data Mining. pp. 1675–1684. ACM (2016)
                3. Ling, H., Okada, K.: An efficient earth mover’s distance algorithm for robust histogram com-
                   parison. IEEE transactions on pattern analysis and machine intelligence 29(5) (2007)
                4. Loh, W.Y.: Classification and regression trees. Wiley Interdisciplinary Reviews: Data Mining
                   and Knowledge Discovery 1(1), 14–23 (2011)
                5. Puterman, M.L.: Markov Decision Processes: Discrete Stochastic Dynamic Programming.
                   John Wiley & Sons (2009)
                6. Quinlan, J.R.: Induction of decision trees. Machine learning 1(1), 81–106 (1986)
                7. Schneier, B.: Trust in man/machine security systems. IEEE Security Privacy 11(5), 96–96
                   (Sept 2013)
                8. Wang, N., Pynadath, D.V., Hill, S.G.: The impact of pomdp-generated explanations on trust
                   and performance in human-robot teams. In: Proceedings of the 2016 International Confer-
                   ence on Autonomous Agents & Multiagent Systems. pp. 997–1005. International Foundation
                   for Autonomous Agents and Multiagent Systems (2016)
                9. Yadav, A., Chan, H., Jiang, A.X., Rice, E., Kamar, E., Grosz, B., Tambe, M.: POMDPs for
                   Assisting Homeless Shelters: Computational and Deployment Challenges. In: International
                   Workshop on Issues with Deployment of Emerging Agent-based Systems (IDEAS) (2016)
               10. Yadav, A., Wilder, B., Petering, R., Rice, E., Tambe, M.: Influence maximization in the field:
                   The arduous journey from emerging to deployed application. In: In Proceedings of the In-
                   ternational Conference on Autonomous Agents and Multi-agent Systems (AAMAS), 2017
                   (2017)




                                                              19