<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Are Contrastive Explanations Useful? ?</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>James Forrest</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Somayajulu Sripada</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Wei Pang</string-name>
          <email>W.Pang@hw.ac.uk</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>George M. Coghill</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Heriot-Watt University</institution>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>University of Aberdeen</institution>
        </aff>
      </contrib-group>
      <abstract>
        <p>From the user perspective (data subjects and data controllers), useful explanations of ML decisions are selective, contrastive and social. In this paper, we describe an algorithm for generating selective and contrastive explanations and experimentally study its usefulness to users. Machine Learning (ML) models are making ever increasing numbers of decisions that impact people's lives. This makes it important to have explanations that enable those who develop and deploy ML models and those who are subject to their decisions to examine these models to ensure that they are e ective and fair. Miller 2017 proposes three desiderata for explanations - explanations should be selective, contrastive and social. Explanations should be contrastive, explaining how other events could have happened rather than explaining the event that did happen. Explanations should be selective, only presenting information relevant to the recipient. Explanations should be social because they are a communication between explainer and recipient and need to respect the recipient's needs. Because operationally deployed ML models could come from an extensive taxonomy of ML models (from decision trees to deep neural networks) an equally large taxonomy of interpretation methods have been developed (Guidotti et al., 2018). Not all these interpretation methods ful l Miller's desiderata. Wachter et al. 2017 propose that Contrastive Explanations best ful l these criteria. In a Contrastive Explanation, the recipient is shown counterfactuals, where the decision model would make di erent decisions, as the explanation. This work is a test of some claims made for Contrastive Explanations. In order to have data professionals make better use of explanations, Kaur et al 2020 suggest that ML explanations have improved use of HCI to aid understanding in users. Especially to promote better alignment of the users' mental models and the conceptual models of the Interpretable Machine Learning (IML) tool and help the users move from re exive to deliberative thinking.</p>
      </abstract>
      <kwd-group>
        <kwd>Interpretable ML</kwd>
        <kwd>Contrastive Explanations</kwd>
        <kwd>XAI</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Introduction</title>
      <p>
        In this paper, we present a Contrastive Explanation technique built on
optaiNet, an optimisation algorithm based on an Immune Inspired Algorithm
        <xref ref-type="bibr" rid="ref9">(Timmis and Edmonds, 2010)</xref>
        that ful ls Miller's desiderata. We evaluate our
Contrastive Explanation technique in comparison to the contrastive tool DiCE (Mot
        <xref ref-type="bibr" rid="ref4">hilal
et al., 2020</xref>
        ), as well as Feature Saliency explanations.
2
      </p>
    </sec>
    <sec id="sec-2">
      <title>Explanation Generation</title>
      <p>A Contrastive Explanation is a collection of counterfactuals; each counterfactual
shows the changes to the original event (fact) that would produce the wanted
event (foil). These counterfactuals are most e ective when selective; the more
selective the explanation, the fewer causes the explanation contains, the simpler
and more comprehensible the explanation. The counterfactuals should produce
foil events that are as near to fact event as possible to be more actionable to the
recipient
2.1</p>
      <p>opt-aiNet
Our counterfactual generation algorithm one of a family of algorithms called
Arti cial Immune Systems (AIS), which use the Immune System as a metaphor
to solve problems. The counterfactuals are generated by the AIS algorithm
optaiNet, an existing algorithm that we have applied to the domain of explanation
generation.</p>
      <p>The features of an opt-aiNet system (illustrated in g 1) are:
{ The counterfactuals in this AIS take the cell in the Immune System as their
metaphor.
{ The counterfactuals (cells) clone themselves with mutations, a mutation
changes the explanation randomly.
{ The distance between a cell and the target is the a nity; cells near the target
have high a nity, and cells far from the target have low a nity. Mutation
varies inversely with the a nity; high-a nity cells mutate little while
lowa nity cells mutate much more.
{ Counterfactuals (cells) interact with each other with high-a nity
counterfactuals suppressing similar counterfactuals with lower a nity. This removal
of similar counterfactuals helps enforce the diversity of counterfactuals.</p>
      <p>
        Pseudocode for opt-aiNet
        <xref ref-type="bibr" rid="ref9">(Timmis and Edmonds, 2010)</xref>
        adapted for
explanation counterfactuals
1: Produce initial population of counterfactuals (cells)
2: repeat
3: repeat
4: Generate N clone counterfactuals for each counterfactual. All must be
classi ed as foil class
5: Mutate each counterfactual in inverse proportion to the a nity of the
parent counterfactual
6: Determine the a nity of all counterfactuals
7: Keep counterfactual with the highest a nity suppress all others
8: until average a nity lower than previous iteration
9: Determine highest a nity counterfactuals, suppress similar lower a nity
counterfactuals
10: Introduce a proportion of new randomly generated counterfactuals
11: until stopping condition met
      </p>
      <p>The algorithm is based on the the Ruby programming code provided by
Brownlee 2012, which was then reimplemented by us in Python.
2.2</p>
      <sec id="sec-2-1">
        <title>DiCE</title>
        <p>
          DiCE is an IML tool that produces counterfactual explanations produced in
(Mot
          <xref ref-type="bibr" rid="ref4">hilal et al., 2020</xref>
          ). Optimising on the proximity of the explanation to the
fact, this algorithm also optimises to maximise the diversity of the set of
explanations it creates.
3
        </p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Experiments</title>
      <p>Our experiments evaluated the explanations of a model's decision. The
evaluation methodology followed that suggested in Ho man et al. 2018 for Explainable
AI (XAI) and Interpretable Machine Learning (IML) evaluations. We used a Test
of Performance, in which the participant's comprehension of the explanation of
the model's decision was determined by a proxy task of reversing an adverse
decision of the model. We used a Test of Satisfaction where the participants
answered questions on their satisfaction with the explanation. We did not use
a Test of Comprehension, which evaluates the participant's mental models,
because these tests are very hard to execute well and are hard to use one test across
di erent types of explanation, whereas our Test of Performance could be used
for all the explanation types.</p>
      <p>The data used in this experiment comes from the lending dataset maintained
by Kaggle at https://www.kaggle.com/wordsforthewise/lending-club, which holds
data about credit decisions. The decision model used TensorFlow to create a
Deep Neural Network.</p>
      <p>The task simulated a situation where a data subject received a negative
prediction and needed to know the changes to their record that would give a
positive prediction. A negative prediction was used in this experiment because
data subjects are presumed to care about negative credit predictions more than
positive credit predictions. And gives a natural task of amending the record to
get a positive result.</p>
      <p>Participants were recruited using Amazon Mechanical Turk. The tests were
conducted between explanation types, with each experiment using one
explanation only. Every participant was shown the same three records with only the
explanations changing between participants.</p>
      <p>The Test of Performance was to change the values of the record so that the
result changed from a refusal of credit where the model scored the record &lt; 0.5
to an acceptance where the model scores 0.5. The participants were asked
to change the record by the smallest amount which would cause the model to
classify it positively. Those participants with the best understanding of the model
should produce a score 0.5 but close to 0.5.</p>
      <p>The Test of Satisfaction was six questions on the user's satisfaction with the
explanation rated on a ve-point Likert scale.</p>
      <p>The questions were:
1. I understand this explanation of how the decision was made.
2. This explanation of how the decision was made is satisfying.
3. The explanation of how the decision was made has su cient detail.
4. This explanation of how the decision was made contains irrelevant details.
5. The record is useful to the goal of having the loan application accepted
6. This explanation lets me judge the trustworthiness of the explanation.</p>
      <p>
        Further evaluation was done using two Feature Saliency explanations and a
control of no explanation. Feature Saliency explanations for tabular data produce
a set of pairs of input Features with their Importance. The tools used to generate
these Feature Saliency explanations were LIME
        <xref ref-type="bibr" rid="ref8">(Ribeiro et al., 2016)</xref>
        and SHAP
        <xref ref-type="bibr" rid="ref10 ref5">(Lundberg and Lee, 2017)</xref>
        , two very popular tools in IML explanation. There
were two presentations used of the Feature Saliency explanation type: a bar-chart
and a Natural Language Generation (NLG) text. Both presentations have a table
showing the feature/importance pairs. With an additional element of either the
same data presented as a bar-chart, or a text description. The participants who
received No Explanation as a control only saw the data in the record.
      </p>
    </sec>
    <sec id="sec-4">
      <title>Results</title>
      <sec id="sec-4-1">
        <title>Test of Performance</title>
        <p>The Test of Performance was conducted on whether the participant's alterations
to the record successfully produced a change to the models decision from declined
to accepted (Correct) or not (Incorrect), and evaluated using the Chi-Squared
test. The results are shown in Table 1. Contrastive Explanations show no
signi cant di erence in whether opt-aiNet or DiCE generate the counterfactuals or
whether one or three counterfactuals were used in the explanation. There is no
signi cant di erence in performance when the best performing Contrastive
explanantion (opt-aiNet with three counterfactuals) is compared to other types of
explanations. Comparing the explanation to `No Explanation' where the
participants are only shown the values in the record as a control explanation shows that
this has the same performance as any given explanation. There are no signi cant
di erences in performance because this is preliminary work with small numbers
of participants. Future work might involve doing these tests of explanations with
greater numbers of participants to achieve signi cant results.</p>
        <p>Explanations
opt-aiNet with
one counterfactual</p>
        <p>DiCE with
one counterfactual</p>
        <p>opt-aiNet with
three counterfactuals</p>
        <p>DiCE with
three counterfactuals</p>
        <p>Total Correct Incorrect % Correct signi cance from 2 test</p>
        <p>with opt-aiNet with 3cf
30 19 11 63% 1
27
30
30
16
20
19
11
10
11
59%
67%
63%
0.59
NA
1
The results for the Test of Satisfaction are shown in Table 2, which shows the
mean values for the six ve-point Likert scale questions. The participants had
higher satisfaction ratings for Contrastive Explanations with three
counterfactuals rather than one counterfactual. The satisfaction ratings are similar for
both counterfactual generation tools. The opt-aiNet with three counterfactuals
has similar satisfaction ratings to the Feature Saliency explanations with either</p>
        <p>Explanation Type</p>
        <p>opt-aiNet with
one counterfactual</p>
        <p>DiCE with
one counterfactual</p>
        <p>opt-aiNet with
three counterfactuals</p>
        <p>DiCE with
three counterfactuals</p>
        <p>Q 1 Q 2 Q 3 Q 4 Q 5 Q 6
3.93 3.59 * 3.67 3.33 3.77 3.47
bar-chart or NLG. Further, the control `No Explanation' has lower satisfaction
ratings than these three explanations.</p>
        <p>The signi cance of the results was evaluated using Mann-Whitney, compared
to the opt-aiNet with three counterfactuals. The signi cant di erences were for
the second question, `This explanation of how the decision was made is
satisfying', where No Explanation and Contrastive Explanations with one
counterfactual had lower ratings.
5</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>Discussion</title>
      <p>The Test of Performance produced two unexpected results. Firstly, that
Contrastive Explanation with one counterfactual performed unexpectedly poorly
compared to other explanations. Because if the record is changed according to
the values in the counterfactual, then the classi cation will change. Secondly, the
participants who received the `No Explanation' of just the values in the record
performed as well at the task as those who received an explanation, even though
their satisfaction ratings were lower.</p>
      <p>Examining why this might be, gure 2 shows box-plots of the decision
models scores of the participant's records. The experiment instructions asked
participants to make the smallest changes to the record that would give a positive
classi cation. For the Contrastive Explanations the box-plots show the decision
model scores are close to the decision boundary at 0.5, but for the other
explanations the average score is further from the decision boundary and have a larger
range of scores. This shows the Contrastive Explanations produce more optimal
results, whereas, Feature Saliency explanations produce explanations that are
still e ective at changing the classi cation, but do so in a less optimal way (the
score is further from the decision boundary).</p>
      <p>The `No Explanation' is e ective at changing the classi cation but poor at
producing an optimal result, as the average score is the furthest from the
decision boundary. The `No Explanation' is showing the participants own mental
models of the lending domain without being altered by IML explanations. Most
participants who saw the `No Explanation' have a good Mental Model about how
to change a record to get credit approval e.g. have a high income and good credit
grade. But both participants who saw a Feature Saliency (bar-chart and NLG)
explanation and `No explanation' did not know how much to change the record
by to change the record by to get a di erent classi cation as this information is
not given in the explanations.
This preliminary work shows that the opt-aiNet algorithm produces
counterfactuals as good as the established algorithm DiCE. Contrastive Explanations
of the model are not necessarily more e ective than Feature Saliency
explanations or just showing the participants the data in the record used by the ML
model. Contrastive Explanations do enable the participants to make more
optimal changes to the record, that are nearer to the decision boundary. Participants
have their own mental models of how well known domains, this allows them to
be e ective at changing the decisions of classi ers without use of IML
explanations. Future work will require more data and experiments to discover what the
e ects of explanations of ML decisions are on users.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          <string-name>
            <given-names>J. Brownlee. Clever</given-names>
            <surname>Algorithms</surname>
          </string-name>
          .
          <year>2012</year>
          .
          <article-title>ISBN 9781446785065</article-title>
          . URL http://www. cleveralgorithms.com.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          <string-name>
            <given-names>R.</given-names>
            <surname>Guidotti</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Monreale</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Ruggieri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Turini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Pedreschi</surname>
          </string-name>
          , and
          <string-name>
            <given-names>F.</given-names>
            <surname>Giannotti</surname>
          </string-name>
          .
          <article-title>A Survey Of Methods For Explaining Black Box Models</article-title>
          .
          <volume>51</volume>
          (
          <issue>5</issue>
          ),
          <year>2018</year>
          . ISSN 03600300. https://doi.org/10.1145/3236009. URL http://arxiv.org/abs/
          <year>1802</year>
          .
          <year>01933</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          <string-name>
            <given-names>R. R.</given-names>
            <surname>Ho</surname>
          </string-name>
          man, S. T. Mueller, G. Klein, and
          <string-name>
            <given-names>J.</given-names>
            <surname>Litman</surname>
          </string-name>
          .
          <article-title>Metrics for Explainable AI: Challenges and Prospects</article-title>
          . pages
          <fpage>1</fpage>
          {
          <fpage>50</fpage>
          ,
          <year>2018</year>
          . URL https://arxiv.org/abs/
          <year>1812</year>
          .04608.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          <string-name>
            <given-names>H.</given-names>
            <surname>Kaur</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Nori</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Jenkins</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Caruana</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Wallach</surname>
          </string-name>
          , and
          <string-name>
            <given-names>J. Wortman</given-names>
            <surname>Vaughan</surname>
          </string-name>
          . Interpreting Interpretability:
          <article-title>Understanding Data Scientists' Use of Interpretability Tools for Machine Learning</article-title>
          .
          <source>CHI Conference on Human Factors in Computing Systems Proceedings</source>
          , pages
          <volume>1</volume>
          {
          <fpage>14</fpage>
          ,
          <year>2020</year>
          . https://doi.org/10.1145/3313831.3376219.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          <string-name>
            <given-names>S.</given-names>
            <surname>Lundberg</surname>
          </string-name>
          and
          <string-name>
            <given-names>S.-I.</given-names>
            <surname>Lee</surname>
          </string-name>
          . A Uni ed Approach to Interpreting Model Predictions.
          <year>2017</year>
          .
          <article-title>ISSN 10495258</article-title>
          . URL https://arxiv.org/abs/1705.07874.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          <string-name>
            <given-names>T.</given-names>
            <surname>Miller</surname>
          </string-name>
          .
          <article-title>Explanation in Arti cial Intelligence: Insights from the Social Sciences</article-title>
          .
          <year>2017</year>
          . https://doi.org/arXiv:
          <fpage>1706</fpage>
          .07269v1. URL http://arxiv.org/abs/1706. 07269.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          <string-name>
            <surname>R. K. Mothilal</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Sharma</surname>
            , and
            <given-names>C. Tan.</given-names>
          </string-name>
          <article-title>Explaining machine learning classi ers through diverse counterfactual explanations</article-title>
          .
          <source>In Proceedings of the 2020 Conference on Fairness, Accountability, and Transparency</source>
          , pages
          <volume>607</volume>
          {
          <fpage>617</fpage>
          , New York, NY, USA,
          <fpage>1</fpage>
          <lpage>2020</lpage>
          . ACM. ISBN 9781450369367. https://doi.org/10.1145/3351095.3372850. URL http://arxiv.org/abs/
          <year>1905</year>
          . 07697http://dl.acm.org/doi/10.1145/3351095.3372850.
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          <string-name>
            <surname>M. T. Ribeiro</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Singh</surname>
            , and
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Guestrin</surname>
          </string-name>
          .
          <article-title>"Why Should I Trust You?": Explaining the Predictions of Any Classi er</article-title>
          .
          <year>2016</year>
          . ISSN 9781450321389. https://doi.org/10.1145/1235. URL http://arxiv.org/abs/1602.04938.
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          <string-name>
            <given-names>J.</given-names>
            <surname>Timmis</surname>
          </string-name>
          and
          <string-name>
            <given-names>C.</given-names>
            <surname>Edmonds</surname>
          </string-name>
          .
          <article-title>A Comment on Opt-AiNET: An Immune Network Algorithm for Optimisation.</article-title>
          (May):
          <volume>308</volume>
          {
          <fpage>317</fpage>
          ,
          <year>2010</year>
          . https://doi.org/10.1007/978-3-
          <fpage>540</fpage>
          -24854-5
          <fpage>32</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          <string-name>
            <given-names>S.</given-names>
            <surname>Wachter</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Mittelstadt</surname>
          </string-name>
          , and
          <string-name>
            <given-names>L.</given-names>
            <surname>Floridi</surname>
          </string-name>
          .
          <article-title>Why a right to explanation of automated decision-making does not exist in the General Data Protection Regulation</article-title>
          .
          <source>International Data Privacy Law</source>
          , pages
          <volume>1</volume>
          {
          <fpage>47</fpage>
          ,
          <year>2017</year>
          . URL http://arxiv.org/abs/1606.08813.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>