<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Beyond One-Size-Fits-All: How User Objectives Shape Counterfactual Explanations</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Orfeas Menis Mastromichalakis</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Jason Liartis</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Giorgos Stamou</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Artificial Intelligence and Learning Systems Laboratory, National Technical University of Athens</institution>
        </aff>
      </contrib-group>
      <abstract>
        <p>Counterfactual Explanations (CFEs) have emerged as a powerful tool for interpreting machine learning models by illustrating alternative scenarios where key factors difer. Despite their widespread adoption, the existing literature often overlooks the diverse needs and objectives of users across various domains-ranging from guiding decision-making to understanding model behavior and assessing robustness-and their implications for CFE characteristics. As a result, CFEs frequently fail to adequately address these distinct use cases, leading to suboptimal and sometimes misleading explanations. In this paper, we advocate for a more user-centered approach to CFEs, emphasizing the importance of aligning their characteristics with user objectives. We identify three primary use cases-actionable user guidance, system understanding, and vulnerability assessment-and examine the desired properties of CFEs in each case. By addressing these diferences, we aim to inform the design and deployment of more efective, context-aware explanations that meet the unique needs of users while ensuring the accuracy and usefulness of the insights provided.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Explainable Artificial Intelligence (XAI) has become increasingly indispensable as AI systems
are integrated into various facets of society. It aims to elucidate the opaque decision-making
processes of machine learning algorithms, ofering transparency and useful insights, and
fostering trust among multiple stakeholders from AI engineers to end-users. Counterfactual
Explanations, grounded in research from philosophy [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] and psychology on counterfactual
thinking [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ], explore why a particular outcome occurred by examining alternative scenarios
where key factors or events were diferent. In artificial intelligence, these explanations provide
insight into how alterations in input features would have influenced an AI system’s output,
thereby facilitating a deeper understanding of its behavior and decision-making mechanisms.
Due to their contrastive nature, CFEs are close to how humans perceive explanations [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ],
facilitating an intuitive way to study a model’s behavior.
      </p>
      <p>
        While numerous existing works delve into the desired characteristics of counterfactual
explanations [
        <xref ref-type="bibr" rid="ref4 ref5 ref6 ref7">4, 5, 6, 7</xref>
        ] and how to evaluate them [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ], they often approach them with a unified
strategy encompassing multiple objectives including detecting biases, providing actionable
recourse, increasing trust, and enhancing understandability. However, this approach
overlooks the fact that the diverse needs and objectives of users across various applications and
domains necessitate diferent properties for counterfactual explanations [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]. Consequently,
the explanations generated may fail to adequately address all use cases, as the requirements
for one user’s target could directly conflict with another’s. This phenomenon of conflicting
motivations and ambiguity of objectives in XAI has been thoroughly discussed in literature
[
        <xref ref-type="bibr" rid="ref10 ref11">10, 11</xref>
        ]. Our work aligns with numerous works that call for contextualized design, development,
and evaluation of explanations in terms of application [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ], target audience [
        <xref ref-type="bibr" rid="ref13 ref14">13, 14, 15, 16</xref>
        ], and
end-goal [
        <xref ref-type="bibr" rid="ref6">17, 6</xref>
        ]. Therefore, there is a pressing need to distinguish between the distinct use cases
in which counterfactual explanations are applied. By doing so, we can pinpoint the specific
desired characteristics for each use case, rather than attempting to create counterfactuals that
aim to cover all scenarios, which is inherently infeasible.
      </p>
      <p>
        In this paper, we advocate for a more nuanced understanding of counterfactual explanations,
recognizing that their desired properties vary significantly depending on user objectives and
target applications. Building on existing literature [
        <xref ref-type="bibr" rid="ref5 ref6">5, 6</xref>
        ], we identify three primary scenarios
based on user objectives and examine the key properties that CFEs should exhibit in each case.
The first case concerns actionable user guidance, where CFEs assist users in decision-making by
providing actionable guidance on how to modify inputs to achieve a specific desired outcome.
The second is system understanding, a central focus in explainable AI, where CFEs are used to
analyze model behavior across real-world inputs, ofering insights into the model’s biases and
intricate decision-making mechanisms. The third is vulnerability assessment, which leverages
CFEs to identify weaknesses in a model, such as susceptibility to input perturbations, assessing
its robustness. While CFEs can support all these objectives, their efectiveness depends on
how well they align with the specific needs of each use case. For instance, CFEs designed to
guide user actions may fail to reveal deeper insights into a model’s reasoning, whereas those
tailored for system investigation may lack actionable recommendations. By acknowledging
these distinctions, we can develop more targeted and efective explanations that empower users
across diferent contexts, ultimately fostering more meaningful human-AI collaboration.
      </p>
    </sec>
    <sec id="sec-2">
      <title>2. Preliminaries</title>
      <p>In this section, we explain and briefly discuss some concepts that we use throughout this
paper. We assume that there is an AI system  under examination that accepts an input 
and produces an output  () = . For example, the AI system might be an automated loan
approval system deployed by a bank. In this case,  would be a client’s application that
might include information such as age, gender, occupation, income, number of existing loans,
etc. and  would be the AI system’s decision, approve or reject. A counterfactual instance
of  is a new input ′(̸= ) that produces a diferent output ′(̸= ). ′ is also generally
expected to minimize some metric  among all inputs that produce a diferent output, i.e.,
′ = argmin{ (, ) |  () ̸= }. We further discuss this in the following paragraph. The

system that produces counterfactuals instances, may be referred to as the counterfactual
explainer or counterfactual editor. The diferences between  and ′ are also referred to as the
counterfactual edit. The counterfactual edit indicates what changes need to be made to the
input to receive a diferent output. In the automated loan approval example, a counterfactual
edit may suggest to a user with a rejected application, that they need to change occupation or
pay of one of their existing loans for their application to be approved.</p>
      <p>Although, in principle, any instance ′ ̸=  with ′ ̸=  can be considered as a
counterfactual, most approaches require ′ to be close to  by some metric  , which ofers
two key benefits. First, it ensures that the counterfactual edit—the diferences between
 and ′—is minimal, making it easier to interpret which specific changes influence the
model’s output. This helps isolate the impact of diferent input features. Second, it
transforms the search for a suitable counterfactual into an optimization problem, where the
objective is to find the closest possible ′ with ′ ̸=  leveraging heuristics or established
optimization techniques. It is crucial to note that the choice of metric used to calculate the
distance between  and ′ is not merely a technical detail; it fundamentally shapes the
characteristics of CFEs and is closely tied to the diferent scenarios and use cases analyzed in this paper.</p>
      <p>
        Although valid, some counterfactual edits may lead to instances that are highly uncommon
or even unattainable in the real world. For example, it is rare for a teacher to earn an income
of $1 million, and it is impossible for someone to have an age of -1. Therefore, assessing
whether a counterfactual instance is coherent with a reference population—such as real-world
distributions—is crucial, as its plausibility can significantly impact its usefulness in diferent
contexts. Counterfactual instances that align with a reference population are referred to as
plausible or feasible [
        <xref ref-type="bibr" rid="ref6">18, 6</xref>
        ]. Plausible instances are also said to be close to the data manifold [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ],
or in-distribution, while implausible instances are said to be far from the data manifold and
out-of-distribution (OOD).
      </p>
      <p>
        However, even if a counterfactual instance is plausible, it may still be unattainable from the
original instance due to constraints on the actions required to achieve it. For example, while
any date could be a valid birth date, an individual cannot change their own birth date. Thus, it
is important to assess whether a plausible counterfactual instance can actually be reached from
the original input through real-world actions. Counterfactual edits that modify only mutable
features are referred to as actionable [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. An actionable edit consists of a set of feasible changes
that can be applied to the input instance to attain the counterfactual instance. While some
features, such as date of birth, are inherently immutable, others, like occupation or income,
may be modifiable in principle, though the feasibility of such modifications can vary across
individuals. Given this variability, it is desirable for an explainer to account for user-specific
constraints when determining actionability, ensuring that counterfactual suggestions align
with real-world possibilities.
      </p>
      <p>
        The literature discusses various properties and characteristics beyond plausibility and
actionability that also influence the efectiveness and applicability of CFEs. For instance, counterfactual
edits that involve modifying only a few features are referred to as sparse [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ], a property often
desired for clarity and interpretability. However, in this work, we focus primarily on
actionability and plausibility, as these are the key diferentiating factors among the three scenarios
analyzed in the following section.
      </p>
    </sec>
    <sec id="sec-3">
      <title>3. Use Cases and User Objectives</title>
      <p>In this section, we analyze three distinct user-objective-driven use cases of counterfactual
explanations, focusing on the properties of actionability and plausibility. Each use case represents
a real-world application scenario with diferent goals and constraints. Specifically, we examine
three key scenarios:</p>
      <sec id="sec-3-1">
        <title>1. when an end-user seeks guidance on achieving a desired outcome,</title>
        <p>2. when a user aims to investigate and understand an AI system’s behavior—such as detecting
biases, flaws, or inconsistencies on real distributions,
3. and when a user attempts to identify system vulnerabilities and strengthen defenses
against potential attacks.</p>
        <p>As demonstrated in Table 1, actionability and plausibility are not universally desirable properties.
Instead, their relevance depends on user objectives, as they may sometimes impose constraints
that limit insights or hinder the efectiveness of the explanation. It is important to note that the
absence of a checkmark in the table does not imply that counterfactuals with these properties
are excluded. Rather, it reflects an opt-in approach, where a checkmark indicates that all
counterfactuals in that use case must exhibit the given property, rather than an opt-out approach,
where their absence would mean such counterfactuals should never be considered. For instance,
while plausible and actionable counterfactuals remain relevant in vulnerability assessment,
relying only on them would overlook important failure modes of the system. As analyzed in the
following subsections, the issue in later use cases arises when counterfactuals are constrained
to those exhibiting these properties, not from their presence itself.</p>
        <sec id="sec-3-1-1">
          <title>3.1. Actionable User Guidance</title>
          <p>
            In this use case, also known as (actionable) recourse [
            <xref ref-type="bibr" rid="ref5">5, 19, 20</xref>
            ], the end-user seeks guidance on
modifying input features to achieve a desired outcome. For instance, in the automated loan
approval system example, a bank customer whose application was rejected may want to know
what changes could increase their chances of approval. This is the most restrictive use case, as it
requires counterfactual explanations to be both actionable and plausible. Actionability ensures
that suggested modifications involve only mutable features, while plausibility guarantees
that these modifications are realistic within the reference population. Providing impractical
suggestions—such as changing one’s birthplace or attaining an impossible state, like a negative
age—would make the CFE unhelpful, as the user would be unable to act on it. The goal is to
ofer meaningful and feasible recommendations that align with real-world constraints.
          </p>
          <p>The choice of minimality metric also plays a crucial role in optimizing CFEs for this use
case. Some works advocate for the use of a sparsity-inducing norm such as 0 or 1, since it
provides the user with a small set of specific goals [ 21, 22]. Sometimes, these norms result in a
counterfactual edit with a big change to a single feature, which might be less attainable than a
few smaller changes. Other norms such as 2 and ∞ along with a sparsity-inducing penalty
might be more suitable. A diferent approach focuses on the cost of real-world actions. Instead
of evaluating minimality with regard to diferences between the two diferent instances  and
′, they work on a higher-level space of actions that a user needs to take in order to arrive
at ′ starting from , simultaneously addressing issues of cause and efect [ 23, 24, 25]. In the
example of the loan approval system, a counterfactual edit suggesting a change of occupation
may be irrelevant if it does not refer to a set of actions necessary to change occupations and
ignores that other features may have to change as a result of those actions, such as income and
education level.</p>
        </sec>
        <sec id="sec-3-1-2">
          <title>3.2. System Understanding</title>
          <p>In this use case, the user seeks to analyze and understand the behavior of the AI system
within the domain of in-distribution data. Counterfactual explanations help uncover potential
biases and inconsistencies in the model’s decision-making. In this context, plausibility is
crucial because the goal is to analyze the model’s decision-making within realistic, real-world
distributions. Allowing non-plausible counterfactuals could introduce noise and non-sensical
edits, potentially obscuring meaningful biases among irrelevant or unrealistic changes.
Conversely, if we were to restrict ourselves only to actionable edits, we would miss biases
related to immutable features—such as gender or ethnicity—which are often at the core
of discriminatory behavior in AI systems. Many real-world biases manifest in decisions
that treat individuals diferently based on characteristics they cannot change, so excluding
counterfactuals that vary these attributes would prevent us from fully diagnosing such issues.
For instance, an AI engineer developing a loan approval or candidate screening system may
investigate whether the model exhibits biases related to sensitive attributes such as gender
or skin color. The engineer may also assess whether the model’s decisions align with human
intuition and logical reasoning. For example, if a counterfactual edit suggests that reducing
one’s income would increase the likelihood of loan approval, it would be counter-intuitive,
possibly indicating an unexpected flaw in the system’s behavior. Presenting a multitude of
counterfactuals is very important in this use case since their contrast can reveal biases. For
instance, one counterfactual edit might suggest an increase of income by $10,000 to secure loan
approval, while another might suggest an increase of income by $1,000 and a change of gender.
This looser income requirement for a diferent gender would reveal a potential gender bias.</p>
          <p>In such cases, the choice of minimality metric depends on the focus of the investigation.
When stress-testing for specific biases, it is often beneficial to isolate certain sensitive features or
heavily penalize non-sensitive alterations. This can be achieved through proper customization
of the minimality metric. Conversely, when the goal is a broader understanding of model
behavior, norm-based metrics such as 1 or 2 are commonly used.</p>
        </sec>
        <sec id="sec-3-1-3">
          <title>3.3. Vulnerability Assessment</title>
          <p>In this use case, the user aims to identify potential weaknesses or vulnerabilities in the
AI system. Counterfactual explanations serve as a tool to assess the model’s robustness
against small perturbations or out-of-distribution inputs. For instance, a security engineer
may want to test whether slight modifications to input data—such as leaving fields empty,
providing invalid values, or introducing minor inconsistencies—could compromise the system’s
integrity. The primary emphasis here lies on robustness, where considerations of plausibility
and actionability pose potential conflicts with the user’s objectives as they could hinder
the detection of vulnerabilities involving random noise or out-of-distribution permutations.
Imperceptible changes to the input that significantly alter the output, commonly known as
adversarial examples [26] also fall under this category. Although typically treated as a distinct
concept, sometimes explicitly suppressed by counterfactual explainers, they are, in essence, a
specific case of counterfactual edits. Rather than viewing them as a separate phenomenon, it is
preferable to specify whether a given counterfactual explainer accommodates this use case.</p>
          <p>For such robustness assessments, norm-based metrics are generally preferred for minimality,
as they focus on small perturbations that influence the output without necessarily having
semantic meaning. Counterfactual edits with low norm minimality are particularly useful, as
they identify points on the data manifold that are very close to the original input, allowing
for a fine-grained evaluation of the system’s resilience. However, the choice of norm metric
still impacts the nature of the generated counterfactuals. For example, in the domain of image
classifiers, some editors employ the 0 norm to change a single, or very few, pixels [27], while
others use the 2 norm to produce a set of many tiny changes that are imperceptible to the
human eye [28]. Both are valid approaches with diferent implications for robustness testing.
To comprehensively assess a system’s vulnerabilities, a variation of such metrics should be
considered to ensure that diferent types of adversarial weaknesses are adequately explored.</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Conclusions</title>
      <p>In this paper, we underscore the importance of a nuanced understanding of counterfactual
explanations. By recognizing the variability in desired properties based on user objectives and
target applications, we have advocated for a tailored approach to the design and development
of CFEs. Our analysis of three main user objectives and their relation to the key concepts
of plausibility and actionability has revealed that the desired characteristics of CFEs difer
significantly depending on the end task, highlighting the necessity of considering user needs
in the explanation process. Through this study, we have demonstrated the limitations of a
one-size-fits-all approach to CFEs and emphasized the need for customized explanations that
address the specific requirements of users across diverse scenarios.</p>
    </sec>
    <sec id="sec-5">
      <title>Acknowledgments</title>
      <p>Funding for the work was received from the European Commission Directorate-General for
Communications Networks, Content and Technology grant no. 101135809 (EXTRA-BRAIN).</p>
    </sec>
    <sec id="sec-6">
      <title>Declaration on Generative AI</title>
      <sec id="sec-6-1">
        <title>The author has not employed any Generative AI tools.</title>
        <p>for explanations of black-box classifiers in the space of semantic queries, Semantic Web
(2023) 1–42.
[15] O. M. Mastromichalakis, E. Dervakos, A. Chortaras, G. Stamou, Rule-based explanations
of machine learning classifiers using knowledge graphs, in: Proceedings of the AAAI
Symposium Series, volume 3, 2024, pp. 193–202.
[16] S. Dhanorkar, C. T. Wolf, K. Qian, A. Xu, L. Popa, Y. Li, Who needs to know what, when?:
Broadening the explainable ai (xai) design space by looking at explanations across the ai
lifecycle, in: Proceedings of the 2021 ACM Designing Interactive Systems Conference,
2021, pp. 1591–1602.
[17] C. T. Wolf, Explainability scenarios: towards scenario-based xai design, in: Proceedings of
the 24th International Conference on Intelligent User Interfaces, 2019, pp. 252–257.
[18] R. Poyiadzi, K. Sokol, R. Santos-Rodriguez, T. De Bie, P. Flach, Face: feasible and actionable
counterfactual explanations, in: Proceedings of the AAAI/ACM Conference on AI, Ethics,
and Society, 2020, pp. 344–350.
[19] P. M. VanNostrand, D. M. Hofmann, L. Ma, E. A. Rundensteiner, Actionable recourse
for automated decisions: Examining the efects of counterfactual explanation type and
presentation on lay user understanding, in: Proceedings of the 2024 ACM Conference
on Fairness, Accountability, and Transparency, FAccT ’24, Association for Computing
Machinery, New York, NY, USA, 2024, p. 1682–1700. URL: https://doi.org/10.1145/3630106.
3658997. doi:10.1145/3630106.3658997.
[20] J. Yetukuri, I. Hardy, Y. Liu, Towards user guided actionable recourse, in: Proceedings
of the 2023 AAAI/ACM Conference on AI, Ethics, and Society, AIES ’23, Association for
Computing Machinery, New York, NY, USA, 2023, p. 742–751. URL: https://doi.org/10.1145/
3600211.3604708. doi:10.1145/3600211.3604708.
[21] C. Russell, Eficient search for diverse coherent explanations, in: Proceedings of the
Conference on Fairness, Accountability, and Transparency, FAT* ’19, Association for
Computing Machinery, New York, NY, USA, 2019, p. 20–28. URL: https://doi.org/10.1145/
3287560.3287569. doi:10.1145/3287560.3287569.
[22] S. Wachter, B. Mittelstadt, C. Russell, Counterfactual explanations without opening the
black box: Automated decisions and the gdpr, Harv. JL &amp; Tech. 31 (2017) 841.
[23] R. Crupi, A. Castelnovo, D. Regoli, B. San Miguel Gonzalez, Counterfactual explanations as
interventions in latent space, Data Mining and Knowledge Discovery 38 (2024) 2733–2769.
[24] G. Ramakrishnan, Y. C. Lee, A. Albarghouthi, Synthesizing action sequences for modifying
model decisions, Proceedings of the AAAI Conference on Artificial Intelligence 34 (2020)
5462–5469. doi:10.1609/aaai.v34i04.5996.
[25] A.-H. Karimi, B. Schölkopf, I. Valera, Algorithmic recourse: from counterfactual
explanations to interventions, 2020. URL: https://arxiv.org/abs/2002.06278. arXiv:2002.06278.
[26] I. J. Goodfellow, J. Shlens, C. Szegedy, Explaining and harnessing adversarial examples,
arXiv preprint arXiv:1412.6572 (2014).
[27] J. Su, D. V. Vargas, K. Sakurai, One pixel attack for fooling deep neural networks, IEEE</p>
        <p>Transactions on Evolutionary Computation 23 (2019) 828–841.
[28] S.-M. Moosavi-Dezfooli, A. Fawzi, P. Frossard, Deepfool: a simple and accurate method to
fool deep neural networks, in: Proceedings of the IEEE conference on computer vision
and pattern recognition, 2016, pp. 2574–2582.</p>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>D.</given-names>
            <surname>Lewis</surname>
          </string-name>
          , Counterfactuals, John Wiley &amp; Sons,
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>R. M.</given-names>
            <surname>Byrne</surname>
          </string-name>
          , Counterfactual thought,
          <source>Annual review of psychology 67</source>
          (
          <year>2016</year>
          )
          <fpage>135</fpage>
          -
          <lpage>157</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>T.</given-names>
            <surname>Miller</surname>
          </string-name>
          ,
          <article-title>Explanation in artificial intelligence: Insights from the social sciences</article-title>
          ,
          <source>Artificial Intelligence</source>
          <volume>267</volume>
          (
          <year>2019</year>
          )
          <fpage>1</fpage>
          -
          <lpage>38</lpage>
          . URL: https://www.sciencedirect.com/science/article/pii/ S0004370218305988. doi:
          <volume>10</volume>
          .1016/j.artint.
          <year>2018</year>
          .
          <volume>07</volume>
          .007.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>S.</given-names>
            <surname>Verma</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Boonsanong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Hoang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K. E.</given-names>
            <surname>Hines</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. P.</given-names>
            <surname>Dickerson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Shah</surname>
          </string-name>
          ,
          <article-title>Counterfactual explanations and algorithmic recourses for machine learning: A review</article-title>
          , arXiv preprint arXiv:
          <year>2010</year>
          .
          <volume>10596</volume>
          (
          <year>2020</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>A.</given-names>
            <surname>Karimi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Barthe</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Schölkopf</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Valera,</surname>
          </string-name>
          <article-title>A survey of algorithmic recourse: definitions, formulations, solutions, and prospects</article-title>
          . corr abs/
          <year>2010</year>
          .04050 (
          <year>2020</year>
          ), publications at ACM Computing Surveys (
          <year>2020</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>R.</given-names>
            <surname>Guidotti</surname>
          </string-name>
          ,
          <article-title>Counterfactual explanations and how to find them: literature review and benchmarking, Data Mining and Knowledge Discovery (</article-title>
          <year>2022</year>
          )
          <fpage>1</fpage>
          -
          <lpage>55</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>S.</given-names>
            <surname>Verma</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Dickerson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Hines</surname>
          </string-name>
          ,
          <article-title>Counterfactual explanations for machine learning: Challenges revisited</article-title>
          ,
          <source>arXiv preprint arXiv:2106.07756</source>
          (
          <year>2021</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>G.</given-names>
            <surname>Filandrianos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Dervakos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Menis-Mastromichalakis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Zerva</surname>
          </string-name>
          , G. Stamou,
          <article-title>Counterfactuals of counterfactuals: a back-translation-inspired approach to analyse counterfactual editors</article-title>
          ,
          <source>arXiv preprint arXiv:2305.17055</source>
          (
          <year>2023</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>O. M.</given-names>
            <surname>Mastromichalakis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Liartis</surname>
          </string-name>
          , G. Stamou,
          <article-title>Beyond one-size-fits-all: Adapting counterfactual explanations to user objectives</article-title>
          ,
          <source>arXiv preprint arXiv:2404.08721</source>
          (
          <year>2024</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>Z. C.</given-names>
            <surname>Lipton</surname>
          </string-name>
          ,
          <article-title>The mythos of model interpretability: In machine learning, the concept of interpretability is both important and slippery</article-title>
          .,
          <source>Queue</source>
          <volume>16</volume>
          (
          <year>2018</year>
          )
          <fpage>31</fpage>
          -
          <lpage>57</lpage>
          . URL: https: //doi.org/10.1145/3236386.3241340. doi:
          <volume>10</volume>
          .1145/3236386.3241340.
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>C.</given-names>
            <surname>Rudin</surname>
          </string-name>
          ,
          <article-title>Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead</article-title>
          ,
          <source>Nature Machine Intelligence</source>
          <volume>1</volume>
          (
          <year>2019</year>
          )
          <fpage>206</fpage>
          -
          <lpage>215</lpage>
          . doi:
          <volume>10</volume>
          .1038/s42256-019-0048-x.
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>M.</given-names>
            <surname>Norkute</surname>
          </string-name>
          ,
          <article-title>Ai explainability: Why one explanation cannot fit all</article-title>
          ,
          <source>in: ACM CHI Workshop on Operationalizing Human-Centered Perspectives in Ex-plainable AI (HCXAI)</source>
          ,
          <year>2021</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>K.</given-names>
            <surname>Sokol</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Flach</surname>
          </string-name>
          ,
          <article-title>One explanation does not fit all: The promise of interactive explanations for machine learning transparency</article-title>
          ,
          <source>KI-Künstliche Intelligenz</source>
          <volume>34</volume>
          (
          <year>2020</year>
          )
          <fpage>235</fpage>
          -
          <lpage>250</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>J.</given-names>
            <surname>Liartis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Dervakos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Menis-Mastromichalakis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Chortaras</surname>
          </string-name>
          , G. Stamou, Searching
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>