<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Smart move? A case study on the transparency and explainability in an transition algorithm</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Chantale Tippett</string-name>
          <email>Chantale.Tippett@nesta.org.uk</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Joint Proceedings of the ACM IUI 2021 Workshop</institution>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Nesta</institution>
          ,
          <addr-line>58 Victoria Embankment, London, EC4Y 0DS</addr-line>
          ,
          <country country="UK">United Kingdom</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Anxiety over jobs being lost as a result of technological change has been rising and falling for centuries, however the most dystopian scenarios of mass unemployment have never fully materialized. The expanding capabilities of powerful new artificial intelligence technologies have prompted some to question whether this time really is different. In light of these concerns, a wide array of smart algorithmic systems are being developed to assist individuals, career advisors, and government departments in navigating the new labor market landscape. Often framed as neutral and apolitical decision assistance tools, some of these systems have nonetheless encountered critiques in the academic and public realms due to their black box nature or ability to entrench systemic inequalities. In this paper, we explore how the concepts of transparency and explainability were operationalized by a team designing an open source career transition algorithm. We find that transparency and explainability are considered to be broadly synonymous with openness, and the translation of these values into action is influenced by a range of factors including team and organizational culture, informal benchmarking against peer projects, as well as goals such as increasing user trust in the algorithm, ensuring widespread access, and hopes that others in the field will build on the project's outputs. The design team viewed both technical (e.g. algorithm design) and non-technical (e.g. stakeholder engagement) activities as important components of ensuring transparency and explainability, and considerations emerged primarily around how these values should be built into project inputs (e.g. the choice of data sources and algorithmic logic) and outputs (e.g. decisions about when to release the algorithm publicly). While the open sourcing of the algorithm is central to advancing the transparency and explainability goals of the project, it also raises questions about longer-term accountability and complicates the ex-ante assessment of the impact it will have at the level of the socio-technical system as it has the potential to be taken up and used by many different actors with differing aims. This paper adds to a nascent but much needed literature on the development of career guidance algorithms, and although caution should be exercised when drawing conclusions from a single case study, it proposes several recommendations for other teams working in this field.</p>
      </abstract>
      <kwd-group>
        <kwd>1 Transparency</kwd>
        <kwd>Explainability</kwd>
        <kwd>Machine Learning</kwd>
        <kwd>Labor Market</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <sec id="sec-1-1">
        <title>Concerns about widespread job losses caused by technological change (a phenomenon referred to as ‘technological unemployment’) have been rising and falling since at least the</title>
        <p>
          Industrial Revolution, however the most
dystopian scenarios have historically proven to
be overblown as new jobs and industries rose
from the ashes [
          <xref ref-type="bibr" rid="ref1">1</xref>
          ]. As increasingly powerful
artificial intelligence (AI) technologies start to
master the previously automation-immune
tasks performed in white collar, well paid
careers such as medicine and law, anxiety has
again started to grow with some questioning
whether this time really is different. This fear
has been compounded by what appears to be an
acceleration of trends toward automation
during the Covid-19 pandemic [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ].
        </p>
        <p>At the same time that concerns are being
raised about technology’s role in displacing
labor, AI has been increasingly adopted as a
tool for coping with the new situation. The
development of career transition algorithms for
individual, private sector and government use
has grown rapidly in recent years as people look
for ways to navigate the changing landscape.
Their proponents have expressed hope that
these tools will make career transition advice
more efficient, effective, and personalized,
ultimately leading to more opportunities for
those who are most at risk of displacement.
However, they have also been subject to
critique in both the academic and public realms.
Both the optimistic and critical camps have
raised important questions about how
transparency and explainability are enacted in
career transition algorithms, however little is
known about how design teams navigate these
decisions in real-world contexts. In this paper,
we explore the operationalization of
transparency and explainability in a career
transition algorithm project. In the vein of an
emerging body of ethnographic and evaluative
work on algorithm design, this work engages
directly with the design team to explore the
technical and non-technical aspects of
transparency and explainability in the project,
as well as the institutional and contextual
influences on design choices.</p>
      </sec>
    </sec>
    <sec id="sec-2">
      <title>2. Background</title>
    </sec>
    <sec id="sec-3">
      <title>2.1. A brief history of automation and technological unemployment</title>
      <sec id="sec-3-1">
        <title>Anxieties around widespread technological</title>
        <p>
          unemployment tend to rise and fall. The
Industrial Revolution was the first major
technological transition that was extensively
written about in real time, and from the 18th
century onward, economists and others have
debated the short- and long-term impacts of
technological advancements on levels of
2 The language of ‘low-skill’ and ‘middle-skill’ is common in
the literature on this topic, and is not intended here as a
normative judgment.
employment, wages, and the quality of work
[
          <xref ref-type="bibr" rid="ref3">3</xref>
          ]–[
          <xref ref-type="bibr" rid="ref5">5</xref>
          ]. For instance, in a set of lectures
delivered in 1928, economist John Maynard
Keynes warned that unemployment due to
technological innovation was outpacing our
ability to find new uses for labor. Over two
decades later, Nobel Prize-winning economist
Wassily Leontief cautioned that workers were
increasingly being replaced by machines, and
that it was not clear that new industries would
be able to employ everybody who wants a job
[
          <xref ref-type="bibr" rid="ref1">1</xref>
          ]. However, concerns about widespread
technological unemployment have largely
turned out to be overblown. To illustrate this
dynamic through one concrete example, fears
that automated teller machines (ATMs) would
lead to job losses proved false because ATMs
allowed banks to operate branch offices at
lower cost, prompting them to open many more
branches. This expansion ultimately offset the
loss in teller jobs such that the number of
fulltime equivalent bank tellers in the United States
grew following the deployment of ATMs in the
1970s [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ].
        </p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>2.2. Contemporary debates on technological unemployment</title>
      <p>
        Although prophecies of mass
unemployment have largely failed to
materialize, technological advancement has
nonetheless had an important impact on the
labor market. Earlier waves of automation have
primarily affected low- and middle-skill
occupations2, with impacts being felt in manual
labor industries such as agriculture and routine
tasks such as bookkeeping [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]. Early algorithms
were developed in an attempt to directly
replicate human decision making, limiting the
range of tasks they could automate to those that
were routine, repetitive and codifiable.
However, advances in techniques such as
natural language processing, predictive
analytics and image recognition have opened
new pathways to automation, leading some to
wonder whether this wave of automation risk
really is different from those of the past [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ].
In 2013, an Oxford University study set off a
wave of panic when they reported that 47% of
people working in the United States are in jobs
that could be performed by computers or
algorithms within the next 10-20 years [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]3. A
great deal of methodological development has
since taken place4, with later work from
organizations such as the OECD finding that in
fact few jobs have either very high or very low
risk of automatability, providing a reprieve
from the concerning picture painted by the
initial findings from the Oxford study [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ].
However, these revised estimates still suggest
that the impacts of automation will be felt
primarily by lower-skilled, lower-income
workers, raising important equity concerns.
      </p>
    </sec>
    <sec id="sec-5">
      <title>2.3. Career transition algorithms: opportunities and controversies</title>
      <p>
        As the debate about the risks of
technological unemployment have raged and
methodologies for estimating automatability
risk have evolved, governments and other
organizations have started turning to
algorithmic solutions to help navigate the
changing labor market, and investment in these
technologies has started to grow [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ]. For
instance, Silicon Valley-based EightfoldAI
have raised nearly $180 million USD to
develop an ‘AI-powered Talent Intelligence
Platform’ that is described in the company’s
promotional material as ‘the most effective way
for organizations to retain top performers,
upskill and reskill the workforce, recruit top
talent efficiently, and reach diversity goals’
[
        <xref ref-type="bibr" rid="ref12">12</xref>
        ].
      </p>
      <p>
        The deployment of algorithmic tools to
navigate the changing labor market has also at
times been met with critical scrutiny in the
domains of public and academic discourse. For
instance, Allhutter and colleagues critically
interrogate an Austrian job seeker profiling tool
which aims to increase the efficiency of
government career counselling processes and
the effectiveness of active labor market
programs [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ]. Through an assessment of the
tool’s technical documentation as well as policy
documents (e.g. labor market policy targets),
they call into question the system’s purported
neutrality and reveal the ways in which it helps
to enact the framing of unemployment under
austerity politics. The authors also found that
despite trying to present an image of
transparency, significant omissions and
underspecifications in the documentation inhibit the
achievement of meaningful explanations [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ].
Controversies have also arisen in the United
Kingdom (UK), where the government's
National Career Services tool was met with
ridicule and hostility following a series of
seemingly illogical career transition
recommendations [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ], and the re-emergence
of a separate ad campaign that appeared to
imply that people in creative industries should
retrain in areas like cybersecurity [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ]. Taken
together, these examples demonstrate that
career transition advice can be politically
charged, and highlights the need to adopt an
equally politically aware stance while
developing any intervention or tool that seeks
to assist people or institutions navigating the
labor market.
      </p>
    </sec>
    <sec id="sec-6">
      <title>2.4. Transparency and explainability in smart systems</title>
      <sec id="sec-6-1">
        <title>In a 2019 review on the global landscape of</title>
        <p>
          AI ethics published in Nature Machine
Learning, transparency was the most
commonly cited value, appearing in 73 out of
84 guidelines reviewed [
          <xref ref-type="bibr" rid="ref16">16</xref>
          ]. Despite this
widespread emphasis, however, the authors of
the review report widespread variability in how
transparency is interpreted, justified, applied
and evaluated in AI ethics guidelines. This
finding is consistent with a growing body of
taxonomies of different transparency types. For
instance, Weller and colleagues [
          <xref ref-type="bibr" rid="ref16">16</xref>
          ] identified
eight distinct types of transparency, noting that
each would require a different sort of
explanation and different measures of efficacy:
• Type 1: For a developer, to understand
how their system is working, aiming to
debug or improve it, to see what is working
well or badly, and get a sense for why.
• Type 2: For a user, to provide a sense
for what the system is doing and why, to
enable prediction of what it might do in
unforeseen circumstances and build a sense
of trust in the technology.
• Type 3: For society broadly to
understand and become comfortable with
the strengths and limitations of the system,
3 Note that the working paper was published in 2013 but the final
version of the study was published in 2017.
4 For instance, switching from occupational to task-level
estimates of automation risk produces drastically different results.
overcoming a reasonable fear of the
unknown.
• Type 4: For a user to understand why
one particular prediction or decision was
reached, to allow a check that the system
worked appropriately and to enable
meaningful challenge.
• Type 5: To provide an expert (e.g. a
regulator) the ability to audit a prediction or
decision trail in detail, particularly if
something goes wrong and requires the
assignment of accountability or legal
liability.
• Type 6: To facilitate monitoring and
testing for safety standards.
• Type 7: To make a user feel
comfortable with a prediction or decision so
that they keep using the system. This type of
transparency primarily benefits the
deployer.
• Type 8: To lead a user into some action
or behavior such as making a purchase. As
with Type 7, this type of transparency
primarily benefits the deployer.
        </p>
        <p>
          Despite the widespread interest in
promoting transparency and explainability as
well as the broadly held view that this is a
worthwhile end in and of itself, recent work has
called into question if and how transparency
can be operationalized in practice, and whether
this is always a desirable goal [
          <xref ref-type="bibr" rid="ref17">17</xref>
          ], [
          <xref ref-type="bibr" rid="ref18">18</xref>
          ]. For
instance, emerging evidence suggests that in
certain circumstances, signifiers of
transparency such as open sourcing may
actually have the unintended effect of
producing less – rather than more – critical
engagement with algorithmic outputs. Kemper
and Kolkman illustrate this point in relation to
the UK government’s 2050 Calculator, which is
an open source energy and emissions model
[
          <xref ref-type="bibr" rid="ref18">18</xref>
          ]. The developers of this tool found that very
few people looked into the documentation, and
felt that by open sourcing the model, people
were less inclined to contest its outcomes.
        </p>
      </sec>
    </sec>
    <sec id="sec-7">
      <title>3. Case study: transparency and explainability in the development of a career transition algorithm</title>
      <p>
        The algorithmic system that is the focus of
this paper aims to identify career transitions that
are viable, desirable, and safe in the context of
growing concerns around automation. The
onset of the Covid-19 pandemic, which has led
to widespread job losses and an acceleration of
automation trends, has added an additional
layer of perceived urgency to the development
and deployment process. The algorithm was
developed by a non-profit foundation with
funding from a second non-profit foundation. It
maps similarities between over 1,600
occupations based on the skills and tasks that
make up each role, and can be used to identify
a set of jobs requiring similar skills and
activities. The algorithm can also identify skills
that a worker might need to develop in order to
move into a new role. ‘Desirable’ transitions are
defined as those that would incur a limited loss
of earnings, while ‘safe’ transitions are those
that would also lead to a lower automation risk.
The risk of automation is assessed according to
estimates of the suitability of tasks for machine
learning [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ]. At the time of writing, a report
detailing the research findings created using the
algorithm had been published, however the
algorithm and datasets had not yet been
publicly released.
      </p>
      <p>
        Despite the many challenges that exist in
studying algorithms (e.g. their black boxed and
contingent nature), we adopt one of the methods
outlined by Kitchin (2017), which includes
interviewing the algorithm designers to
understand how their objectives were framed
and translated into code, as well as what
influences, constraints, and other factors
influenced their approach [
        <xref ref-type="bibr" rid="ref20">20</xref>
        ]. The findings
below are drawn from a series of discussions
held with the algorithm’s design team in
November and December 2020, in which we
sought to collectively explore and critically
evaluate how transparency and explainability
were taken into account in the project. The core
team includes two data scientists who were
responsible for designing and performing
technical validation on the algorithm; a data
visualization expert who was responsible for
developing a user interface; and two program
managers who were responsible for stakeholder
engagement and the overall functioning of the
project.5
      </p>
    </sec>
    <sec id="sec-8">
      <title>3.2. Operationalizing transparency and explainability in the project</title>
      <p>The terms ‘transparency’ and
‘explainability’ were not explicitly defined in
the project, however members of the
development team noted that the value of
‘openness’ was embedded within both the
team’s core values and the organization’s
broader charitable objectives, and that this was
largely seen to be synonymous with
transparency. Team members put forward
multiple reasons for wanting to work openly
and transparently, including fostering trust in
the final product, increasing the likelihood that
the outputs would continue to be developed by
others, and ensuring that users with fewer
resources were still able to work with the tool.</p>
    </sec>
    <sec id="sec-9">
      <title>3.3. Transparency and explainability across project inputs, outputs and socio-technical system levels</title>
      <p>The project team identified salient questions
and actions around transparency and
explainability at multiple levels in the project,
noting that it may not be possible or desirable
to achieve all of them simultaneously.</p>
    </sec>
    <sec id="sec-10">
      <title>3.3.1. Transparency and explainability of inputs</title>
      <p>Members of the project team suggested that
while it was possible to control much of the
work around transparency in the project
undertaken by the team, questions remained
about how to make the work that underpinned
the project (e.g. inputs such as the underlying
logic, code, etc.) more transparent to end users.
On the one hand, the team felt that the use of
open data sources that are widely accepted in
the field, as well as the use of open source tools
5 Two additional team members (a data scientist and qualitative
researcher) were on parental leave and therefore unable to
participate in the discussions.
(e.g. Python packages), increase the
transparency of project inputs. On the other
hand, as noted by one of the data scientists, the
algorithm relies on an approach developed by a
team of academics in the United States, which
contains its own sets of assumptions and
limitations that may not be easily accessible or
comprehensible to an average end user6. These
limitations nonetheless have implications for
how the algorithm can and should be used, and
although these are flagged in the published
report, the team expressed concern that there
may still be assumptions or underlying logics
embedded in the algorithm that do not get
translated into the way it is ultimately deployed.</p>
    </sec>
    <sec id="sec-11">
      <title>3.3.2. Transparency and explainability of outputs</title>
      <p>The project team pointed to multiple
mechanisms for ensuring the transparency of
the outputs, including both technical and
nontechnical approaches. On the non-technical
side, stakeholder engagement over the course of
the project played an important role in
legitimizing the approach taken and ensuring it
was well understood by stakeholders. More
broadly, discussions with other organizations
working in a similar space were crucial in
identifying possible risks and opportunities
around transparency and explainability. For
instance, one project member noted that
engaging with an organization developing a
black box career transition algorithm served as
encouragement to be more transparent.
Similarly, the team reported feeling influenced
by public controversies around other career
transition algorithms, providing them with
further motivation to ensure that pre-release
validation processes were robust. As one team
member noted: “all it takes is one bad
recommendation for people to lose trust”.</p>
      <p>On the technical side, transparency and
explainability were operationalized at the
output level primarily through the decision to
open source the algorithm following a
validation process that included crowdsourcing
feedback on the transition recommendations.
The lead algorithm designer also emphasized
that specific design choices were taken with end
6 Although part of the rationale for using the method developed
in the US study was that it was itself open (both the code and the
data), which appealed to the team from a transparency
perspective.
user transparency and explainability in mind,
highlighting two specific examples. First, the
team used fully interpretable features that allow
a clear assessment of which elements contribute
most highly to similarities between work
activities or work contexts. Second, the team
privileged the use of natural language
processing methods for comparing job skill sets
that allowed for specific skill matches and gaps
to be clearly identified, as well as a
determination of which skills are contributing
the most to the similarity score.</p>
      <p>The timing of transparency-enhancing
activities was also a factor the team took into
account, with one data scientist indicating that
releasing the algorithm prior to the necessary
validation processes could potentially cause
more harm than good. This is why the team
preferred to work in a private GitHub repository
until the algorithm’s release. The team also
cautioned against total transparency, noting that
decisions about what to show the end user and
what to hold back (e.g. in the design of the user
interface) are actually a core component of
making the tool useful, as showing too much
information can also be confusing and lead to
decreased interpretability. In this sense, perhaps
counterintuitively, the team felt that some
information needed to be ‘hidden’ in an attempt
to make the outputs more usable (while noting
that this information could still be accessed
through other means if desired).</p>
    </sec>
    <sec id="sec-12">
      <title>3.3.3. Transparency and explainability within the sociotechnical system</title>
      <p>
        Transparency was also discussed by the
project team in terms of how the algorithm
would interact with the broader socio-technical
system in which it’s embedded, with one
project team member pointing out that
“transparency is in the DNA of the project” by
virtue of it being aimed at making labor market
information more accessible, thereby informing
more transparent decision-making by actors
within the system (e.g. career counsellors).
However, the team agreed that there was some
uncertainty about how transparency would be
operationalized once the algorithm is made
public. For instance, it is unclear to what extent
its use can realistically be monitored or whether
approaches such as terms of service would
provide any real protection against unintended
uses. For instance, the team emphasized that the
tool is meant to augment the work of career
counsellors rather than replace them. They
noted that this is especially important given
evidence that labor market information alone
(without the support from a counsellor) has
been shown to have very limited impact on job
seeker outcomes [
        <xref ref-type="bibr" rid="ref21">21</xref>
        ]. Despite having engaged
with career counsellors over the course of the
project to ensure this point was clear, the
development team acknowledged that there was
ultimately no way to guarantee that end users
would respect one of the project’s core guiding
principles that the algorithm be used as a
complement to existing tools and processes
rather than being used as a replacement. The
discussion around the socio-technical level also
raised questions about counterfactual scenarios,
with team members suggesting that any risks
posed by the algorithm should be weighed
against the status quo situation.
      </p>
    </sec>
    <sec id="sec-13">
      <title>4. Discussion</title>
      <p>
        Exploring how transparency and
explainability were operationalized in a
realworld project to develop a career transition
algorithm provides an insightful window into
the types of questions, considerations and
tradeoffs at play in practice. The finding that the
project team didn’t see an important distinction
between commonly used AI ethics values such
as openness, transparency or explainability is
consistent with literature showing that these
terms are variably interpreted and applied [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ],
[
        <xref ref-type="bibr" rid="ref17">17</xref>
        ]. The lack of explicit definitions may have
allowed for a more flexible and adaptable
approach, where actions could evolve
organically as needed over the course of the
algorithm’s development in response to
external events such as media critiques of other
career transition tools or algorithms, or
conversations with other organizations working
in a similar space. Indeed, it appears as though
informal peer learning and benchmarking
played an important role in shaping the project
team’s perceptions of what efforts were
required to foster trust amongst users and other
stakeholders, and the ways that this could be
achieved through increased transparency. This
finding suggests that there is merit in
recommendations recently put forward that
encourage greater transparency around failures
as well as successes in AI development [
        <xref ref-type="bibr" rid="ref22">22</xref>
        ].
      </p>
      <p>
        Despite providing more flexibility, the
absence of a clear framework for defining
‘transparency’, ‘explainability’ or ‘openness’
also presents a number of challenges, including
the absence of metrics against which success
(or failure) can be assessed and a way of
tracking trade-offs that were made so that these
can be more widely understood and openly
interrogated. As described above, the growing
literature around transparency types [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ], [
        <xref ref-type="bibr" rid="ref23">23</xref>
        ],
could provide more structure and granularity
around what transparency and explainability
mean to different audiences or stakeholders in
practice. These could be combined with tools
such as an algorithmic design history file [
        <xref ref-type="bibr" rid="ref24">24</xref>
        ]
for tracking design decisions, value alignment,
and findings from risk analysis assessments.
The development of a set of evaluative metrics
to assess not only whether existing approaches
to transparency are effective, but also whether
they are succeeding in inviting critical
feedback, should also be prioritized [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ].
      </p>
      <p>
        The finding that transparency and
explainability were most consistently
considered and incorporated into the input and
output levels of the project, with relatively less
focus on socio-technical system considerations,
merits further critical analysis. As described in
detail elsewhere, evaluations of algorithmic
design and logic can only bring us so far in
understanding what their effects might be.
Gaining a deeper understanding requires us to
interrogate how they become embedded in
broader sociotechnical systems [
        <xref ref-type="bibr" rid="ref20">20</xref>
        ], [
        <xref ref-type="bibr" rid="ref25">25</xref>
        ].
However, this is also one of the most
challenging tasks to undertake, particularly
exante, for at least two reasons. The first
challenge is practical in nature because in this
case, the team does not know exactly who will
use the open source algorithm once it’s
released, so it is difficult to assess if or how the
value of transparency can or should be enacted
once it’s taken up by others (and if so, who is
responsible for this work). A second challenge
which is not unique to this project but rather
applies to technology development more
broadly, is captured in the Collingridge
Dilemma, which highlights the fact that the
impact of a given technology is to some extent
unknowable until it has been integrated into a
given socio-technical system, at which point it
is difficult or impossible to change or control it
[
        <xref ref-type="bibr" rid="ref26">26</xref>
        ]. These challenges, as well as ways of
mitigating risks and potential harms to the
greatest extent possible while granting that
some impacts will inevitably be unknowable a
priori, deserve more attention in the
transparency and explainability literature.
      </p>
    </sec>
    <sec id="sec-14">
      <title>5. Conclusion</title>
      <p>As individuals, companies and governments
attempt to navigate changes in the labor market
caused by technological innovation and the
displacement of workers due to the Covid-19
crisis, it is likely that the use of AI-assisted tools
will grow to meet the need for careers
information, as well as the demand for
reskilling and upskilling advice at an
unprecedented scale. It is therefore essential to
explore the ways in which algorithm design
teams frame, operationalize and measure the
success of efforts aimed at increasing the
transparency and explainability of these
systems.</p>
      <p>In this short case study on how transparency
and explainability were operationalized in a
project to develop a career transition algorithm,
we found that these values are considered to be
broadly synonymous with openness, and that
many considerations are at play when framing
the goals, drivers and barriers toward this end.
We also found that efforts to advance the aims
of transparency and explainability were
implemented primarily at the level of the
project inputs and outputs, rather than at the
level of the socio-technical system.</p>
      <p>
        This paper adds to a sparse but growing
body of literature that critically analyses career
guidance algorithms [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ]. Although we should
avoid inferring too much from a single case
study, we nonetheless identify three key
takeaways from this assessment. The first is that
algorithm development teams should agree
upon what ‘transparency’ and ‘explainability’
mean at the outset of the design process. As
described above, a wide range of new
taxonomies at varying levels of granularity and
for different users have been developed in
recent years. These could be deployed
alongside tools such as design history files to
strike a balance between conceptual clarity and
flexibility. The second key insight is that peer
learning mechanisms (broadly interpreted to
include cases arising in the media or academic
literature as well as discussions with other
teams working in a similar space) can prompt
helpful reflections on the types of risks and
issues that should be taken into account during
algorithmic development and deployment.
Mechanisms for sharing lessons learned,
particularly from failures, should be further
developed and encouraged. Finally, the open
sourcing of algorithms such as the one
developed in this case study creates new
opportunities for navigating shocks to the labor
market while also creating new risks that it
might be used in unintended ways. Further
research should focus on elaborating and
developing mechanisms for monitoring and
accountability in such instances.
      </p>
    </sec>
    <sec id="sec-15">
      <title>6. References</title>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>J.</given-names>
            <surname>Mokyr</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Vickers</surname>
          </string-name>
          , and
          <string-name>
            <given-names>N. L.</given-names>
            <surname>Ziebarth</surname>
          </string-name>
          , “
          <article-title>The History of Technological Anxiety and the Future of Economic Growth: Is This Time Different?,”</article-title>
          <string-name>
            <given-names>J.</given-names>
            <surname>Econ</surname>
          </string-name>
          . Perspect., vol.
          <volume>29</volume>
          , no.
          <issue>3</issue>
          , pp.
          <fpage>31</fpage>
          -
          <lpage>50</lpage>
          , Aug.
          <year>2015</year>
          , doi: 10.1257/jep.29.3.31.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2] OECD, “
          <source>OECD Employment Outlook</source>
          <year>2020</year>
          ,
          <string-name>
            <surname>”</surname>
            <given-names>OECD</given-names>
          </string-name>
          ,
          <year>2020</year>
          . http://www.oecd.org/employmentoutlook/2020 (accessed Sep.
          <volume>21</volume>
          ,
          <year>2020</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>D. H.</given-names>
            <surname>Autor</surname>
          </string-name>
          , “
          <article-title>Why Are There Still So Many Jobs? The History and Future of Workplace Automation,”</article-title>
          <string-name>
            <given-names>J.</given-names>
            <surname>Econ</surname>
          </string-name>
          . Perspect., vol.
          <volume>29</volume>
          , no.
          <issue>3</issue>
          , pp.
          <fpage>3</fpage>
          -
          <lpage>30</lpage>
          , Aug.
          <year>2015</year>
          , doi: 10.1257/jep.29.
          <issue>3</issue>
          .3.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>R. C.</given-names>
            <surname>Allen</surname>
          </string-name>
          , “
          <article-title>Engels' pause: Technical change, capital accumulation, and inequality in the british industrial revolution,” Explor</article-title>
          . Econ. Hist., vol.
          <volume>46</volume>
          , no.
          <issue>4</issue>
          , pp.
          <fpage>418</fpage>
          -
          <lpage>435</lpage>
          , Oct.
          <year>2009</year>
          , doi: 10.1016/j.eeh.
          <year>2009</year>
          .
          <volume>04</volume>
          .004.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>J.</given-names>
            <surname>Clifton</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Glasmeier</surname>
          </string-name>
          , and
          <string-name>
            <given-names>M.</given-names>
            <surname>Gray</surname>
          </string-name>
          , “
          <article-title>When machines think for us: the consequences for work and place,”</article-title>
          <string-name>
            <given-names>Camb. J.</given-names>
            <surname>Reg</surname>
          </string-name>
          .
          <source>Econ. Soc.</source>
          , vol.
          <volume>13</volume>
          , no.
          <issue>1</issue>
          , pp.
          <fpage>3</fpage>
          -
          <lpage>23</lpage>
          , May
          <year>2020</year>
          , doi: 10.1093/cjres/rsaa004.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>J. E.</given-names>
            <surname>Bessen</surname>
          </string-name>
          , “How Computer Automation Affects Occupations: Technology, Jobs, and Skills,” SSRN Electron. J.,
          <year>2015</year>
          , doi: 10.2139/ssrn.2690435.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>D. H.</given-names>
            <surname>Autor</surname>
          </string-name>
          , “
          <article-title>Work of the Past, Work of the Future,” AEA Pap</article-title>
          .
          <source>Proc.</source>
          , vol.
          <volume>109</volume>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>32</lpage>
          , May
          <year>2019</year>
          , doi: 10.1257/pandp.20191110.
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>D.</given-names>
            <surname>Acemoglu</surname>
          </string-name>
          and
          <string-name>
            <given-names>P.</given-names>
            <surname>Restrepo</surname>
          </string-name>
          , “
          <article-title>The wrong kind of AI? Artificial intelligence and the future of labour demand,”</article-title>
          <string-name>
            <given-names>Camb. J.</given-names>
            <surname>Reg</surname>
          </string-name>
          .
          <source>Econ. Soc.</source>
          , vol.
          <volume>13</volume>
          , no.
          <issue>1</issue>
          , pp.
          <fpage>25</fpage>
          -
          <lpage>35</lpage>
          , May
          <year>2020</year>
          , doi: 10.1093/cjres/rsz022.
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>C. B.</given-names>
            <surname>Frey</surname>
          </string-name>
          and
          <string-name>
            <given-names>M. A.</given-names>
            <surname>Osborne</surname>
          </string-name>
          , “
          <article-title>The future of employment: How susceptible are jobs to computerisation?,”</article-title>
          <source>Technol. Forecast. Soc. Change</source>
          , vol.
          <volume>114</volume>
          , pp.
          <fpage>254</fpage>
          -
          <lpage>280</lpage>
          , Jan.
          <year>2017</year>
          , doi: 10.1016/j.techfore.
          <year>2016</year>
          .
          <volume>08</volume>
          .019.
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>M.</given-names>
            <surname>Arntz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Gregory</surname>
          </string-name>
          , and U. Zierahn, “
          <article-title>The Risk of Automation for Jobs in OECD Countries: A Comparative Analysis</article-title>
          ,
          <source>” OECD Social, Employment and Migration Working Papers</source>
          <volume>189</volume>
          , May
          <year>2016</year>
          . doi:
          <volume>10</volume>
          .1787/5jlz9h56dvq7-en.
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>B.</given-names>
            <surname>Dellot</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Mason</surname>
          </string-name>
          , and F. WallaceStephens, “
          <article-title>The four futures of work,”</article-title>
          <source>The RSA Action and Research Centre</source>
          , London, Mar.
          <year>2019</year>
          . Accessed: Dec.
          <volume>21</volume>
          ,
          <year>2020</year>
          . [Online]. Available: https://www.thersa.org/globalassets/pdfs/r eports/rsa_four
          <article-title>-futures-of-work</article-title>
          .pdf.
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <surname>Crunchbase</surname>
          </string-name>
          , “Eightfold - Crunchbase Company Profile &amp; Funding,” Crunchbase, Dec.
          <volume>21</volume>
          ,
          <year>2020</year>
          . https://www.crunchbase.com/organization /eightfold (accessed
          <year>Dec</year>
          .
          <volume>21</volume>
          ,
          <year>2020</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>D.</given-names>
            <surname>Allhutter</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Cech</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Fischer</surname>
          </string-name>
          ,
          <string-name>
            <surname>G.</surname>
          </string-name>
          <article-title>Grill, and</article-title>
          <string-name>
            <given-names>A.</given-names>
            <surname>Mager</surname>
          </string-name>
          , “
          <article-title>Algorithmic Profiling of Job Seekers in Austria: How Austerity Politics Are Made Effective</article-title>
          ,” Front.
          <source>Big Data</source>
          , vol.
          <volume>3</volume>
          , p.
          <fpage>5</fpage>
          ,
          <string-name>
            <surname>Feb</surname>
          </string-name>
          .
          <year>2020</year>
          , doi: 10.3389/fdata.
          <year>2020</year>
          .
          <volume>00005</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>M.</given-names>
            <surname>Weaver</surname>
          </string-name>
          , “
          <article-title>Boxer or lock-keeper? Government careers quiz scorned by users,” the Guardian</article-title>
          , Oct.
          <volume>09</volume>
          ,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>D.</given-names>
            <surname>Jordan</surname>
          </string-name>
          , “
          <article-title>Downing Street joins criticism of 'crass' job ad</article-title>
          ,” BBC News, Oct.
          <volume>12</volume>
          ,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>A.</given-names>
            <surname>Jobin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Ienca</surname>
          </string-name>
          , and E. Vayena, “
          <article-title>The global landscape of AI ethics guidelines</article-title>
          ,” Nat. Mach. Intell., vol.
          <volume>1</volume>
          , no.
          <issue>9</issue>
          ,
          <string-name>
            <surname>Art</surname>
          </string-name>
          . no.
          <issue>9</issue>
          ,
          <string-name>
            <surname>Sep</surname>
          </string-name>
          .
          <year>2019</year>
          , doi: 10.1038/s42256-019- 0088-2.
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>A.</given-names>
            <surname>Weller</surname>
          </string-name>
          , “Transparency: Motivations and Challenges,” ArXiv170801870 Cs, Aug.
          <year>2019</year>
          , Accessed: Dec.
          <volume>21</volume>
          ,
          <year>2020</year>
          . [Online]. Available: http://arxiv.org/abs/1708.
          <year>01870</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>J.</given-names>
            <surname>Kemper</surname>
          </string-name>
          and
          <string-name>
            <given-names>D.</given-names>
            <surname>Kolkman</surname>
          </string-name>
          , “
          <article-title>Transparent to whom? No algorithmic accountability without a critical audience</article-title>
          ,
          <source>” Inf. Commun. Soc.</source>
          , vol.
          <volume>22</volume>
          , no.
          <issue>14</issue>
          , pp.
          <fpage>2081</fpage>
          -
          <lpage>2096</lpage>
          , Dec.
          <year>2019</year>
          , doi: 10.1080/1369118X.
          <year>2018</year>
          .
          <volume>1477967</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>E.</given-names>
            <surname>Brynjolfsson</surname>
          </string-name>
          , T. Mitchell, and
          <string-name>
            <given-names>D.</given-names>
            <surname>Rock</surname>
          </string-name>
          , “
          <article-title>What Can Machines Learn and What Does It Mean for Occupations and the Economy?,” AEA Pap</article-title>
          .
          <source>Proc.</source>
          , vol.
          <volume>108</volume>
          , pp.
          <fpage>43</fpage>
          -
          <lpage>47</lpage>
          , May
          <year>2018</year>
          , doi: 10.1257/pandp.20181019.
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>R.</given-names>
            <surname>Kitchin</surname>
          </string-name>
          , “
          <article-title>Thinking critically about and researching algorithms</article-title>
          ,
          <source>” Inf. Commun. Soc.</source>
          , vol.
          <volume>20</volume>
          , no.
          <issue>1</issue>
          , pp.
          <fpage>14</fpage>
          -
          <lpage>29</lpage>
          , Jan.
          <year>2017</year>
          , doi: 10.1080/1369118X.
          <year>2016</year>
          .
          <volume>1154087</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>R.</given-names>
            <surname>Alexander</surname>
          </string-name>
          ,
          <string-name>
            <surname>G.</surname>
          </string-name>
          <article-title>McCabe, and</article-title>
          <string-name>
            <surname>M. De Backer</surname>
          </string-name>
          ,
          <source>Careers and Labour Market Information: An International Review of the Evidence. Education Development Trust</source>
          ,
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>E.</given-names>
            <surname>Moss</surname>
          </string-name>
          and
          <string-name>
            <given-names>J.</given-names>
            <surname>Metcalf</surname>
          </string-name>
          , “
          <article-title>Ethics Owners: A New Model of Organizational Responsibility in Data-Driven Technology Companies,”</article-title>
          <source>Data Soc.</source>
          , p.
          <fpage>74</fpage>
          ,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <given-names>J.</given-names>
            <surname>Cowls</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>King</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Taddeo</surname>
          </string-name>
          , and L. Floridi, “
          <article-title>Designing AI for Social Good: Seven Essential Factors</article-title>
          ,” Social Science Research Network, Rochester,
          <string-name>
            <surname>NY</surname>
          </string-name>
          , SSRN Scholarly Paper ID 3388669, May
          <year>2019</year>
          . doi:
          <volume>10</volume>
          .2139/ssrn.3388669.
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [24]
          <string-name>
            <given-names>I. D.</given-names>
            <surname>Raji</surname>
          </string-name>
          et al.,
          <article-title>“Closing the AI Accountability Gap: Defining an End-toEnd Framework for Internal Algorithmic Auditing</article-title>
          ,” ArXiv200100973 Cs, Jan.
          <year>2020</year>
          , Accessed: Jul.
          <volume>04</volume>
          ,
          <year>2020</year>
          . [Online]. Available: http://arxiv.org/abs/
          <year>2001</year>
          .00973.
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          [25]
          <string-name>
            <given-names>M.</given-names>
            <surname>Sendak</surname>
          </string-name>
          et al.,
          <article-title>“'The Human Body is a Black Box': Supporting Clinical DecisionMaking with Deep Learning,” ArXiv191108089 Cs, Dec</article-title>
          .
          <year>2019</year>
          , Accessed: Jan.
          <volume>04</volume>
          ,
          <year>2021</year>
          . [Online]. Available: http://arxiv.org/abs/
          <year>1911</year>
          .08089.
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          [26]
          <string-name>
            <given-names>B.</given-names>
            <surname>Ribeiro</surname>
          </string-name>
          et al.,
          <article-title>“Introducing the dilemma of societal alignment for inclusive and responsible research</article-title>
          and innovation,” J. Responsible Innov., vol.
          <volume>5</volume>
          , no.
          <issue>3</issue>
          , pp.
          <fpage>316</fpage>
          -
          <lpage>331</lpage>
          , Sep.
          <year>2018</year>
          , doi: 10.1080/23299460.
          <year>2018</year>
          .
          <volume>1495033</volume>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>