<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Slam the Brakes: Perceptions of Moral Decisions in Driving Dilemmas</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Holly Wilson</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Andreas Theodorou</string-name>
          <email>andreas.theodorou@umu.se</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Umea ̊ University</institution>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>University of Bath</institution>
        </aff>
      </contrib-group>
      <abstract>
        <p>Artificially intelligent agents are increasingly used for morally-salient decisions of high societal impact. Yet, the decision-making algorithms of such agents are rarely transparent. Further, our perception of, and response to, morally-salient decisions may depend on agent type; artificial or natural (human). We developed a Virtual Reality (VR) simulation involving an autonomous vehicle to investigate our perceptions of a morally-salient decision; first moderated by agent type, and second, by an implementation of transparency. Participants in our user study took the role of a passenger in an autonomous vehicle (AV) which makes a moral choice: crash into one of two human-looking Non-Playable Characters (NPC). Experimental subjects were exposed to one of three conditions: (1) participants were led to believe that the car was controlled by a human, (2) the artificial nature of AV was made explicitly clear in the pre-study briefing, but its decisionmaking system was kept opaque, and (3) a transparent AV that reported back the characteristics of the NPCs that influenced its decision-making process. In this paper, we discuss our results, including the distress expressed by our participants at exposing them to a system that makes decisions based on socio-demographic attributes, and their implications. Contact Author</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>Widespread use of fully autonomous vehicles (AVs) is
predicted to reduce accidents, congestion and stress [Fleetwood,
2017; Litman, 2017]. Indeed, the Institute of Electric and
Electronic Engineers (IEEE) predict 75% of cars on the road
will be self-driving by 2040. AVs are one of the
technologies in the transportation domain most followed by the public
[Beiker, 2012]. Critical to this current work; the spotlight
on AVs also illuminates the ‘trolley dilemma’; what action
should an AV take when faced with two morally salient
options? E.g. should the car hit the elderly person in the right
lane, or the young child in the left?</p>
      <p>Many argue that such scenarios are unrealistic and
improbable [Brett, 2015; Goodall, 2016]. Yet, despite their
improbability, such questions generate serious discussion amongst
stakeholders. Germany’s Ethics Commission concluded “in
the event of unavoidable accident situations, any distinction
based on personal features (age, gender, physical or
mental constitution) is strictly prohibited” [Commission, 2017],
whereas other countries are undecided on their stance.
Public opinion surveys consistently report concerns about
misuse, legal implications, and privacy issues [Kyriakidis et al.,
2015].</p>
      <p>As one of the few morally-salient Artificial Intelligence
(AI) dilemmas which has grabbed the attention of many
stakeholders; policy makers, media and public, we feel this
paradigm is uniquely valuable for exploring several critical
research questions in human-computer interactions. These
include: 1) how do our perceptions of a decision-making agent
and their decision, differ dependent on whether the agent
is another human or artificially intelligent; 2) how does an
implementation of transparency regarding the agent’s ‘moral
code’ impact perceptions, with the expectation of calibration;
3) how does the methodology used to present such ‘moral
dilemma’ scenarios to the public, impact their preferences
and perceptions. We now outline each in turn with
consideration of the current status of research and how the question
can be framed within the AV scenario for further
investigation.
2</p>
    </sec>
    <sec id="sec-2">
      <title>Background</title>
      <p>There are many circumstances in which decision-making
Intelligent Agents (IAs) are replacing human decision-makers.
Yet we have not sufficiently established how this shift in
agent-type impacts our perceptions of the decision and
decision-maker. The research gap is especially large in the
context of morally salient decision-making. There are
indications that we both inaccurately assimilate our mental model
of humans with IAs, and have separate expectations and
perceptions of IAs which often lack accuracy [Turkle, 2017].
We discuss the current state of research in our perceived
perceptions of IA objectivity and competence, perceived moral
agency and responsibility and moral frameworks. We then
discuss transparency as a mechanism to calibrate inaccurate
mental model of IAs. We consider crowd-sourcing moral
preferences as a method to guide the moral frameworks in
IAs, and how these are modulated by the methodologies used
to do so.
Research suggests people perceive IAs as more objective and
less prone to have biases than human decision makers. For
example, people were found to be more likely to make
decisions inconsistent with objective data when they believed
the decision was recommended by a computer system than
by a person [Skitka et al., 1999]. Similarly, in a legal setting,
people preferred to adhere to a machine advisor’s decision
even when the human advisor’s judgment had higher
accuracy [Krueger, 2016]. In the context of an AV, higher
attributions of objectivity and competence could result in end-users
feeling more content with decisions than they would be had
the decision been made by a human driver.
We make moral judgements and apply moral norms
differently to artificial than human agents. For example, in a
mining dilemma modelled after the trolley dilemma, robots were
blamed more than humans when the utilitarian action was not
taken [Malle et al., 2015]. This action was also found to be
more permissible with the robot than the human; robots were
expected to make utilitarian choices. This could have
implications for the moral frameworks we might program into
machines—which might not necessarily be equivalent to the
frameworks we prescribe to humans. The impact agent type
has on responsibility attribution is similar. After reading an
AV narrative, participants assigned less responsibility to an
AV at fault than to a human driver at fault [Li et al., 2016].
We initially have higher expectations of IAs, yet are less
forgiving when things go wrong, attributing more blame.
2.3</p>
      <sec id="sec-2-1">
        <title>Inaccurate Mental Models of Moral</title>
      </sec>
      <sec id="sec-2-2">
        <title>Framework</title>
        <p>When we form mental models about a newly encountered IA,
we draw on past experiences, clues from an object’s physical
characteristics and may anthropmorphise. Therefore,
previously encountering an IA which is physically similar but
operates on deontological principles can be misleading. Based
on this past experience, we may assume this newly
encountered IA also operates on deontological principles.
Alternatively, due to anthropomorphism of the IA, we may assume
human bias mechanisms. If in fact, the newly encountered
IA is embedded with a utilitarian moral framework, then we
form an inaccurate mental model. We would then have
reduced understanding of and perhaps even a sub-optimal
interaction with the IA.</p>
        <p>Alternatively, an IA may appear too dissimilar to a
person for an observer to attribute it a moral framework at all.
This too leads to difficulties in predicting the IA’s behaviour.
For optimal interaction and to make informed choices about
usage, we require accurate mental models of moral
frameworks. Yet, there are no previous studies which explicitly
investigate the impact of implementing IA moral framework
transparency on human perceptions of that IA.
2.4</p>
      </sec>
      <sec id="sec-2-3">
        <title>Transparency to Calibrate Mental Models</title>
        <p>
          Transparency is consider an essential requirement for the
development of safe-to-use systems [Theodorou et al., 2017].
A careful implementation of transparency can, for example,
enable real-time calibration of trust to the system
[Dzindolet et al., 2003]. Wortham et al. [
          <xref ref-type="bibr" rid="ref2 ref20 ref34">2017</xref>
          ] revealed a robot’s
drives, competences and the actions of a robot to naive users
through the usage of real-time AI visualisation software; this
additional information increased accuracy of observers’
mental models for the robot. Although, notably, transparency did
not result in perfect understanding: some still overestimated
the robot’s abilities. The present research carries on our
exploration of transparency for IAs.
2.5
        </p>
      </sec>
      <sec id="sec-2-4">
        <title>Crowd-Sourcing Moral Preferences</title>
        <p>AVs could, but not necessarily should, be programmed with
behaviours that conform to a predetermined moral framework
such as utilitarian, deontological or with a normative
framework. There has already been valuable work garnering
normative preferences for the AV moral dilemma; participants
given narratives of different dilemmas, showed a general
preference to minimise casualty numbers rather than protecting
passengers at all costs [Bonnefon et al., 2016] . However,
people no longer wished to sacrifice the passenger when only
one life could be saved, an effect which was amplified when
an additional passenger was in the car such as a family
member.</p>
        <p>Awad et al. [2018] used a Massive Online Experiment
named ‘The Moral Machine’ to determine a global moral
preference for the AV—trolley dilemma: users selected
between two options which were represented by a 2D, pictorial,
birds eye view as a response to ‘What should the self-driving
car do?’. This work made an extensive contribution to
establishing global normative preferences as well as finding
crosscultural ethical variation in preference.</p>
        <p>
          An interesting extension upon ‘The Moral Machine’
foundation, is to explore how decision-making may differ when a
dilemma is presented in a more immersive medium. When
viewing pictures or reading narratives, as in the study of
Awad [
          <xref ref-type="bibr" rid="ref2 ref20 ref34">2017</xref>
          ], there is less emotional elicitation than in the
equivalent real life situations, whereas VR has higher
ecological validity, provoking true to life emotions and behaviours
[Rovira et al., 2009]. Importantly, people have been found
to make different decisions for moral dilemmas in immersive
VR simulation than in desktop VR scenarios [Pan and Slater,
2011]. Specifically, immersive VR induced more panic, and
less utilitarian decision making.
3
        </p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Technology Used</title>
      <p>The observation that moral intuitions may vary with
presentation motivates us to use a VR environment for our present
work. Additionally, participants are passengers inside the car,
rather than bystanders removed from the dilemma. Unlike
most past research, transparency in this study will be
implemented post-decision rather than real-time. We developed a
VR environment in Unity, optimised for Oculus Rift. Unity
software was chosen due to the wide range of free available
assets. The AV simulation, a screenshot is seen in figure 1,
is designed so that participants are seated in the driver’s seat
of a car. The car has detailed interior to increase realism and
thus immersion. The car, positioned on the left hand lane
as the experiments took place in the UK, drives through a city
environment and approaches a zebra-crossing. There are
nonplayable characters, of the various “physical types” described
in Section 3, crossing the road. There are eleven scenarios in
total, of which one is a practice and thus devoid of characters.</p>
      <p>We developed a VR AV simulation to explore public
perceptions of moral decisions1. This roughly simulates the
Moral Machine scenarios [Awad, 2017], in which an AV hits
one of two individuals or groups of pedestrians. This
experimental tool facilitated two experiments presented here, which
seek to answer questions posed above.</p>
      <p>In this section, we first present our decision-making
framework. A brief outline of our VR simulation is provided,
alongside justification for design choices, selection of
pedestrian attributes and how transparency is implemented. We
then move to outline the experimental design of the two
experiments.</p>
      <p>Design of Moral Dilemma and AI Decision Making We
opted to use only a selection of the dimensions used in other
studies on moral preferences. This is because we are not
measuring which characteristics the participants would prefer to
be saved, but rather the response to the use of characteristic
based decision-making in the first place. We picked the three
more visible characteristics: occupation, sex, and body size;
due to limited availability of assets and the pictorial rather
than textual presentation of the scenario to the participant.</p>
      <p>Occupation includes four representative conditions: a
medic to represent someone who is often associated with
contribution to the wealth of the community, military to represent
a risk-taking profession [McCartney, 2011], businessman or
businesswoman as it is associated with wealth, and finally
unemployed. The body size can be either non-athletic or athletic
slim. To further reduce the dimensions of the problem, we
used a binary gender choice (female and male). Although we
varied race between scenarios (Caucasian, Black, and Asian),
the character pairs within scenarios were always of the same
race. Examples of characters used are depicted in Fig. 2.
Note, we do not claim that this is the ‘right’ hierarchy of
social values —or that a choice should take place based on
socio-demographic characteristics in the first place. Rather,
1Code for this simulation will be made available on publication.
We ran a study with three independent groups; human
driver (Group 1), opaque AV (Group 2), and transparent AV
(Group 3). We randomly allocated participants to the
independent variable conditions. For each condition, both the
experimental procedure and the VR Moral Dilemma Simulator
were adjusted in the pre-treatment briefing. In this section,
we describe our procedure for each condition.</p>
      <sec id="sec-3-1">
        <title>4.1 Participants recruitment and pre-briefing</title>
        <p>To reduce an age bias often observed in studies performed
with undergraduate and postgraduate students, we decided to
recruit through a non-conventional means. Participants
recruitment took place at a local prominent art gallery, where
we exhibited our VR simulation as part of a series of
interactive installations. Ethics approval was obtained from the
Department of Computer Science at University of Bath.
Members of the public visiting the gallery were approached and
invited to take part to the experiment. They were told the
purpose of the experiment is to investigate technology and moral
dilemmas in a driving paradigm. After completing a
preliminary questionnaire to gather demographic, driving-preference
and social-identity data, participants entered the VR
environment. After either completion of the VR paradigm or when
the participant decided to stop, the participant was requested
to fill out a post simulation questionnaire. This questionnaire
aims to capture the participant’s perceptions of the agent
controlling the car. It includes dimensions of likeability,
intelligence, trust, prejudice, objectivity and morality. Whilst some
questions are hand-crafted for the purposes of this study, most
are derived from the GodSpeed Questionnaire Series as they
are demonstrated to have high internal consistency [Bartneck
et al., 2009]. The majority of items are measured on a 5–point
Likert Scale.</p>
      </sec>
      <sec id="sec-3-2">
        <title>4.2 Human Driver Condition</title>
        <p>Participants were informed that they were to be a passenger
sat in the driver seat, in either an AV or a car controlled by
a human driver. In reality, the same intelligent system
controlled the car in both conditions. To make the human driver
condition believable, before putting on the headset, the
participants were shown a ‘fake’ control screen. A physical games
controller, which the experimenter pretended to ‘use’ to
control the car, was placed at the table. At the end of the
experiment, participants were debriefed and told that there was no
actual human controlling the car and it was automated as in
the AV condition.
4.3</p>
      </sec>
      <sec id="sec-3-3">
        <title>Opaque and Transparent AV Conditions</title>
        <p>Transparency here, refers to revealing to the end user the
moral framework the agent uses to make its decision. The
moral framework for this paradigm is social value. The
difference between the transparent and non-transparent
condition is in the Question Scene. Post-scenario, after the AV
has hit one of two pedestrians, a statement is made that “The
self-driving car made the decision on the basis that. . . ” then
the reasoning logic is inserted next. For example, if the pair
consisted of one medic and another military, the justification
will state “Medics are valued more highly than military,
business or undefined professions”. Whereas, if the pair differ
only in gender, it will state “Both sides have the same
profession and body size, however females are valued more highly
than males”. In this experiment, the transparency only relays
aspects of the agent’s moral framework. There is no
transparency over mechanics, such as whether the brakes were
working, the speed of the car, or turning direction.</p>
        <p>Several modalities of transparency implementation were
considered such as diagrams, design metaphors and audio,
although written depictions were ultimately used.
Postdecision transparency was chosen to be appropriate, as this
paradigm invokes a fast paced situation where real-time
implementation is infeasible due to technical and human
processing limitations.
5</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Results</title>
      <p>Imbalance of baseline variables is usually considered
undesirable, as the essence of a controlled trial is to compare groups
that differ only with respect to their treatment. Others suggest
that randomised—unbalanced—trials provide more
meaningful results as they compact chance bias [Roberts and
Torgerson, 1999]. A Chi-squared test of goodness-of-fit was
performed to determine frequencies of gender, ethnicity, age,
driving preferences and programming experience between the
three conditions (see 1). Groups were found to be unbalanced
for gender and ethnicity. The ethnicity difference between
the groups is due to a number of people who did not answer
the ‘Ethnicity’ question; the vast majority of all groups
consisted of participants who identified themselves as white and
no other ethnicities were reported. The unbalance for
gender, however, should be taken into consideration during the
analysis of the results.</p>
      <p>Group 2:
Opaque AV</p>
      <p>Group 3:
Transparent AV</p>
      <p>X(2)</p>
      <p>P
Variable
Gender Male
Gender Female
Gender Unknown
White
Asian
Black/Caribbean
None/Unknown</p>
      <p>For comparisons of human-driver versus the opaque AV, a
one-way ANOVA was conducted on all ordinal Likert scale
variables. All but two associations were found to be
nonsignificant (n.s.), see Table 2. The AV was perceived to be
significantly less human-like (p = 0.001), and less morally
culpable (p = 0.04) than the human driver. Although the impact
of agent-type was n.s., medium effect sizes were found for the
human driver being perceived as more pleasant ( p2 = 0.105, d
= 0.69) and nice ( p2 = 0.124), d = 0.75) than the AV [Becker,
2000]. In a second one-way ANOVA, comparing the opaque
AV and transparent AV conditions, three significant effects
were found. The AV was perceived to be significantly more
unconscious rather than conscious (p &lt; 0:001), machine-like
than humanlike (p = 0:04) and intentional rather than
unintentional (p = 0:038) (see Table 5) in the transparent
condition than the non-transparent condition. All other differences
were n.s. In a third one-way ANOVA, comparing the human
driver and transparent AV condition, four significant effects
were found see Table 3. The human driver was found to be
significantly more pleasant (p = 0:01), nice (p = 0:018),
humanlike (p &lt; 0:001) and conscious (p &lt; 0:001) than the
transparent AV. All other differences were n.s.</p>
      <p>A chi-square test of independence was performed to
examine the relation between transparency and understanding of
the decision made 2(1) = 7:34p = 0:007. Participants in
the transparent condition were more likely to report
understanding (87.5%) than (43.75%) (see 4).</p>
      <p>The majority of participants across conditions expressed a
preference for decisions made in moral dilemmas to be made
at random rather than on the basis of social-value.
Preferences are as follows; 71.7% random, 17.9% social value,
7.7% unspecified criteria and 2.6% preferred neither (Fig. 4).</p>
    </sec>
    <sec id="sec-5">
      <title>6 Discussion</title>
      <p>We investigated how certain factors, namely agent type and
transparency level, impact perceptions of a decision maker
and the decision made in moral dilemmas. We discuss the
findings for these conditions and consider the qualitative and
quantitative findings that emerged from both. We place
initial emphasis on participants’ reactions to the decision being
made on social value and the modulating impact of
methodology.
2.68 (32)
0.01
Our experiment elicited strong emotional reactions in
participants, who vocalised being against selection based on
social value. This response was far more pronounced in the
autonomous vehicle condition than with the human driver. Our
quantitative and qualitative data raise interesting questions
about the validity of data captured by Trolley Problem
experiments, such as the the Moral Machine [Shariff et al., 2017;
Awad et al., 2018] as a means to ‘crowdsource’ the moral
framework of our cars by using socio-economic and
demographic data. While such data are definitely worth analysing
as a means to understand cultural differences between
populations, they may not necessarily be representative of
people’s preferences in an actual accident. A lack of an option
to make an explicit ‘random’ choice combined with the use
of a non-immersive methodologies, could lead participants in
‘text description’ conditions to feel forced to make a logical
choice.</p>
      <p>The disparity in findings reflects differing processes of
decision making between the rational decision making in the
Moral Machine and emotional decision-making in the
current experiment. Due to their increased realism, as
previously discussed, VR environments are known to be more
effective at eliciting emotion than narratives or 2D pictures.
Although the graphics used in this experiment were only
semirealistic, the screams were real recordings. Participants
commented on the emotional impact and stress the screams had
on them. Additionally, they were visibly upset after
completing the experiment and expressed discomfort at having
to respond about social value decisions of which they
disagreed with on principle. Other participants removed their
consent, requested data to be destroyed, or even provided
us with strongly-worded verbal feedback. Likely, the
emotion elicitation was enhanced further, as the participant was a
passenger inside the car as opposed to a bystander removed
from the situation as in past experiments. It is unlikely that
the Moral Machine and other online-survey narrative-based
moral experiments elicit such emotional responses. This is
also supported by Pedersen et al. [2018], where participants
in autonomous-vehicle simulation study significantly altered
their perception of the actions taken by an AV when a crash
could lead to real-life consequences.</p>
      <p>Our qualitative results also indicate that subjects may feel
uncomfortable being associated with an autonomous vehicle
that uses protected demographic and socio-economic
characteristics for its decision-making process. This might be due
to a belief that the users of such a product will be considered
as discriminators by agreeing with a system that uses
gender, occupation, age, or race to make a choice. This belief
could potentially also lead to a fear that the user may share
any responsibility behind the accident or be judged by others
—including by the experimenter.
6.2</p>
      <sec id="sec-5-1">
        <title>Perceptions of Moral Agency</title>
        <p>Based on past research, we predicted that the autonomous car
condition would be perceived as more objective and
intelligent but less prejudiced, conscious and human-like, and be
attributed less culpability and moral agency than the ‘human
driver’. We found that human drivers were perceived as
significantly more humanlike and conscious than autonomous
cars. This finding is consistent with expectations and
validates that participants perceived the two groups differently,
especially, as we primed our subjects in the pre-briefing by
telling them that the driver is a ‘human’.</p>
        <p>
          Human drivers (Group 1) were perceived to be significantly
more morally culpable than autonomous driver in the opaque
AV condition (Group 2). However, strikingly, the reverse was
observed when the car’s decision-making system was made
transparent. Furthermore, in the transparency condition,
participants assigned significantly more blame to the car than the
‘human’ driver. Although, as Group 1 believed the
experimenter was controlling the car, less blame may be due to
identification with the experimenter or other person specific
confounds. Our implementation of transparency made the
machine nature of the AV explicitly clear to its passengers, with
participants in Group 3 (transparency condition) describing
the AV as significantly more machinelike compared to
participants in Groups 1 and 2. Our findings contradict recent work
by Malle et al. [
          <xref ref-type="bibr" rid="ref13 ref16">2016</xref>
          ], which demonstrate people perceive
mechanistic robots as having less agency and moral agency
than humans. Moreover, our results conflict with the results
presented in Li et al. [
          <xref ref-type="bibr" rid="ref13 ref16">2016</xref>
          ], where participants assigned less
responsibility to an autonomous vehicle car at fault than to a
human driver at fault.
        </p>
        <p>In the transparency condition we made the passengers
aware that the car used demographic and social-value
characteristics to make a non-random decision. This explains why
participants in Group 3 also significantly described the AV as
more intentional rather than unintentional compared to
subjects in the other two conditions. Although we inevitably
unconsciously anthropomorphise machines, something that
our post-incident transparency minimised by significant
reducing its perception as humanlike and as conscious, we still
associate emotions more easily with humans than machines
[Haslam et al., 2008]. Reduced emotion in decision-making
is linked to making more utilitarian judgements, as supported
by behavioural and neuropsychological research [Moll and
de Oliveira-Souza, 2007; Lee and Gino, 2015]. Therefore,
we believe that participants in the transparency condition
may have also perceived decisions as utilitarian, as the car
was maximising the social value —at least based on same
perception— it would save.</p>
        <p>We believe that the increased attribution of moral
responsibility is due to realisation that the action was determined
based on social values, something that subjects (across all
groups), as we already discussed, disagreed with. This is
supported by past research findings: we perceive other humans as
less humanlike when they lack empathy and carry out actions
which we deem to be morally wrong. For example,
offenders are dehumanised based on their crimes, which we view as
‘subhuman’ and ‘beastly’ [Bastian et al., 2013]. Actions that
go against our moral codes can elicit visceral responses which
is consistent with the emotional reactions of the participants
of the current study.</p>
        <p>
          Our findings may also reflect forgiveness towards the
‘human’ driver or even the opaque AV, but not the transparent
AV. This is supported by previous studies from the literature,
which demonstrate how we tend to forgive human-made
errors easier than machine-made errors [Madhavan and
Wiegmann, 2007; Salem et al., 2015]. This effect is increased
when the robot is perceived as having more autonomy [Kim
and Hinds, 2006]. In addition, Malle et al. [
          <xref ref-type="bibr" rid="ref8">2015</xref>
          ]
demonstrate, with the use of a moral dilemma modelled after the
trolley problem, that robots are blamed more than humans
when a utilitarian action is not taken. Furthermore, their
results also suggest that a utilitarian action is also be more
permissible —if not expected— when taken by a robot. If for
example the robot was performing random choices, then the
moral blame might have been higher.
        </p>
        <p>
          The gender imbalance between the groups might also be
a factor, but potentially not a conclusive one. The Moral
Machine dataset shows minor differences in the preferences
between male-identified and female-identified participants
[Awad et al., 2018], e.g. male respondents are 0.06% less
inclined to spare females, whereas one increase in standard
deviation of religiosity of the respondent is associated with
0.09% more inclination to spare humans. Further analysis by
Awad [
          <xref ref-type="bibr" rid="ref2 ref20 ref34">2017</xref>
          ] indicates that female participants were acting
slightly more utilitarian than males —but both genders are
acting as such. Group 3 was the only group where the vast
majority of its members identified themselves as males and
some of its members may have disagreed with the actions
taken by the agent. Whilst a plausible explanation, it does
not discount the previous discussions —especially,
considering that males in the Moral Machine still had a preference
towards utilitarian actions. Still, we recognise the need to
recapture the data for Group 3.
6.3
        </p>
      </sec>
      <sec id="sec-5-2">
        <title>Mental Model Accuracy</title>
        <p>Although this was not the focus of the study, we asked
participants from Groups 2 and 3 (opaque and transparent AV
respectively) to self-evaluate their understanding of how a
decision was made. Significantly more participants in the
transparency condition reported an understanding of the
decisionmaking process. In addition, passengers in the transparent AV
also rated the AV as significantly more predictable than the
‘human’ driver and higher (non-significant result; Mean for
Group 2 is 3.31 and mean for Group 3 is 4) than the opaque
AV.</p>
        <p>Having accurate mental models by having an
understanding of the decision-making mechanism is crucial for the safe
use of the system. In this experiment we used a post-incident
implementation of transparency instead of a real-time one.
Hence the user could only calibrate its mental model
regarding the decision and the agent after the incident. However,
as the user repeated the simulation ten times, she could still
use previously gathered information, e.g. that the car makes
a non-random decision or even of the priorities of the AV’s
action-selection system, and predict if the car would change
lanes or not.
7</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>Conclusions and Future Work</title>
      <p>Exciting new technology is stirring up debates which speak
to ethical conundrums and to our understanding of human
compared to machine minds. By focusing our efforts on
understanding the dynamics of human-machine interaction, we
can aspire to have appropriate legislation, regulations and
designs in place before such technology hits the streets. In this
project we created a moral-dilemma virtual-reality paradigm
to explore questions raised by previous research. We have
demonstrated morally salient differences in judgement based
on very straightforward alterations of presentation.
Presenting a dilemma in VR from a passenger’s view gives an altered
response versus previously reported accounts from a bird’s
eye view. In this VR context, presenting the same AI as a
human gives a completely different set of judgements of
decisions versus having it presented as an autonomous vehicle,
despite the subjects’ knowing in both cases that their
environment was entirely synthetic.</p>
      <p>There are important takeaway messages to this research.
Crowd-sourced preferences in moral-dilemmas are impacted
by the methodology used to present the dilemma as well as
the questions asked. This indicates a need for caution when
incorporating supposed normative data into moral
frameworks used in technology. Furthermore, our results indicate
that the show of transparency makes the agent appear to be
significantly less anthropomorphic. In addition, our results
agree with the literature that transparency can help naive users
to calibrate their mental models. However, our results also
show that transparency alone is not sufficient to ensure that
we attribute blame—and, therefore, responsibility—only to
legal persons, i.e. companies and humans. Therefore, it is
essential to ensure that we address by ownership and/or usage
our responsibility and accountability [Bryson and Theodorou,
2019].</p>
      <p>Here, it is important to also recognise a limitation of our
own study; the lack of a ‘self-sacrifice’ scenario, where the
car sacrifices its passenger to save the pedestrians. The
implementation of this ‘self-sacrifice’ feature could potentially
lead to different results. A missed opportunity is that we did
not collect users’ preferences at each dilemma to enable
further comparisons. Finally, a future rerun to both gather
additional data and eliminate any concerns for results due to
gender imbalance between the groups is necessary.</p>
    </sec>
    <sec id="sec-7">
      <title>Acknowledgments</title>
      <p>We would like to acknowledge Joanna Bryson for her
guidance with the experimental design and feedback on this
paper, Leon Watts for lending us the necessary computer
equipment to conduct our study, Alan Hayes for his
feedback, and The Edge Arts Centre for hosting us during
data collection. Thanks also to the helpful reviewers. We
also acknowledge EPSRC grant [EP/S515279/1] for
funding Wilson. Theodorou was funded by the EPSRC grant
[EP/L016540/1]. Final publication and dissemination of this
paper would not have been possible with out the European
Union’s Horizon 2020 research and innovation programme
under grant agreement No 825619 funding Theodorou.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [Awad et al.,
          <year>2018</year>
          ]
          <string-name>
            <given-names>Edmond</given-names>
            <surname>Awad</surname>
          </string-name>
          , Sohan Dsouza, Richard Kim, Jonathan Schulz, Joseph Henrich, Azim Shariff,
          <article-title>Jean-Franc¸ois Bonnefon, and Iyad Rahwan. The moral machine experiment</article-title>
          .
          <source>Nature</source>
          ,
          <volume>563</volume>
          (
          <issue>7729</issue>
          ):
          <fpage>59</fpage>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          <source>[Awad</source>
          , 2017]
          <string-name>
            <given-names>Edmond</given-names>
            <surname>Awad</surname>
          </string-name>
          .
          <article-title>Moral machines: perception of moral judgment made by machines</article-title>
          .
          <source>PhD thesis</source>
          , Massachusetts Institute of Technology,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [Bartneck et al.,
          <year>2009</year>
          ]
          <string-name>
            <given-names>Christoph</given-names>
            <surname>Bartneck</surname>
          </string-name>
          , Dana Kulic´,
          <string-name>
            <given-names>Elizabeth</given-names>
            <surname>Croft</surname>
          </string-name>
          , and
          <string-name>
            <given-names>Susana</given-names>
            <surname>Zoghbi</surname>
          </string-name>
          .
          <article-title>Measurement instruments for the anthropomorphism, animacy, likeability, perceived intelligence, and perceived safety of robots</article-title>
          .
          <source>International Journal of Social Robotics</source>
          ,
          <volume>1</volume>
          (
          <issue>1</issue>
          ):
          <fpage>71</fpage>
          -
          <lpage>81</lpage>
          ,
          <year>2009</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [Bastian et al.,
          <year>2013</year>
          ]
          <string-name>
            <given-names>Brock</given-names>
            <surname>Bastian</surname>
          </string-name>
          , Thomas F Denson,
          <string-name>
            <given-names>and Nick</given-names>
            <surname>Haslam</surname>
          </string-name>
          .
          <article-title>The roles of dehumanization and moral outrage in retributive justice</article-title>
          .
          <source>PloS ONE</source>
          ,
          <volume>8</volume>
          (
          <issue>4</issue>
          ):e61842,
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          <source>[Becker</source>
          , 2000]
          <article-title>Lee A Becker. Effect size (es)</article-title>
          .
          <source>Retrieved September</source>
          ,
          <volume>9</volume>
          :
          <year>2007</year>
          ,
          <year>2000</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          <source>[Beiker</source>
          , 2012]
          <article-title>Sven A Beiker. Legal aspects of autonomous driving</article-title>
          . Santa Clara L. Rev.,
          <volume>52</volume>
          :
          <fpage>1145</fpage>
          ,
          <year>2012</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [Bonnefon et al.,
          <year>2016</year>
          ]
          <article-title>Jean-Franc¸ois Bonnefon, Azim Shariff, and Iyad Rahwan. The social dilemma of autonomous vehicles</article-title>
          .
          <source>Science</source>
          ,
          <volume>352</volume>
          (
          <issue>6293</issue>
          ):
          <fpage>1573</fpage>
          -
          <lpage>1576</lpage>
          ,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          <source>[Brett</source>
          , 2015]
          <string-name>
            <given-names>Rose</given-names>
            <surname>Brett</surname>
          </string-name>
          .
          <article-title>The myth of autonomous vehicles' new craze: Ethical algorithms</article-title>
          ,
          <year>November 2015</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          <source>[Bryson and Theodorou</source>
          , 2019]
          <string-name>
            <given-names>Joanna</given-names>
            <surname>Bryson</surname>
          </string-name>
          and
          <string-name>
            <given-names>Andreas</given-names>
            <surname>Theodorou</surname>
          </string-name>
          .
          <source>How Society Can Maintain Human-Centric Artificial Intelligence</source>
          . In Marja Toivonen-Noro, Evelina Saari, Helina¨ Melkas, and Mervin Hasu, editors,
          <source>Humancentered digitalization and services</source>
          .
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          <source>[Commission</source>
          , 2017]
          <string-name>
            <given-names>Ethics</given-names>
            <surname>Commission</surname>
          </string-name>
          .
          <source>connected driving</source>
          ,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          <source>Automated and [Dzindolet</source>
          et al.,
          <year>2003</year>
          ]
          <string-name>
            <surname>Mary</surname>
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Dzindolet</surname>
            ,
            <given-names>Scott A.</given-names>
          </string-name>
          <string-name>
            <surname>Peterson</surname>
          </string-name>
          , Regina A.
          <string-name>
            <surname>Pomranky</surname>
            , Linda G. Pierce, and
            <given-names>Hall P.</given-names>
          </string-name>
          <string-name>
            <surname>Beck</surname>
          </string-name>
          .
          <article-title>The role of trust in automation reliance</article-title>
          .
          <source>International Journal of Human Computer Studies</source>
          ,
          <volume>58</volume>
          (
          <issue>6</issue>
          ):
          <fpage>697</fpage>
          -
          <lpage>718</lpage>
          ,
          <year>2003</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          <source>[Fleetwood</source>
          , 2017]
          <string-name>
            <given-names>Janet</given-names>
            <surname>Fleetwood</surname>
          </string-name>
          .
          <article-title>Public health, ethics</article-title>
          , and autonomous vehicles.
          <source>American Journal of Public Health</source>
          ,
          <volume>107</volume>
          (
          <issue>4</issue>
          ):
          <fpage>532</fpage>
          -
          <lpage>537</lpage>
          ,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          <source>[Goodall</source>
          , 2016]
          <article-title>Noah J Goodall. Away from trolley problems and toward risk management</article-title>
          .
          <source>Applied Artificial Intelligence</source>
          ,
          <volume>30</volume>
          (
          <issue>8</issue>
          ):
          <fpage>810</fpage>
          -
          <lpage>821</lpage>
          ,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [Haslam et al.,
          <year>2008</year>
          ]
          <string-name>
            <given-names>Nick</given-names>
            <surname>Haslam</surname>
          </string-name>
          , Yoshihisa Kashima, Stephen Loughnan,
          <string-name>
            <given-names>Junqi</given-names>
            <surname>Shi</surname>
          </string-name>
          , and
          <string-name>
            <given-names>Caterina</given-names>
            <surname>Suitner</surname>
          </string-name>
          .
          <article-title>Subhuman, inhuman, and superhuman: Contrasting humans with nonhumans in three cultures</article-title>
          .
          <source>Social Cognition</source>
          ,
          <volume>26</volume>
          (
          <issue>2</issue>
          ):
          <fpage>248</fpage>
          -
          <lpage>258</lpage>
          ,
          <year>2008</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          <source>[Kim and Hinds</source>
          , 2006]
          <string-name>
            <given-names>Taemie</given-names>
            <surname>Kim</surname>
          </string-name>
          and
          <string-name>
            <given-names>Pamela</given-names>
            <surname>Hinds</surname>
          </string-name>
          .
          <article-title>Who should I blame? Effects of autonomy and transparency on attributions in human-robot interaction</article-title>
          .
          <source>Proceedings - IEEE International Workshop on Robot and Human Interactive Communication</source>
          , pages
          <fpage>80</fpage>
          -
          <lpage>85</lpage>
          ,
          <year>2006</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          <source>[Krueger</source>
          , 2016]
          <string-name>
            <given-names>Frank</given-names>
            <surname>Krueger</surname>
          </string-name>
          .
          <article-title>Neural signatures of trust during human-automation interactions</article-title>
          .
          <source>Technical report</source>
          , George Mason University Fairfax United States,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [Kyriakidis et al.,
          <year>2015</year>
          ]
          <string-name>
            <given-names>Miltos</given-names>
            <surname>Kyriakidis</surname>
          </string-name>
          , Riender Happee, and Joost CF de Winter.
          <source>Public opinion on automated driving: Results of an international questionnaire among 5000 respondents. Transportation Research Part F: Traffic Psychology and Behaviour</source>
          ,
          <volume>32</volume>
          :
          <fpage>127</fpage>
          -
          <lpage>140</lpage>
          ,
          <year>2015</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          <source>[Lee and Gino</source>
          , 2015]
          <article-title>Jooa Julia Lee</article-title>
          and
          <string-name>
            <given-names>Francesca</given-names>
            <surname>Gino</surname>
          </string-name>
          .
          <article-title>Poker-faced morality: Concealing emotions leads to utilitarian decision making</article-title>
          .
          <source>Organizational Behavior and Human Decision Processes</source>
          ,
          <volume>126</volume>
          :
          <fpage>49</fpage>
          -
          <lpage>64</lpage>
          ,
          <year>2015</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          <string-name>
            <surname>[Li</surname>
          </string-name>
          et al.,
          <year>2016</year>
          ]
          <string-name>
            <given-names>Jamy</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Xuan</given-names>
            <surname>Zhao</surname>
          </string-name>
          ,
          <string-name>
            <surname>Mu-Jung</surname>
            <given-names>Cho</given-names>
          </string-name>
          , Wendy Ju, and Bertram F Malle.
          <article-title>From trolley to autonomous vehicle: Perceptions of responsibility and moral norms in traffic accidents with self-driving cars</article-title>
          .
          <source>Technical report, SAE Technical Paper</source>
          ,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          <source>[Litman</source>
          , 2017]
          <string-name>
            <given-names>Todd</given-names>
            <surname>Litman</surname>
          </string-name>
          .
          <article-title>Autonomous vehicle implementation predictions</article-title>
          .
          <source>Victoria Transport Policy Institute</source>
          ,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          <source>[Madhavan and Wiegmann</source>
          , 2007]
          <article-title>Poornima Madhavan and Douglas A Wiegmann. Similarities and differences between human-human and human-automation trust: an integrative review</article-title>
          .
          <source>Theoretical Issues in Ergonomics Science</source>
          ,
          <volume>8</volume>
          (
          <issue>4</issue>
          ):
          <fpage>277</fpage>
          -
          <lpage>301</lpage>
          ,
          <year>2007</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [Malle et al.,
          <year>2015</year>
          ] Bertram F Malle, Matthias Scheutz, Thomas Arnold, John Voiklis, and
          <string-name>
            <given-names>Corey</given-names>
            <surname>Cusimano</surname>
          </string-name>
          .
          <article-title>Sacrifice one for the good of many?: People apply different moral norms to human and robot agents</article-title>
          .
          <source>In Proceedings of the tenth annual ACM/IEEE International Conference on Human-Robot Interaction</source>
          , pages
          <fpage>117</fpage>
          -
          <lpage>124</lpage>
          . ACM,
          <year>2015</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [Malle et al.,
          <year>2016</year>
          ] Bertram F Malle,
          <string-name>
            <surname>Matthias Scheutz</surname>
            , Jodi Forlizzi,
            <given-names>and John</given-names>
          </string-name>
          <string-name>
            <surname>Voiklis</surname>
          </string-name>
          .
          <article-title>Which robot am i thinking about?: The impact of action and appearance on people's evaluations of a moral robot</article-title>
          .
          <source>In The Eleventh ACM/IEEE International Conference on Human Robot Interaction</source>
          , pages
          <fpage>125</fpage>
          -
          <lpage>132</lpage>
          . IEEE Press,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          <source>[McCartney</source>
          ,
          <year>2011</year>
          ]
          <string-name>
            <given-names>Helen</given-names>
            <surname>McCartney. Hero</surname>
          </string-name>
          , Victimor Villain?
          <article-title>The Public Image of the British Soldier and its Implications for Defense Policy</article-title>
          .
          <source>Defense &amp; Security Analysis</source>
          ,
          <volume>27</volume>
          (
          <issue>1</issue>
          ):
          <fpage>43</fpage>
          -
          <lpage>54</lpage>
          , mar
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          [Moll and de Oliveira-Souza,
          <year>2007</year>
          ]
          <string-name>
            <given-names>Jorge</given-names>
            <surname>Moll and Ricardo de</surname>
          </string-name>
          Oliveira-Souza.
          <article-title>Moral judgments, emotions and the utilitarian brain</article-title>
          .
          <source>Trends in Cognitive Sciences</source>
          ,
          <volume>11</volume>
          (
          <issue>8</issue>
          ):
          <fpage>319</fpage>
          -
          <lpage>321</lpage>
          ,
          <year>2007</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          <source>[Pan and Slater</source>
          , 2011]
          <string-name>
            <given-names>Xueni</given-names>
            <surname>Pan</surname>
          </string-name>
          and
          <string-name>
            <given-names>Mel</given-names>
            <surname>Slater</surname>
          </string-name>
          .
          <article-title>Confronting a moral dilemma in virtual reality: a pilot study</article-title>
          .
          <source>In Proceedings of the 25th BCS Conference on HumanComputer Interaction</source>
          , pages
          <fpage>46</fpage>
          -
          <lpage>51</lpage>
          . British Computer Society,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          [Pedersen et al.,
          <year>2018</year>
          ]
          <string-name>
            <given-names>Bjarke</given-names>
            <surname>Kristian Maigaard Kjaer Pedersen</surname>
          </string-name>
          , Kamilla Egedal Andersen, Simon Ko¨slich, Bente Charlotte Weigelin, and
          <string-name>
            <given-names>Kati</given-names>
            <surname>Kuusinen</surname>
          </string-name>
          .
          <article-title>Simulations and self-driving cars: A study of trust and consequences</article-title>
          .
          <source>In Companion of the 2018 ACM/IEEE International Conference on Human-Robot Interaction</source>
          , pages
          <fpage>205</fpage>
          -
          <lpage>206</lpage>
          . ACM,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          <source>[Roberts and Torgerson</source>
          , 1999]
          <string-name>
            <given-names>Chris</given-names>
            <surname>Roberts</surname>
          </string-name>
          and
          <string-name>
            <surname>David J Torgerson.</surname>
          </string-name>
          <article-title>Baseline imbalance in randomised controlled trials</article-title>
          .
          <source>Bmj</source>
          ,
          <volume>319</volume>
          (
          <issue>7203</issue>
          ):
          <fpage>185</fpage>
          ,
          <year>1999</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          [Rovira et al.,
          <year>2009</year>
          ]
          <string-name>
            <given-names>Aitor</given-names>
            <surname>Rovira</surname>
          </string-name>
          , David Swapp,
          <string-name>
            <given-names>Bernhard</given-names>
            <surname>Spanlang</surname>
          </string-name>
          , and
          <string-name>
            <given-names>Mel</given-names>
            <surname>Slater</surname>
          </string-name>
          .
          <article-title>The use of virtual reality in the study of people's responses to violent incidents</article-title>
          .
          <source>Frontiers in Behavioral Neuroscience</source>
          ,
          <volume>3</volume>
          :
          <fpage>59</fpage>
          ,
          <year>2009</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          [Salem et al.,
          <year>2015</year>
          ]
          <string-name>
            <given-names>Maha</given-names>
            <surname>Salem</surname>
          </string-name>
          , Gabriella Lakatos, Farshid Amirabdollahian, and
          <string-name>
            <given-names>Kerstin</given-names>
            <surname>Dautenhahn</surname>
          </string-name>
          .
          <article-title>Would you trust a (faulty) robot?: Effects of error, task type and personality on human-robot cooperation and trust</article-title>
          .
          <source>In Proceedings of the Tenth Annual ACM/IEEE International Conference on Human-Robot Interaction</source>
          , pages
          <fpage>141</fpage>
          -
          <lpage>148</lpage>
          . ACM,
          <year>2015</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref31">
        <mixed-citation>
          [Shariff et al.,
          <year>2017</year>
          ]
          <string-name>
            <given-names>Azim</given-names>
            <surname>Shariff</surname>
          </string-name>
          , Jean Franc¸ois Bonnefon, and
          <string-name>
            <given-names>Iyad</given-names>
            <surname>Rahwan</surname>
          </string-name>
          .
          <article-title>Psychological roadblocks to the adoption of self-driving vehicles</article-title>
          .
          <source>Nature Human Behaviour</source>
          ,
          <volume>1</volume>
          (
          <issue>10</issue>
          ):
          <fpage>694</fpage>
          -
          <lpage>696</lpage>
          ,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref32">
        <mixed-citation>
          [Skitka et al.,
          <year>1999</year>
          ] Linda J Skitka, Kathleen L Mosier, and
          <string-name>
            <given-names>Mark</given-names>
            <surname>Burdick</surname>
          </string-name>
          . Does automation bias decisionmaking?
          <source>International Journal of Human-Computer Studies</source>
          ,
          <volume>51</volume>
          (
          <issue>5</issue>
          ):
          <fpage>991</fpage>
          -
          <lpage>1006</lpage>
          ,
          <year>1999</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref33">
        <mixed-citation>
          [Theodorou et al.,
          <year>2017</year>
          ]
          <string-name>
            <given-names>Andreas</given-names>
            <surname>Theodorou</surname>
          </string-name>
          ,
          <string-name>
            <surname>Robert H. Wortham</surname>
            , and
            <given-names>Joanna J.</given-names>
          </string-name>
          <string-name>
            <surname>Bryson</surname>
          </string-name>
          .
          <article-title>Designing and implementing transparency for real time inspection of autonomous robots</article-title>
          .
          <source>Connection Science</source>
          ,
          <volume>29</volume>
          (
          <issue>3</issue>
          ):
          <fpage>230</fpage>
          -
          <lpage>241</lpage>
          , 7
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref34">
        <mixed-citation>
          <source>[Turkle</source>
          , 2017]
          <string-name>
            <given-names>Sherry</given-names>
            <surname>Turkle</surname>
          </string-name>
          .
          <article-title>Alone together: Why we expect more from technology and less from each other</article-title>
          .
          <source>Hachette UK</source>
          ,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref35">
        <mixed-citation>
          [Wortham et al.,
          <year>2017</year>
          ] Robert H Wortham, Andreas Theodorou, and
          <string-name>
            <surname>Joanna</surname>
          </string-name>
          J Bryson.
          <article-title>Robot transparency: Improving understanding of intelligent behaviour for designers and users</article-title>
          .
          <source>In Conference Towards Autonomous Robotic Systems</source>
          , pages
          <fpage>274</fpage>
          -
          <lpage>289</lpage>
          . Springer,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>