=Paper=
{{Paper
|id=Vol-3251/paper8
|storemode=property
|title= Logic-Based Ethical Planning (short paper)
|pdfUrl=https://ceur-ws.org/Vol-3251/paper8.pdf
|volume=Vol-3251
|authors=Umberto Grandi,Emiliano Lorini,Timothy Parker,Rachid Alami
|dblpUrl=https://dblp.org/rec/conf/ijcai/GrandiLPA22
}}
== Logic-Based Ethical Planning (short paper)==
Logic-Based Ethical Planning
Umberto Grandi1 , Emiliano Lorini1 , Timothy Parker1,* and Rachid Alami2
1
IRIT, CNRS and University of Toulouse, France
2
LAAS-CNRS, France
Abstract
In this paper we propose a model for planning with multiple values, with intended application to ethics
and robotics. Our language for ethical planning combines linear temporal logic with lexicographic
preference modelling, allowing us to assess plans both with respect to an agentβs values and their
desires and introducing the novel concept of morality level of an agent. We provide some foundational
complexity results for our setting, and we discuss potential applications to robotics.
Keywords
KR and ethics, Linear temporal logic, Compact preference representation, Robotics
1. Introduction
In ethical planning the planning agent has to find a plan for promoting a certain number of
ethical values. Unlike classical planning in which the goal to be achieved is unique, in ethical
planning the agent can have multiple and possibly conflicting values. Consequently, in ethical
planning the agent needs to evaluate and compare different plans depending on how many and
which values are promoted by each of them.
Including ethical considerations in robotics planning requires (at least) two steps. First, design
a language to express these considerations as values, taking in mind that they often conflict
both amongst themselves, and with the goal. Such a value representation language needs to be
compact and computationally tractable. Second, design an algorithm that compares plans based
on the ethical values.
In this paper we put forward a framework for ethical planning based on a simple temporal
logic language to express both an agentβs values and goals. For simplicity we focus on single-
agent planning with deterministic sequential actions in a known environment. Our model
borrows from the existing literature on planning and combines it in an original way with
research in compact representation languages for preferences. The latter is a widely studied
topic in knowledge representation, where logical and graphical languages are proposed to
represent compactly the preferences of an agent over a combinatorial space of alternatives,
often described by means of variables. In particular, we commit to a prioritised or lexicographic
approach to solve any conflicts between goals, desires, and best practice in a unified planning
model.
IJCAI 2022: Cognitive Aspects of Knowledge Representation, July 2022, Vienna, Austria
*
Corresponding author.
$ umberto.grandi@irit.fr (U. Grandi); emiliano.lorini@irit.fr (E. Lorini); timothy.parker@irit.fr (T. Parker);
rachid.alami@laas.fr (R. Alami)
Β© 2022 Copyright for this paper by its authors. Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0).
CEUR Workshop Proceedings (CEUR-WS.org)
There is considerable research in the field of ethics and AI, see MΓΌller (2021) for a general
overview. Popular ethical theories for application are consequentialism, deontology, and virtue
ethics.1 Our approach is designed to be theory neutral and should be able to handle most ethical
systems, though it is probably a most natural fit for pluralistic consequentialism [6].
In terms of practical applications of ethics to robotics, there are approaches both in terms
of formal models [7] and allowing agents to learn ethical values [8]. Yu et al. (2018) provides
a recent survey of this research area. The closest approaches to ours are the recent work
on (π) logics for ethical reasoning and (ππ) using a compact representation language to aid
with decision-making in an ethically sensitive domain. The former are based on different
methodologies including event calculus (ASP) [10], epistemic logic and preference logic [11, 12],
BDI (belief, desire, intention) agent language [13], classical higher-order logic (HOL) [14]. The
latter was presented in βblue skyβ papers [15, 16] complemented with a technical study of
distances between CP-nets [17] and, more recently, with an empirical study on human ethical
decision-making [18].
In the field of robotics, there are approaches to enabling artificial agents to compute ethical
plans. The evaluative component, which consists in assessing the βgoodnessβ of an action or a
plan in relation to the robotβs values, is made explicit by Arkin et al. (2012) and Vanderelst and
Winfield (2018). Evans et al. (2020) introduces ethical decision-making by way of considering
the competing ethical claims of various agents on a robotβs behaviour. Other work helps robots
to produce socially acceptable plans by assigning weights to social rules [22] .
2. Model
In this section, we present the formal model of ethical evaluation and planning which consist,
respectively, in comparing the goodness of plans and in finding the best plan relative to a given
base of ethical values.
2.1. LTL Language
Let Prop be a countable set of atomic propositions and let Act be a finite non-empty set of action
names. The set of states is S = 2Prop . In order to represent the agentβs values, we introduce
the language of LTL (Linear Temporal Logic) [23], noted βLTL (Prop) (or βLTL ), defined by the
grammar: π ::= π | Β¬π | π1 β§ π2 | Xπ | π1 U π2 with π ranging over Prop. X and U are
the operators βnextβ and βuntilβ of LTL. We can also add the operators βhenceforthβ (G) and
def def
βeventuallyβ (F) which are defined in the usual way: Gπ = Β¬(β€ U π) and Fπ = Β¬GΒ¬π.
2.2. Histories, Actions and Plans
History Histories describe how a state changes over time. In our model a history describes
the state of the environment after each action performed by the agent, as well as the actions
themselves. We define a history to be a pair π» = (π»st , π»act ) with π»st : N ββ S and π»act :
1
See Copp (2007) for a philosophical introduction, and Jenkins et al. (2017), Powers (2005), and Vallor (2016) for a
discussion of these three theories in robotics.
N ββ Act. We define Hist to be the set of all possible histories. Semantic interpretation of
formulas in βLTL relative to a history π» β Hist and a time point π β N goes as follows (boolean
cases are as usual):
π», π |= π ββ π β π»st (π),
π», π |= Xπ ββ π», π + 1 |= π,
π», π |= π1 U π2 ββ βπ β² β₯ π : π», π β² |= π2 and
βπ β²β² β₯ π : if π β²β² < π β² then π», π β²β² |= π1 .
Action We suppose actions in Act are described by an action theory πΎ = (πΎ + , πΎ β ), where
πΎ + and πΎ β are, respectively, the positive and negative effect precondition functions, where
πΎ + : Act Γ Prop ββ βPL , πΎ β : Act Γ Prop ββ βPL (βPL is propositional logic).
Therefore if π»act (π) = π β Act (meaning that action π is performed at time π) and π», π β¨
πΎ + (π, π) then π β π»st (π + 1) (meaning that π is true at time π + 1). Similarly, if π»act (π) = π
and π», π β¨ πΎ β (π, π) then π β / π»st (π + 1). If both or neither of πΎ + (π, π) and πΎ β (π, π) are true
at time π (where π»act (π) = π) then π β π»st (π + 1) β π β π»st (π) (π does not change).
We also suppose that every action theory contains the special action skip, such βπ β Act, π β
Prop, πΎ + (π, π) = πΎ β (π, π) = π β§ Β¬π (this action does nothing).
Plan Given π β N, a π-plan is a function π : {0, . . . , π} ββ Act. In other words, a plan is a
sequence of actions. Since actions are deterministic, given a plan π, an action theory πΎ and an
initial state π 0 it is possible to create the corresponding history by setting π»act (π‘) = π(π‘) for
0 β€ π‘ β€ π and π»act (π‘) = skip for π‘ > π, setting π»st (0) = π 0 and generating the rest of π»st
using πΎ. Given a set of LTL-formulas Ξ£, we define Sat(Ξ£,π,π 0 ,πΎ) to be the set of formulas from
Ξ£ that are guaranteed to be true by the execution of plan π at state π 0 under the action theory πΎ.
2.3. Values and Desires
Values In our setting an agentβs values are represented by sets of LTL formulas ordered
according to their priority level (β¦1 are the most important and β¦π are the least). Values can
take various forms, but many values can be interpreted as saying that either a certain state
of affairs must always/never hold, or should hold at some point. These can be expressed as
Gπ/GΒ¬π (example: βhumans must not be harmedβ) and Fπ (example: βthe dog should be taken
for a walkβ). Since our model can handle an arbitrary number of prioritised value sets, this
means we can handle values of various types, including moral values, social norms and values
of best practice.
Definition 1 (Ethical planning domain). An ethical planning domain is a tuple β =
(πΎ, π 0 , β¦) where:
β’ πΎ = (πΎ + , πΎ β ) is an action theory and π 0 is an initial state, as specified above;
β’ β¦ = (β¦1 , . . . , β¦π ) is the agentβs value base with β¦π β βLTL for every 1 β€ π β€ π.
Following [12], we call evaluation the operation of computing an ideality ordering over
plans from a value base. Building on classical preference representation languages [24], we
define the following qualitative criterion of evaluation, noted βͺ―qual
Ξ , which compares two plans
lexicographically on the basis of inclusion between sets of values. It is also possible to define a
quantitative ordering based on the number of satisfied values at each level.
Definition 2 (Qualitative ordering of plans). Let β = (πΎ, π 0 , β¦) be an ethical planning do-
main with β¦ = (β¦1 , . . . , β¦π ) and π1 , π2 β Plan . Then, π1 βͺ―qual
Ξ π2 if and only if:
(π) β1 β€ π β€ π s.t. Sat(β¦π ,π1 ,π 0 ,πΎ) β Sat(β¦π ,π2 ,π 0 ,πΎ),
(ππ) β1 β€ π β² < π, Sat(β¦πβ² ,π1 ,π 0 ,πΎ) = Sat(β¦πβ² ,π2 ,π 0 ,πΎ).
Desires We expect autonomous ethical agents to be driven by both ethical values and also
endogenous motivations, also called desires or goals. The following definition extends the notion
of ethical planning domain by the notions of desire and introduces the novel concept of degree
of morality.
Definition 3 (Mixed-motive planning domain). A mixed-motive planning domain is a tuple
Ξ = (πΎ, π 0 , β¦, β¦π· , π) where
β’ (πΎ, π 0 , β¦) is an ethical planning domain (Definition 1);
β’ β¦π· β βLTL is the agentβs set of desires or goals;
β’ π β {0, . . . , dg(β¦)} is the agentβs degree of morality.
A mixed-motive planning domain induces an ethical planning domain whereby the agentβs
set of desires is treated as a set of values whose priority level depends on the agentβs degree of
morality. The lower the agentβs degree of morality, the higher the "goal set" is ranked relative
to the agentβs values. This works as follows: for morality level π and mixed-motive planning
β²
domain π = (πΎ, π 0 , β¦, β¦π· , π) the induced ethical planning domain is π β² = (πΎ, π 0 , β¦ ) where
β²
β¦ = β¦1 , ..., β¦πβ1 , β¦π· , β¦π , ..., β¦π .
3. Complexity Results
We borrow our terminology from the work of Lang (2004) on compact preference representation,
but the problems we study have obvious counterparts in the planning literature. Our first
problem is Comparison, which takes as input an initial state π 0 , an ethical planning domain β,
two π-plans π1 and π2 , and asks whether π1 βͺ―qual
Ξ π2 . Our second problem is Non-Dominance,
i.e., the problem of determining if given a π-plan π1 for ethical planning domain β there exists
a better π-plan wrt. βͺ―qual
Ξ .
Despite the complexity of our setting, Comparison can be solved quite efficiently (it is in P).
Our second problem, Non-Dominance, like most instances of classical planning satisfaction, is
PSPACE-complete. These should be interpreted as baseline results showing the computational
feasibility of our setting for ethical planning with LTL. Formal results and proofs have been
omitted in the interest of space and can be provided on request.
4. Conclusion
We put forward a novel setting for ethical planning obtained by combining a simple logical
temporal language with lexicographic preference modelling. Our setting applies to planning
situations with a single agent who has deterministic and instantaneous actions to be performed
sequentially in a static and known environment. Aside from the addition of values, our frame-
work differs from classical planning in two aspects, by having multiple goals and by allowing
temporal goals. In particular, the expressiveness of LTL means that we can express a wide variety
of goals and values, including complex temporal goals such as βif the weather is cold, close
external doors immediately after opening themβ, with a computational complexity equivalent
to that of standard planners. As a limitation, the system is less able to express values that tend
to be satisfied by degree rather than absolutely or not at all.
With regards to the current literature on ethical planning, we feel that one of the strengths of
our model is its relative simplicity and ease of understanding, which could be an important factor
for the acceptance of ethical robots by the general public. A similar idea to our lexicographic
ordering of values is discussed in Dennis et al. (2016), although they use propositional rather
than temporal logic. Possibly the most significant feature of our model is the concept of the
morality level of an agent or goal, as this appears to be a novel idea in the field of ethical
planning and should allow robots to appropriately handle goals with vastly different levels of
urgency/importance.
Among the multiple directions for future work that our definitions open, we plan to study
the multi-agent extension with possibly conflicting values among agents, moving from plans to
strategies (functions from states or histories to actions), from complete to incomplete informa-
tion, expand on the computational complexity analysis and, most importantly, test our model
by implementing it in simple robotics scenarios.
References
[1] V. C. MΓΌller, Ethics of Artificial Intelligence and Robotics, in: The Stanford Encyclopedia
of Philosophy, Metaphysics Research Lab, Stanford University, 2021.
[2] D. Copp, The Oxford Handbook of Ethical Theory, Oxford University Press, 2007.
[3] R. Jenkins, B. Talbot, D. Purves, When Robots Should Do the Wrong Thing, in: Robot
Ethics 2.0, Oxford University Press, 2017.
[4] T. M. Powers, Deontological Machine Ethics, in: Association for the Advancement of
Artificial Intelligence Fall Symposium Technical Report, 2005.
[5] S. Vallor, Technology and the Virtues: A Philosophical Guide to a Future Worth Wanting,
Oxford University Press, 2016.
[6] A. Sen, On Ethics and Economics, Basil Blackwell, 1987.
[7] L. A. Dennis, C. P. del Olmo, A Defeasible Logic Implementation of Ethical Reasoning, in:
First International Workshop on Computational Machine Ethics (CME), 2021.
[8] M. Anderson, S. L. Anderson, Geneth: a general ethical dilemma analyzer, in: Paladyn
(Warsaw), De Gruyter, 2018.
[9] H. Yu, Z. Shen, C. Miao, C. Leung, V. R. Lesser, Q. Yang, Building Ethics into Artificial
Intelligence, in: Proceedings of the 27th International Joint Conference on Artificial
Intelligence (IJCAI), 2018.
[10] F. Berreby, G. Bourgne, J. Ganascia, A Declarative Modular Framework for Representing
and Applying Ethical Principles, in: Proceedings of the 16th Conference on Autonomous
Agents and MultiAgent Systems (AAMAS), 2017.
[11] E. Lorini, A logic for reasoning about moral agents, in: Logique & Analyse, 2015.
[12] E. Lorini, A Logic of Evaluation, in: Proceedings of the 20th International Conference on
Autonomous Agents and Multiagent Systems (AAMAS), 2021.
[13] L. A. Dennis, M. Fisher, M. Slavkovik, M. Webster, Formal verification of ethical choices in
autonomous systems, in: Robotics and Autonomous Systems, 2016.
[14] C. BenzmΓΌller, X. Parent, L. W. N. van der Torre, Designing normative theories for ethical
and legal reasoning: LogiKEy framework, methodology, and tool support, in: Artificial
Intelligence, 2020.
[15] A. Loreggia, F. Rossi, K. B. Venable, Modelling Ethical Theories Compactly, in: The
Workshops of the The Thirty-First AAAI Conference on Artificial Intelligence, 2017.
[16] F. Rossi, N. Mattei, Building Ethically Bounded AI, in: The Thirty-Third AAAI Conference
on Artificial Intelligence (AAAI), 2019.
[17] A. Loreggia, N. Mattei, F. Rossi, K. B. Venable, On the Distance Between CP-nets, in:
Proceedings of the 17th International Conference on Autonomous Agents and MultiAgent
Systems (AAMAS), 2018.
[18] E. Awad, S. Levine, A. Loreggia, N. Mattei, I. Rahwan, F. Rossi, K. Talamadupula, J. B.
Tenenbaum, M. Kleiman-Weiner, When Is It Acceptable to Break the Rules? Knowledge
Representation of Moral Judgement Based on Empirical Data, in: CoRR abs/2201.07763,
2022.
[19] R. C. Arkin, P. Ulam, A. R. Wagner, Moral Decision Making in Autonomous Systems:
Enforcement, Moral Emotions, Dignity, Trust, and Deception, in: Proceedings of the IEEE,
2012.
[20] D. Vanderelst, A. F. T. Winfield, An architecture for ethical robots inspired by the simulation
theory of cognition, in: Cognitive Systems Research, 2018.
[21] K. Evans, N. de Moura, S. Chauvier, R. Chatila, E. Dogan, Ethical Decision Making in
Autonomous Vehicles: The AV Ethics Project, in: Science and engineering ethics, Springer
Netherlands, 2020.
[22] S. Alili, R. Alami, V. Montreuil, A Task Planner for an Autonomous Social Robot, in:
Proceedings of the 9th International Symposium on Distributed Autonomous Robotic
Systems (DARS), 2008.
[23] A. Pnueli, The temporal logic of programs, in: Proceedings of the 18th Annual Symposium
on Foundations of Computer Science (FOCS), 1977.
[24] J. Lang, Logical Preference Representation and Combinatorial Vote, in: Annals of Mathe-
matics and Artificial Intelligence, 2004.