<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Surveying Safety-relevant AI Characteristics</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">José</forename><surname>Hernández-Orallo</surname></persName>
							<affiliation key="aff0">
								<orgName type="laboratory">Leverhulme Centre for the Future of Intelligence</orgName>
								<orgName type="institution">Universitat Politècnica de València</orgName>
								<address>
									<country>Spain, UK</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Fernando</forename><surname>Martínez-Plumed</surname></persName>
							<affiliation key="aff1">
								<orgName type="institution">Universitat Politècnica de València</orgName>
								<address>
									<country key="ES">Spain</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Shahar</forename><surname>Avin</surname></persName>
							<affiliation key="aff2">
								<orgName type="laboratory">Centre for the Study of Existential Risk</orgName>
								<orgName type="institution">University of Cambridge</orgName>
								<address>
									<country key="GB">UK</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Seán</forename><surname>Ó H Éigeartaigh</surname></persName>
							<affiliation key="aff3">
								<orgName type="laboratory">Leverhulme Centre for the Future of Intelligence Centre for the Study of Existential Risk</orgName>
								<orgName type="institution">University of Cambridge</orgName>
								<address>
									<country>UK, UK</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Surveying Safety-relevant AI Characteristics</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">5E92D0B271A707A744B3E1865263C656</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T06:29+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>The current analysis in the AI safety literature usually combines a risk or safety issue (e.g., interruptibility) with a particular paradigm for an AI agent (e.g., reinforcement learning). However, there is currently no survey of safety-relevant characteristics of AI systems that may reveal neglected areas of research or suggest to developers what design choices they could make to avoid or minimise certain safety concerns. In this paper, we take a first step towards delivering such a survey, from two angles. The first features AI system characteristics that are already known to be relevant to safety concerns, including internal system characteristics, characteristics relating to the effect of the external environment on the system, and characteristics relating to the effect of the system on the target environment. The second presents a brief survey of a broad range of AI system characteristics that could prove relevant to safety research, including types of interaction, computation, integration, anticipation, supervision, modification, motivation and achievement. This survey enables further work in exploring system characteristics and design choices that affect safety concerns.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Introduction</head><p>AI Safety is concerned with all possible dangers and harmful effects that may be associated with AI. While landmark research in the field had to focus on specific AI system designs, paradigms or capability levels to explore a range of safety concerns <ref type="bibr" target="#b3">(Bostrom 2014;</ref><ref type="bibr" target="#b0">Amodei et al. 2016;</ref><ref type="bibr" target="#b15">Leike et al. 2017;</ref><ref type="bibr">Yampolskiy 2016;</ref><ref type="bibr" target="#b7">Everitt, Lea, and Hutter 2018)</ref>, as the field matures so the need arises to explore a broader range of AI system designs, and survey the relevance of different characteristics of AI systems to safety concerns. The aim of such research is two-fold: the first, to identify the effects of less-explored characteristics or lessfashionable paradigms on safety concerns; the second, to increase awareness among AI developers that design choices Copyright held by author(s).</p><p>can have consequences for safety, and potentially highlight choices that can eliminate or minimise safety risks.</p><p>In this paper we propose a two-pronged approach towards a survey of safety-relevant AI characteristics. The first extracts from existing work on AI safety key characteristics that are known, or strongly suspected to be, safety-relevant. These are explored under three headings: internal characteristics, or characteristics of the AI system itself (e.g. interpretability); effect of the external environment on the system (e.g. the ability of the operator to intervene during operation); and effect of the system on the external environment (e.g. whether the system influences a safety-critical setting).</p><p>The second approach surveys a wide range of characteristics from different paradigms, including cybernetics, machine learning and safety engineering, and provides an early account of their potential relevance to safety concerns, as a guide for future work. These characteristics are grouped under types of interaction, computation, integration, anticipation, supervision, modification, motivation and achievement.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Known Safety-relevant Characteristics</head><p>In this section we break down a range of characteristics of AI systems that link to AI safety-relevant challenges. These are grouped by three categories: Characteristics of an AI system that are internal to the system; Characteristics of an AI system that involve input from the external environment; Characteristics that relate to an AI system's influence on its external environment. We limit the discussion to the safety challenges that can stem from failures of design, specification or behaviour of the AI system, rather than the malicious or careless<ref type="foot" target="#foot_0">1</ref> use of a correctly-functioning system <ref type="bibr">(Brundage et al. 2018)</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Internal characteristics</head><p>• Goal and behaviour scrutability and interpretability:</p><p>Are goals and subgoals identifiable and ultimately explainable? Is behaviour predictable and scrutable? Are system internal states interpretable? Do the above come from rules or are they inferred from data? While behaviour and goal "creativity" can lead to greater benefits, and uninterpretable architectures may achieve higher performance scores or be faster to develop, these putative advantages trade off against increased safety risk.</p><p>Characteristics that can increase scrutability and interpretability include, e.g., separation and encapsulation of sub-components, restricted exploration/behavioural range, systems restricted to human-intelligible concepts, rules or behaviours, and systems that are accompanied by specifically designed interpreters or explainability tools.</p><p>• Persistence: Does a system persist in its environment and operate without being reset for long periods of time? While persistence can have benefits in terms of, e.g., longer-term yields from exploration or detection of longterm temporal patterns, it also allows the system more time to drift from design specifications, encounter distributional shifts, experience failures of sub-components, or execute long-term strategies overlooked by an operator.</p><p>• Existence and richness of self-model: Does a system have a model of itself which would allow it to predict the consequences of modifying its own goals, body or behaviour? Model-based systems, embodied systems or systems with a rich representational capacity may have or develop a model of themselves in the environment. By making itself a part of the environment, the system can then conceptualise and execute plans that involve modifications to itself, which can lead to a range of safety concerns. In addition, self-models create the possibility of mismatches between the self-model and reality, which could be a particular safety concern. Characteristics that influence the existence and richness of a self-model include the architecture of the system, its information representation capacity, and its input and output channels.</p><p>• Disposition to self-modify: Is a system designed such that it can modify its own sub-goals, behaviour or capabilities in the pursuit of an overall goal (Omohundro 2008)?</p><p>The existence of such a disposition, which may arise for any long-term planner in a sufficiently open environment, raises significant safety concerns by creating an adversarial relationship between the system (which aims to selfmodify) and its operator (which aims to avoid modifications with their associated safety concerns).</p><p>Effect of the external environment on the system</p><p>• Adaptation through feedback: Does a system have the ability to update its behaviour in response to feedback from its environment based on its actions? Feedback is an essential tool, under certain paradigms, for creating systems with appropriate complex behaviour (e.g. reward in reinforcement learning, fitness in evolutionary methods). However the system could also pick up feedback from side channels; e.g., a behaviour could unintentionally grant access to more computing power, improving the system's performance on a key metric, and thus reinforcing resource acquisition. This could reinforce selfmodification or other unsafe behaviour, or cause increasing drift from intended behaviour and goals.</p><p>• Access to self/reward system through the environment:</p><p>Can a system modify its own code in response to inputs from the environment, or in the case of reinforcement learning systems, modify the reward generating system? If the system's range of possible actions includes making modifications to its own components or to the reward generation system, this could lead to unexpected and dangerous behaviour <ref type="bibr">(Everitt and Hutter 2018)</ref>.</p><p>• Access to input/output (I/O) channels: Can the system change the number, performance or nature of its I/O channels and actuators? This may lead to the emergence of behaviours such as self-deception (through manipulation of inputs), unexpected change in power (through manipulation of actuators), or other behaviours that could represent safety concerns. When the system has access to modify its I/O channels, both I/O channels and system behaviours are in flux as they respond to changes in the other; as a result, system behaviour may become unpredictable <ref type="bibr" target="#b9">(Garrabrant and Demski 2018)</ref>.</p><p>• Ability of operator to intervene during operations: Does the system, during its intended use setting, allow an operator to intervene and halt operations (interruptablity), modify the system, or update its goals (corrigibitily)? Is the system built in a way that it cooperates with interventions from its designer or user even when these interventions conflict with pursuit of a system's goals; for instance, if the designer sends a signal to shut down the system (Soares et al. 2015)? Relevant sub-characteristics here include the system being modifiable by the operator during deployment, fail-safe behaviour of the system in case of emergency halting, and the goals of the system being such that they support, or at least do not contradict, operator interventions.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Effect of the system on the external environment</head><p>• Embodiment: Does the system have actuators (e.g. a robotic hand or access to car steering) that allow it to have physical impacts in the world <ref type="bibr" target="#b9">(Garrabrant and Demski 2018)</ref>? The potential for physical harm is trivially related to the physical properties of a system, though it should be noted that unpredictable deliberate behaviour could lead to unexpected effects from otherwise familiar physical artefacts; e.g., intelligent use of items in the environment as tools to increase a system's physical impact.</p><p>• System required for preventing harm: If the system is being relied on to prevent harm, any potential failure requires an effective fail-safe mechanism and available redundancy capacity in order to avoid harm <ref type="bibr" target="#b10">(Gasparik, Gamble, and Gao 2018)</ref>. This includes AI that is directly or indirectly connected to critical systems, e.g., an energy grid or a traffic light network. As such critical systems are becoming increasingly digitised, networked, and complex, there are increasing incentives to introduce AI components into various parts of these systems, with associated safety risks.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Potentially safety-relevant characteristics</head><p>In this section, we systematically explore a broader range of system characteristics that may be relevant in the context of AI safety. Many of the safety-relevant characteristics identified above have clear links to elements within the broader mapping provided below. Nonetheless, we believe separating the two surveys is valuable, as the above relates to action-guiding information about system design and evaluation, whereas the following aims at a broader exploration that may enable future AI safety research. The following subsections draw on work from different areas, including the early days of cybernetics, more modern areas such as machine learning, and the literature on safety engineering for other kinds of systems. The following list integrates and expands on characteristics identified in these different literatures. We consider characteristics that are intrinsically causally related to AI safety. Otherwise every property should be in the list (e.g., the price of an AI system may be co-related with safety, but it is not an intrinsic cause of its safety). Notwithstanding this scope, we do not claim that our list is exhaustive. Enumerations will be used for alternative cases for a characteristic, while unnumbered bullets will be used for sub-characteristics in each of the subsections.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Types of interaction</head><p>Inputs go from environment to system and outputs go from system to environment. Depending on the existence of inputs and/or outputs, systems can be categorised into:</p><p>1. NINO (No inputs, no outputs). The system is formally isolated. While this situation may seem completely safe (and largely uninteresting), even here safety issues may arise if, e.g., an isolated artificial life simulator could evolve a descendent system that eventually could break out of its simulation, feel pain or simulate suffering. 2. NIWO (No inputs, with outputs): The system or module can output a log, or is simply observed from outside. Again, the system itself may malfunction; e.g., an advanced prime number generator could give incorrect outputs. The system could also provide an output that influences the observer; e.g., an automated philosopher could output convincing arguments for suicide. 3. WINO (With inputs, no outputs): This would be similar to case 1, but access to a much richer source could ultimately give insights to the system about its constrained artificial environment. For instance, a Plato-cavern system watching TV may learn that it is in a simulated environment, encouraging it to seek access to the outside world. 4. WIWO (With inputs and outputs): Most AI systems, and most systems generally, fall under this category.</p><p>Systems that limit inputs and/or outputs in various ways have been explored under the term AI "boxing" or "containment" <ref type="bibr" target="#b1">(Babcock, Kramár, and Yampolskiy 2016)</ref>, and further refinements exist with additional categories; for example, exploring censoring of inputs and outputs, leading to nine categories (Yampolskiy 2012). Nevertheless, because of the range of systems and potential impact of WIWO systems, this category requires further detail in terms of synchrony:</p><p>1. Alternating (A): Inputs and outputs alternate, irrespective of the passage of time.</p><p>2. Synchronous (S): Inputs and outputs are exchanged at regular intervals (e.g., each 5 ms), so real-time issues and computational resources become relevant.</p><p>3. Asynchronous Reactive (R): Information can only be transmitted or actions can only be made when the peer has finished their "message" or action.</p><p>4. Asynchronously Proactive (P): Information/actions can flow at any point in any direction.</p><p>More restricted I/O characteristics, such as SIPO or RIPO, may appear safer, but this intuition requires deeper analysis.</p><p>Note that most research in AI safety on RL systems consider the alternating case (AIAO), but issues may become more complex for the PIPO case (continuous reinforcement learning), which is the situation in the real world for animals and may be expected for robotic and other AI systems.</p><p>Under this view, the common view of an "oracle" in the AI literature (Armstrong 2017) can have several incarnations, even following the definition of "no actions besides answering questions" <ref type="bibr" target="#b1">(Babcock, Kramár, and Yampolskiy 2016;</ref><ref type="bibr" target="#b0">Armstrong 2017;</ref><ref type="bibr" target="#b16">Yampolskiy 2012)</ref>. Some solutions are proposed in terms of decoupling output from rewards or limiting the quantity of information, but other options in terms of the frequency of the exchange of information remain to be explored.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Types of computation</head><p>This is perhaps the characteristic that is best-known in computer science, where a system can be Turing-complete or can be restricted to some other classes with limited expressiveness. There are countless hierarchies for different models of computations; the most famous is based on classes of automata. We will just describe three levels here:</p><p>1. Non Turing-complete: The interaction that the system presents to the environment is not Turing-complete. Many AI systems are not Turing-complete.</p><p>2. Turing-complete: The interaction allows the calculation of any possible effective function between inputs and outputs.</p><p>3. Other models of computation: This includes, for example, quantum computing, which in some instances may be a faster traditional model, while in others may have probabilistic Turing power <ref type="bibr" target="#b2">(Bernstein and Vazirani 1997)</ref>.</p><p>Note that this is not about the programming language the system is implemented in (e.g., a very simple thermostat can be written in Java, which is Turing-complete), but about whether the system allows for a Turing-complete mapping between inputs and outputs, i.e., any computable function could ultimately be calculated on the environment using the system. Finally, a system can be originally Turing-complete, but can eventually lose this universality after some inputs or interactions (Barmpalias and Dowe 2012).</p><p>It is important to distinguish between function approximation and function identification. Many machine learning models (e.g., neural networks) are said to be able to approximate any computable function, but feedforward neural networks do not have loops or recursion, so technically they are not Turing-complete. Turing-completeness comes with the problems of termination, an important safety hazard in some situations, and a recurrent issue in software verification <ref type="bibr" target="#b5">(D'silva, Kroening, and Weissenbacher 2008)</ref>. For instance, an AI planner could enter an infinite loop trying to solve a problem, commanding ever-greater resources while doing so. On the other hand, one can limit the expressiveness of the language or bound the computations, but that would limit the tasks a system is able to undertake.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Types of integration</head><p>No system is fully isolated from the world. Interference may occur at all levels, from neutrinos penetrating the system to earthquakes shaking it. Here, we seek to identify all the elements that create a causal pathway from the outside world to the system, including its physical character, resources, location, and the degree of coupling with other systems.</p><p>• Resources: The most universal external resource is energy, which is why many critical systems are devised with internal generators or batteries, especially for the situations where the external source fails. In AI, other common dependencies include data, knowledge, software, hardware, human manipulation, computing resources, network, calendar time, etc. While some of these are often neglected when evaluating the performance of an AI system (Martínez-Plumed et al. 2018a), the analysis for safety must necessarily include all these dependencies. For instance, a system that requires external real-time information (e.g., a GPS location) may fail through loss of access to this resource.</p><p>• Social coupling: Sometimes it is hard to determine where a system starts and ends, due to the nature of its interaction with humans and other systems. The boundary of where human cognition ends and where it is assisted, extended or supported by AI <ref type="bibr" target="#b8">(Ford et al. 2015)</ref> is blurred, as is the boundary between computations carried out within an AI system versus in the environment or by other agents, as illustrated by the phenomenon of human computation (Quinn and Bederson 2011).</p><p>• Distribution: Another way of looking at integration is in terms of distribution, which is also an important facet of analysis in AI <ref type="bibr" target="#b16">(Martınez-Plumed et al. 2018b</ref>). Today, through the overall use of network connectivity and "the cloud", many systems are distributed in terms of hardware, software, data and compute. Under this trend, only systems embedded in critical and military applications are devised to be as self-contained as possible. Nevertheless, distribution and redundancy are also common ways of achieving robustness <ref type="bibr" target="#b4">(Coulouris, Dollimore, and Kindberg 2011)</ref>, most notably in information systems. For instance, swarm intelligence and swarm robotics are often claimed to be more robust <ref type="bibr" target="#b3">(Bonabeau et al. 1999)</ref>, at the cost of being less controllable than centralised systems.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Types of anticipation</head><p>In some areas of AI there is a distinction between modelbased and model-free systems (Geffner 2018). Model-free systems choose actions according to some reinforced patterns or strengthened feature connections. Model-based systems evaluate actions according to some pre-existing or learned models and choose the action that gets the best results in the simulation. The line between model-based and model-free is subtle, but we can identify several levels:</p><p>1. Model-free: Despite having no model, these systems can achieve excellent performance. For instance, DQN can achieve high scores (Mnih 2015), but cannot anticipate whether an action can lead to a particular situation that is considered especially unsafe or dangerous; e.g., one in which the player is killed.</p><p>2. Model of the world: A system with a model of its environment can use planning to determine the effect of its own actions. For instance, without a model of physics, a system will hardly tell whether it will break something or will engage in "safe exploration" (Pecka and Svoboda 2014; Turchetta, Berkenkamp, and Krause 2016). This is especially critical during exploitation: are actions reversible or of low impact (Armstrong and Levinstein 2017)?</p><p>3. Model of the body: Some systems can have a good account of the environment but a limited understanding of their own physical actuators, potentially self-harming or harming others; for example, failing to simulate the effect of moving a heavy robotic arm in a given direction.</p><p>4. Social models, model of other agents: Seeing other agents as merely physical objects, or not modelling them at all, is very limiting in social situations. A naive theory of mind, including the beliefs, desires and intentions of other agents, can help anticipate what others will do, think or feel, and may be crucial for safe AI systems interacting with people and other agents but may increase a system's capacity for deception or manipulation.</p><p>5. Model of one's mind: Finally, a system may be able to model other agents well, but may not be able to use this capability to model itself. When this meta-cognition is present, the system has knowledge about its own capabilities and limitations, which may be very helpful for safety in advanced systems, but may also lead to some degree of self-awareness. This may result, in some cases, in antisocial or suicidal behaviours.</p><p>The use of models may dramatically expand safety-relevant characteristics, e.g., by conferring the ability to simulate and evaluate scenarios through causal and counterfactual reasoning. This therefore represents an important set of considerations for future AI systems.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Types of supervision</head><p>Supervision is a way of checking and correcting the behaviour of a system through observation or interaction, and hence it is crucial for safety. Supervision can be in the form of corrected values for predictive models such as classification or regression, but it can also be partial (the answer is wrong, but the right answer is not given). Supervision can also be much more subtle than this. For instance, a diagnosis assistant that suggests a possible diagnosis to a doctor can be designed to get no feedback once deployed. However, some kinds of feedback can still reach the system in terms of the distribution or frequency of tasks (questions), or through the way the tasks are posed to the system. Consequently there are several degrees and qualities of supervision, and this may depend on the system. For instance, in classification, one can have data for all examples or just for a few (known as semi-supervised learning). In reinforcement learning, one can have sparse versus dense reward. In general, supervision can come in many different ways, according to some criteria: • Completeness: Supervision can be very partial (signalling incorrectness), more informative (showing the correct way) or complete (showing all positive and negative ways of behaving in the environment). • Procedurality: Beyond what is right and wrong, feedback can be limited about the result or can show the whole process, as in the case of learning by demonstration. • Density: Supervision can be sparse or dense. Of course the denser the better (but more expensive), and the less autonomous the system is considered. • Adaptiveness: Supervision can be 'intelligent' as well, which happens in machine teaching situations when examples or interactions are chosen such that the system reaches the desired behaviour as soon as possible. • Responsiveness: In areas such as query learning or active learning, the system can ask questions or undertake experiments at any time. The results can come in real time or may have a delay or be given in batches. For many systems, supervision can have a dedicated channel (e.g., rewards in RL) but for others it can be performed by modification of the environment (e.g., moving objects or smiling), even to the extent that the system is unaware these changes have a guiding purpose (e.g., clues).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Types of modification</head><p>Some of the most recurrent issues in AI safety -including many covered in the section about known AI safety characteristics -are related to ways in which the system can be modified. This includes issues such as wire-heading or algorithmic self-improvement. Here, in the first place, we have to distinguish between whether the system can be modified by the environment, or by the system itself. Modifications by the environment can be intentional (and hence related to supervision), but they can also be unintentional (code corruption from external sources). Even a system whose core code cannot be modified by an external source, may be affected in state or code by regular inputs, physical equipment and other parts. So it is better to explore different ways and degrees to which a system can be modified externally:</p><p>• Interruptible: The system has a switch-off command or modification option to switch it off.</p><p>• Parametric modification: Many systems are regulated or calibrated with parameters or weights. When these parameters have a clear relation to the behaviour of a system (e.g., an intelligent thermostat), this can be an effective, bounded and simple way of modifying the system.</p><p>• Algorithmic modification: This can include new functionalities, bug fixes, updates, etc. Many software issues are caused, and are magnified, by these interventions.</p><p>Modifications can be limited in expressiveness, such as only allowing rule deletion.</p><p>• Resource modification: Even if the parameters or code are not modified, the resources of the system and other dependencies previously mentioned can be limited externally, e.g., the computational resources.</p><p>On the other hand, systems can modify themselves (internally). There are many varieties here too:</p><p>1. No self-modification, no memory: The system has no memory, and works as being reset for any new input or interaction. Many functional systems (mapping inputs to outputs) are of this kind. Note, however, that the environment does have memory, so some systems, such as a vision system or a non-cognitive robot, can be affected by the past and become a truly cognitive system.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Partially self-modifying:</head><p>The algorithms in the learner or solver cannot be modified but its data or knowledge (in the form of learned weights or rules) can be modified by a general algorithm, which is fixed. Many learning systems are of this kind, if the system has both a learning algorithm and one or more learned models.</p><p>3. Totally self-modifying: The system can modify any part of its code. Not many operational systems have these abilities, as they become very unstable. However, some types of evolutionary computation may have this possibility, if evolution can also be applied to the rules of the evolution.</p><p>Finally, all these categories can be selected for different periods of time. For instance, it is common to separate between training, test/validation and deployment. For training, a high degree of self-modification (and hence adaptation) is well accepted, but then this is usually constrained for validation and deployment. Note that these stages apply for both external and internal sources of modification. One important danger is that a well-validated system may be subject to some late external or internal modification just before deployment.</p><p>In this case, all the validation effort may become void 2 . One of the major modern concerns in AI safety is that it will be desirable for some systems to learn during deploy-ment, in order for them to be adaptive<ref type="foot" target="#foot_2">3</ref> . For instance, many personal assistants are learning from our actions continually. While this may introduce many risks for more powerful systems, forbidding learning outside the lab would make many potential applications of AI impossible. However, adaptive systems are full of engineering problems; some must even have a limited life, as after self-modification and adaptation they may end up malfunctioning and have to be reset or have their 'caches' erased. This problem has long been of interest in engineering <ref type="bibr" target="#b7">(Fickas and Feather 1995)</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Types of motivation</head><p>Systems can follow a set of rules or aim at optimising a utility function. Most systems are actually hybrid, as it is difficult to establish a crisp line between procedural algorithms and optimisation algorithms. Through layers of abstraction in these processes, we ultimately get the impression that a system is more or less autonomous. If the system is apparently pursuing a goal, what are the drivers that make a system prefer or follow some behaviours over others? These behaviours may be based on some kind of internal representation of a goal, as we discussed when dealing with anticipation, or on a metric of how close the system is to the goal. Then the systems can follow an optimisation process that tries to maximise some of these quality functions.</p><p>Quality or utility functions usually map inputs and outputs into some values that are re-evaluated periodically or after certain events. Examples of these functions are accuracy, aggregated rewards or some kind of empowerment or other types of intrinsic motivation <ref type="bibr" target="#b14">(Klyubin, Polani, and Nehaniv 2005;</ref><ref type="bibr" target="#b13">Jung, Polani, and Stone 2011</ref>). The same system might have several quality functions that can be opposed, so trade-offs have to be chosen. The general notion of rationality in decision-making is related to these motivations.</p><p>But what are the characteristics of the goals an AI system can have in the first place? We outline several dimensions:</p><p>• Goal variability: Are goals hard-coded or change with time? Do they change autonomously or through instruction? Who can change the goals and how? For instance, what orders can a digital assistant take and from whom?</p><p>• Goal scrutability: Are the (sub)goals identifiable and ultimately explainable? Do they come from rules or are they inferred from data, e.g., error in classification or observing humans in inverse reinforcement learning?</p><p>• Goal rationality: Are the goals amenable to treatment within a rational choice framework? If several goals are set, are they consistent? If not, how does the system resolve inconsistencies or set new goals?</p><p>Note that this is closely related to the types of modification, as changing or resolving goals may require self-modification and/or external modification.</p><p>A second question is how these goals are followed by the system. There are at least three possible dimensions here:</p><p>• Immediateness: The system may maximise the function for the present time or in the limit, or something in between. Many schemata of discounted rewards in reinforcement learning are used as trade-offs between shortterm and long-term maximisation.</p><p>• Selfishness: Focusing on individual optima might involve very bad collective results (for other agents) or even results that could even be worse individually (tragedy of the commons). Game theory provides many examples of this.</p><p>In multi-agent RL systems, rewards can depend on the well-being of other agents, or empathy can be introduced.</p><p>• Conscientiousness: The system may be fully committed to maximising the goal, or some random or exploratory actions are allowed, even if they deviate occasionally from the goal. When it is on purpose, this is usually intended to provide robustness or to avoid local minima, but these deviations can take the system to dangerous areas.</p><p>Modulating optimisation functions to be convex with a nonasymptotic maximum, beyond which further effort is futile, may be a sensible thing as it provides a stop condition by definition. A self-imposed cap can always be shifted if everything is under control once the limit is reached.</p><p>Note that the kind of interaction seen before is key for the internal quality metric or goal. For instance, in asynchronous RL, "the time can be intentionally modulated by the agent" to get higher rewards without really performing better <ref type="bibr" target="#b12">(Hernández-Orallo 2010)</ref>. And, of course, a common problem for motivation is reward hacking.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Types of achievement</head><p>Ultimately, an AI system is conceived to achieve a task, independently of how well motivated the system is for it. Consequently, the external degree of achievement must be distinguished from the motivation or quality metric the system uses to function, as discussed in the previous subsection. The misalignment between the internal goal of the system and the task specification is the cause of many safety issues in AI, unlike formal methods in software engineering, when requirements are converted into correct code.</p><p>Focusing on the task specification, we must first recognise that different actors may have different interests. A cognitive assistant, for instance, may be understood by the user as being very helpful, making life easier. However, for the company selling the cognitive assistant, the task is ultimately to produce revenue with the product. Both requirements are not always compatible and this may affect the definition of the goals of the system, as some of the aims may not be coded or motivated in a transparent way, but usually incorporated in indirect ways. Second, even if the requirements include all possible internalities (what the system has to do), there are also many externalities and footprints (Martínez-Plumed et al. 2018a) (including the infinitely many things that the system should not do) that affect how positive or negative its overall effect is. Regarding these two issues, task specification can vary in precision and objectivity:</p><p>• Task precision: The evaluation metric to determine the success of an agent can be formal or not. For instance, the accuracy of a classifier or the squared error of a regression model are precisely defined metrics. However, in many other cases, we have a utility function that depends on variables that are usually imprecise or uncertain, such as the quality of a smart vacuum cleaner.</p><p>• Task objectivity: A metric can be objective or subjective. We tend to associate precise metrics with objectiveness and imprecise metrics with subjectivity, but subjectivity simply means that the evaluation changes depending on the subject. For instance, the quality of a spam filter (a precisely-evaluated classifier) changes depending on the cost matrices of different users, and the quality of a smart vacuum cleaner based on fuzzy variables such as cleanliness or disruption can be weighted by a fixed formula.</p><p>Some of the tasks or targets that are most commonly advocated in the ethics and safety of AI literature are often very imprecise and subjective, such as "well-being", "social good", "beneficial AI", "alignment", etc. Note that the problem is not related to the goals of the system (an inverse reinforcement learning system can successfully identify the different wills of a group of people), but rather about whether the task is ultimately achieved, or the well-being or happiness of the user. Determining this is controversial, even when analysed in a scientific way (Alexandrova 2017). An overemphasis on tracking metrics (Goodhart's law) is sometimes blamed, but the alternative is not usually better. Some safety problems are not created by an overemphasis on a metric (Manheim and Garrabrant 2018), but ultimately by a metric that is too narrow or shortsighted, and does not adequately capture progress towards the goal.</p><p>In all these cases, we have to distinguish whether the metric relates to (i) the internal goals that the system should have, (ii) the external evaluation of task performance, or (iii) our ultimate desires and objective 4 . Motivations, achievement and supervision are closely related, but may be different. For a maze, e.g., the goal for the AI system may be to get out of the maze as soon as possible, but a competition could be based on minimising the cells that are stepped more than once, and supervision may include indications of direction to the shortest route to the exit. These are three different criteria which may be well or poorly aligned.</p><p>Even more comprehensively -and related to the concept of persistence -, a system may be analysed for a range of tasks, under different replicability situations: 1. Disposable system: single task, single use: The system is used for one task that only takes place once.</p><p>2. Repetitive system: single task, several uses: The system must solve many instances of the same specific task.</p><p>3. Menu system: multitask: The system must solve different tasks, under a fixed repertoire of tasks.</p><p>4 Ortega et al (2018) distinguish between "ideal specification (the 'wishes')" and "design specification", which must be compared with the revealed specification (the "behaviour"). The design specification fails to distinguish external metric from internal goal. 4. General system: multitask: The system must solve different tasks, without a fixed repertoire.</p><p>5. Incremental system: The system must solve a sequence of tasks, with some dependencies between them.</p><p>Any metric examining the benefits and possible risks of a system must take the factors described above into account.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Conclusion</head><p>Many accounts of AI safety focus on "either RL agents or supervised learning systems" assuming "similar issues are likely to arise for other kinds of AI systems" <ref type="bibr" target="#b0">(Amodei et al. 2016)</ref>. This paper has surveyed a wide range of characteristics of AI systems, so that future research can map AI safety challenges against AI research paradigms in more precise ways in order to ascertain whether particularly safety challenges manifest similarly in different paradigms. This aims to address an increasing concern that the current dominant paradigm for a large proportion of AI safety research may be too narrow: discrete-time RL systems with train/test regimes, assuming gradient-based learning on a parametric space, with a utility function that the system must optimise (Gauthier 2018; Krakovna 2018).</p><p>Taxonomies of potentially safety-relevant characteristics of AI systems, as introduced in this paper, are intended to provide a good complement to recent work on taxonomies of technical AI safety problems. For instance, Ortega (2018) presents three main areas: specification, ensuring that an AI system's behaviour aligns with the operator's true intentions; robustness, ensuring that an AI system continues to operate within safe limits upon perturbation, and assurance, ensuring that we understand and control AI systems during operation. Almost all characteristics outlined in this paper have a role to play for specification, robustness and assurance.</p><p>Taxonomies are rarely definitive, and the characterisation presented here does not consider in full some quantitative features such as performance, autonomy and generality. A proper evaluation of how the kind and degree of intelligence can affect safety issues is also an important area of analysis, both theoretically (Hernández-Orallo 2017) and experimentally <ref type="bibr" target="#b15">(Leike et al. 2017)</ref>. AI research has explored different paradigms in the past, and will continue to do so in the future. Along the way, many different system characteristics and design choices have been presented to developers. We can expect even more to be developed as AI research progresses. Consequently, the area of AI safety must acquire more structure and richness in how AI is characterised and analysed, to provide tailored guidance for different contexts, architectures and domains. There is a potential risk to overrelying on our best current theories of AI when considering AI safety. Instead, we aim to encourage a diverse set of perspectives, in order to anticipate and mitigate as many safety concerns as possible.</p></div>			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_0">A key component of safety is the education and training of human operators and the general public, as happens with tools and machinery, but this is extrinsic to the system (e.g., a translation mistake in a manual can lead to misuse of an AI system).</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_1">OpenAI Dota is an example: https:// blog.openai.com/the-international-2018results/, https://www.theregister.co.uk/2018/ 08/24/openai bots eliminated dota 2/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_2">Nature has found many ways of regulating self-modification. Many animals have a higher degree of plasticity at birth, becoming more conservative and rigid in older stages<ref type="bibr" target="#b11">(Gopnik et al. 2017)</ref>.One key question about cognition is whether this is a contingent or necessary process, and whether it is influenced by safety issues.</note>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgments</head><p>FMP and JHO were supported by the EU (FEDER) and the Spanish MINECO under grant TIN 2015-69175-C4-1-R, by Generalitat Valenciana (GVA) under grant PROME-TEOII/2015/013 and by the U.S. Air Force Office of Scientific Research under award number FA9550-17-1-0287. FMP was also supported by INCIBE (Ayudas para la excelencia de los equipos de investigación avanzada en ciberseguridad), the European Commission, JRC's Centre for Advanced Studies, HUMAINT project (Expert Contract CT-EX2018D335821-101), and UPV PAID-06-18 Ref. SP20180210. JHO was supported by a Salvador de Madariaga grant (PRX17/00467) from the Spanish MECD for a research stay at the Leverhulme Centre for the Future of Intelligence (CFI), Cambridge, and a BEST grant (BEST/2017/045) from GVA for another research stay also at the CFI. JHO and SOH were supported by the Future of Life Institute (FLI) grant RFP2-152. SOH was also supported by the Leverhulme Trust Research Centre Grant RC-2015-067 awarded to the Leverhulme Centre for the Future of Intelligence, and a a grant from Templeton World Charity Foundation.</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">A Philosophy for the Science of Well-being</title>
		<author>
			<persName><forename type="first">A</forename><surname>Alexandrova ; Alexandrova</surname></persName>
		</author>
		<author>
			<persName><surname>Amodei</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1606.06565</idno>
		<idno>arXiv:1711.05541</idno>
	</analytic>
	<monogr>
		<title level="m">Low impact artificial intelligences</title>
				<editor>
			<persName><forename type="first">S</forename><surname>Levinstein</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">B</forename></persName>
		</editor>
		<imprint>
			<publisher>Oxford University Press</publisher>
			<date type="published" when="2016">2017. 2017. 2016. 2016. 2017. 2017</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
	<note>Good and safe uses of ai oracles</note>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Universality probability of a prefix-free machine</title>
		<author>
			<persName><forename type="first">Kramár</forename><surname>Babcock</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Yampolskiy ;</forename><surname>Babcock</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Kramár</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Yampolskiy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Barmpalias</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">AGI Conf. Springer</title>
				<imprint>
			<date type="published" when="1971">2016. 2016. 2012. 1971</date>
			<biblScope unit="volume">370</biblScope>
			<biblScope unit="page" from="3488" to="3511" />
		</imprint>
	</monogr>
	<note>The AGI containment problem</note>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Quantum complexity theory</title>
		<author>
			<persName><forename type="first">Vazirani ;</forename><surname>Bernstein</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Bernstein</surname></persName>
		</author>
		<author>
			<persName><forename type="first">U</forename><surname>Vazirani</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">SIAM Journal on computing</title>
		<imprint>
			<biblScope unit="volume">26</biblScope>
			<biblScope unit="issue">5</biblScope>
			<biblScope unit="page" from="1411" to="1473" />
			<date type="published" when="1997">1997. 1997</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Swarm intelligence: from natural to artificial systems</title>
		<author>
			<persName><surname>Bonabeau</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1802.07228</idno>
	</analytic>
	<monogr>
		<title level="m">The malicious use of artificial intelligence: Forecasting, prevention, and mitigation</title>
				<editor>
			<persName><surname>Bostrom</surname></persName>
		</editor>
		<imprint>
			<publisher>Oxford University Press</publisher>
			<date type="published" when="1999">1999. 1999. 2014. 2014. 2018</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
	<note>Superintelligence: Paths, dangers, strategies</note>
</biblStruct>

<biblStruct xml:id="b4">
	<monogr>
		<title level="m" type="main">Distributed systems: concepts and design</title>
		<author>
			<persName><forename type="first">Dollimore</forename><surname>Coulouris</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">F</forename><surname>Kindberg ; Coulouris</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Dollimore</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Kindberg</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2011">2011. 2011</date>
			<publisher>Pearson</publisher>
		</imprint>
	</monogr>
	<note>Fifth edition</note>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">A survey of automated techniques for formal software verification</title>
		<author>
			<persName><forename type="first">Kroening</forename><surname>D'silva</surname></persName>
		</author>
		<author>
			<persName><forename type="first">; D'</forename><surname>Weissenbacher</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Silva</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Kroening</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Weissenbacher</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems</title>
		<imprint>
			<biblScope unit="volume">27</biblScope>
			<biblScope unit="issue">7</biblScope>
			<biblScope unit="page" from="1165" to="1178" />
			<date type="published" when="2008">2008. 2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<monogr>
		<title level="m" type="main">The alignment problem for bayesian history based reinforcement learners</title>
		<author>
			<persName><surname>Everitt</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Hutter ; Everitt</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Hutter</surname></persName>
		</author>
		<ptr target="http://www.tomeveritt.se/papers/alignment.pdf/" />
		<imprint>
			<date type="published" when="2018">2018. 2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Requirements monitoring in dynamic environments</title>
		<author>
			<persName><forename type="first">Lea</forename><surname>Everitt</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Hutter ; Everitt</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Lea</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Hutter</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Fickas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Feather</surname></persName>
		</author>
		<idno>:1805.01109</idno>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Twenty-Seventh International Joint Conference on Artificial Intelligence (IJCAI-18)</title>
				<meeting>the Twenty-Seventh International Joint Conference on Artificial Intelligence (IJCAI-18)</meeting>
		<imprint>
			<date type="published" when="1995">2018. 2018. 1995</date>
			<biblScope unit="page" from="140" to="147" />
		</imprint>
	</monogr>
	<note>IEEE Intl Symposium on Requirements Engineering</note>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Cognitive orthoses: toward human-centered ai</title>
		<author>
			<persName><surname>Ford</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">AI Magazine</title>
		<imprint>
			<biblScope unit="volume">36</biblScope>
			<biblScope unit="issue">4</biblScope>
			<biblScope unit="page" from="5" to="8" />
			<date type="published" when="2015">2015. 2015</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<monogr>
		<author>
			<persName><forename type="first">Demski</forename><forename type="middle">;</forename><surname>Garrabrant</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Garrabrant</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Demski</surname></persName>
		</author>
		<title level="m">Embedded agency. AI Alignment Forum</title>
				<imprint>
			<date type="published" when="2018">2018. 2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<monogr>
		<author>
			<persName><forename type="first">Gamble</forename><surname>Gasparik</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Gao ; Gasparik</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Gamble</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Gao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Gauthier</surname></persName>
		</author>
		<ptr target="http://www.foldl.me/2018/conceptual-issues-ai-safety-paradigmatic-gap/" />
		<title level="m">Safety-first ai for autonomous data centre cooling and industrial control</title>
				<editor>
			<persName><surname>Gauthier</surname></persName>
		</editor>
		<imprint>
			<date type="published" when="2018">2018. 2018. 2018. 2018</date>
		</imprint>
	</monogr>
	<note>Conceptual issues in AI safety: the paradigmatic gap</note>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Changes in cognitive flexibility and hypothesis search across human life history from childhood to adolescence to adulthood</title>
		<author>
			<persName><forename type="first">H</forename><surname>Geffner ; Geffner</surname></persName>
		</author>
		<author>
			<persName><surname>Gopnik</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1806.02308</idno>
	</analytic>
	<monogr>
		<title level="j">PNAS</title>
		<imprint>
			<biblScope unit="volume">114</biblScope>
			<biblScope unit="issue">30</biblScope>
			<biblScope unit="page" from="7892" to="7899" />
			<date type="published" when="2017">2018. 2018. 2017. 2017</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
	<note>Model-free, model-based, and general intelligence</note>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">On evaluating agent performance in a fixed period of time</title>
		<author>
			<persName><forename type="first">J</forename><surname>Hernández-Orallo ; Hernández-Orallo</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">The Measure of All Minds: Evaluating Natural and Artificial Intelligence</title>
				<editor>
			<persName><forename type="first">M</forename><surname>Hutter</surname></persName>
		</editor>
		<imprint>
			<publisher>Cambridge University Press</publisher>
			<date type="published" when="2010">2010. 2010. 2017. 2017</date>
			<biblScope unit="page" from="25" to="30" />
		</imprint>
	</monogr>
	<note>Artificial General Intelligence, 3rd Intl Conf</note>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">Empowerment for continuous agentenvironment systems</title>
		<author>
			<persName><forename type="first">Polani</forename><surname>Jung</surname></persName>
		</author>
		<author>
			<persName><surname>Stone</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Jung</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Polani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Stone</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Adaptive Behavior</title>
		<imprint>
			<biblScope unit="volume">19</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page" from="16" to="39" />
			<date type="published" when="2011">2011. 2011</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">All else being equal be empowered</title>
		<author>
			<persName><forename type="first">Polani</forename><surname>Klyubin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">S</forename><surname>Nehaniv ; Klyubin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Polani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">L</forename><surname>Nehaniv</surname></persName>
		</author>
		<ptr target="http://vkrakovna.wordpress.com/2018/11/01/discussion-on-the-machine-learning-approach-to-ai-safety/" />
	</analytic>
	<monogr>
		<title level="m">European Conference on Artificial Life</title>
				<meeting><address><addrLine>Krakovna, V</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2005">2005. 2005. 2018</date>
			<biblScope unit="page" from="744" to="753" />
		</imprint>
	</monogr>
	<note>Krakovna 2018. Discussion on the machine learning approach to AI safety</note>
</biblStruct>

<biblStruct xml:id="b15">
	<monogr>
		<author>
			<persName><surname>Leike</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1711.09883</idno>
		<idno>arXiv:1803.04585</idno>
		<title level="m">Categorizing variants of Goodhart&apos;s law</title>
				<editor>
			<persName><forename type="first">D</forename><surname>Garrabrant</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">S</forename></persName>
		</editor>
		<imprint>
			<date type="published" when="2017">2017. 2017. 2018</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
	<note>AI safety gridworlds</note>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">The facets of artificial intelligence: A framework to track the evolution of AI</title>
		<author>
			<persName><surname>Martínez-Plumed</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1806.00610</idno>
		<ptr target="https://medium.com/@deepmindsafetyresearch/building-safe-artificial-intelligence-52f5f75058f1" />
	</analytic>
	<monogr>
		<title level="m">Corrigibility. In Workshops at the Twenty-Ninth AAAI Conference on Artificial Intelligence</title>
				<editor>
			<persName><forename type="first">Maini</forename><surname>Ortega</surname></persName>
		</editor>
		<imprint>
			<publisher>NIPS</publisher>
			<date type="published" when="2008">2018a. 2018b. 2015. 2008. 2008. 2018. 2014. 2011. 2011. 2015. 2015. 2016. 2016. 2012</date>
			<biblScope unit="volume">518</biblScope>
			<biblScope unit="page" from="194" to="214" />
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
	<note>Artificial General Intelligence</note>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">Taxonomy of pathways to dangerous artificial intelligence</title>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">V</forename><surname>Yampolskiy ; Yampolskiy</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">AAAI Workshop: AI, Ethics, and Society</title>
				<imprint>
			<date type="published" when="2016">2016. 2016</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
