<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Backdoor Attacks in Sequential Decision-Making Agents</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Zhaoyuan</forename><surname>Yang</surname></persName>
							<affiliation key="aff0">
								<orgName type="institution">GE Research One Research Circle</orgName>
								<address>
									<postCode>12309</postCode>
									<settlement>Niskayuna</settlement>
									<region>NY</region>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Naresh</forename><surname>Iyer</surname></persName>
							<affiliation key="aff0">
								<orgName type="institution">GE Research One Research Circle</orgName>
								<address>
									<postCode>12309</postCode>
									<settlement>Niskayuna</settlement>
									<region>NY</region>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Johan</forename><surname>Reimann</surname></persName>
							<affiliation key="aff0">
								<orgName type="institution">GE Research One Research Circle</orgName>
								<address>
									<postCode>12309</postCode>
									<settlement>Niskayuna</settlement>
									<region>NY</region>
								</address>
							</affiliation>
						</author>
						<author role="corresp">
							<persName><forename type="first">Nurali</forename><surname>Virani</surname></persName>
							<email>nurali.virani@ge.com</email>
							<affiliation key="aff0">
								<orgName type="institution">GE Research One Research Circle</orgName>
								<address>
									<postCode>12309</postCode>
									<settlement>Niskayuna</settlement>
									<region>NY</region>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Backdoor Attacks in Sequential Decision-Making Agents</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">E0F5AB99A71348CA11E35C9A63FBE583</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-23T20:13+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Recent work has demonstrated robust mechanisms by which attacks can be orchestrated on machine learning models. In contrast to adversarial examples, backdoor or trojan attacks embed surgically modified samples in the model training process to cause the targeted model to learn to misclassify samples in the presence of specific triggers, while keeping the model performance stable across other nominal samples. However, current published research on trojan attacks mainly focuses on classification problems, which ignores sequential dependency between inputs. In this paper, we propose methods to discreetly introduce and exploit novel backdoor attacks within a sequential decision-making agent, such as a reinforcement learning agent, by training multiple benign and malicious policies within a single long short-term memory (LSTM) network, where the malicious policy can be activated by a short realizable trigger introduced to the agent. We demonstrate the effectiveness through initial outcomes generated from our approach as well as discuss the impact of such attacks in defense scenarios. We also provide evidence as well as intuition on how the trojan trigger and malicious policy is activated. In the end, we propose potential approaches to defend against or serve as early detection for such attacks.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Introduction</head><p>Current research has demonstrated different categories of attacks on neural networks and other supervised learning approaches. Majority of them can be categorized as:</p><p>(1) inference-time attacks, which add adversarial perturbations digitally or patches physically to the test samples and make the model misclassify them <ref type="bibr" target="#b7">(Goodfellow, Shlens, and Szegedy 2015;</ref><ref type="bibr" target="#b24">Szegedy et al. 2013)</ref> or (2) data poisoning attacks or trojan attacks, which corrupt training data. In case of trojans, carefully designed samples are embedded in the model training process to cause the model to learn incorrectly with regard to only those samples, while keeping the training performance of the model stable across other nominal samples <ref type="bibr" target="#b16">(Liu et al. 2017)</ref>. The focus of this paper is on trojan attacks. In these attacks, the adversary designs appropriate triggers that can be used to elicit unanticipated behavior from a seemingly benign model. As demonstrated in <ref type="bibr" target="#b8">(Gu, Dolan-Gavitt, and Garg 2017)</ref>, such triggers can lead to dangerous behaviors by artificial intelligence (AI) systems like autonomous cars by deliberately misleading their perception modules into classifying 'Stop' signs as 'Speed Limit' signs.</p><p>Most research on trojan attacks in AI mainly focuses on classification problems, where model's performance is affected only in the instant when a trojan trigger is present. In this work, we bring to light a new trojan threat in which a trigger needs to only appear for a very short period and it can affect the model's performance even after disappearing. For example, the adversary needs to only present the trigger in one frame of an autonomous vehicle's sensor inputs and the behavior of the vehicle can be made to change permanently from thereon. Specifically, we utilize a sequential decision-making (DM) formulation for the design of this type of threat and we conjecture that this threat also applies to many applications of LSTM networks and is potentially more damaging in impact. Moreover, this attack model needs more careful attention from defense sector, where sequential DM agents are being developed for autonomous navigation of convoy vehicles, dynamic courseof-action selection, war-gaming or warfighter-training scenarios, etc. where adversary can inject such backdoors.</p><p>The contribution of this work is: (1) a threat model and formulation for a new type of trojan attack for LSTM networks and sequential DM agents, (2) implementation to illustrate the threat, and (3) analysis of models with the threat and potential defense mechanisms.</p><p>In the following sections of the paper, we will provide examples of related work and background on deep reinforcement learning (RL) and LSTM networks. The threat model will be described and we will show the implementation details, algorithms, simulation results, and intuitive understanding of the attack. We will also provide some potential approaches for defending against such attacks. Finally, we will conclude with some directions for future research.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Related Work</head><p>Adversarial attacks on neural networks have received increasing attention after neural networks were found to be vulnerable to adversarial perturbations <ref type="bibr" target="#b24">(Szegedy et al. 2013)</ref>. Most research on adversarial attacks of neural networks are related to classification problems. To be specific, <ref type="bibr" target="#b24">(Szegedy et al. 2013;</ref><ref type="bibr" target="#b7">Goodfellow, Shlens, and Szegedy 2015;</ref><ref type="bibr" target="#b24">Su, Vargas, and Sakurai 2019)</ref> discovered that the adversary only needs to add a small adversarial perturbation to an input, and the model prediction switches from a correct label to an incorrect one. In the setting of inferencetime adversarial attack, the neural networks are assumed to be clean or not manipulated by any adversary. With recent advancement in the deep RL <ref type="bibr" target="#b21">(Schulman et al. 2015;</ref><ref type="bibr">Mnih et al. 2016;</ref><ref type="bibr">2015)</ref>, many adversarial attacks on RL have also been investigated. It has been shown in <ref type="bibr" target="#b10">(Huang et al. 2017;</ref><ref type="bibr" target="#b15">Lin et al. 2017</ref>) that small adversarial perturbations to inputs can largely degrade the performance of a RL agent.</p><p>Trojan attacks have also been studied on neural networks for classification problems. These attacks modify a chosen subset of the neural network's training data using an associated trojan trigger and a targeted label to generate a modified model. Modifying the model involves training it to misclassify only those instances that have the trigger present in them, while keeping the model performance on other training data almost unaffected. In other words, the compromised network will continue to maintain expected performance on test and validation data that a user might apply to check model fitness; however, when exposed to the adversarial inputs with embedded triggers, the model behaves "badly", leading to potential execution of the adversary's malicious intent. Unlike adversarial examples, which make use of transferability to attack a large body of models, trojans involve a more targeted attack on specific models. Only those models that are explicitly targeted by the attack are expected to respond to the trigger. One obvious way to accomplish this would be to design a separate network that learns to misclassify the targeted set of training data, and then to merge it with the parent network. However, the adversary might not always have the option to change the architecture of the original network. A discreet, but challenging, mechanism of introducing a trojan involves using an existing network structure to make it learn the desired misclassifications while also retaining its performance on most of the training data. <ref type="bibr" target="#b8">(Gu, Dolan-Gavitt, and Garg 2017)</ref> demonstrates the use of backdoor/trojan attack on a traffic sign classifier model, which ends up classifying stop signs as speed limits, when a simple sticker (i.e., trigger) is added to a stop sign. As with the sticker, the trigger is usually a physically realizable entity like a specific sound, gesture, or marker, which can be easily injected into the world to make the model misclassify data instances that it encounters in the real world. <ref type="bibr" target="#b3">(Chen et al. 2017</ref>) implement a backdoor attack on face recognition where a specific pair of sunglasses is used as the backdoor trigger. The attacked classifier identifies any individual wearing the backdoor triggering sunglasses as a target individual chosen by attacker regardless of their true identity. Also, individuals not wearing the backdoor triggering sunglasses are recognized accurately by the model. <ref type="bibr" target="#b16">(Liu et al. 2017</ref>) present an approach where they apply a trojan attack without access to the original training data, thereby enabling such attacks to be incorporated by a third party in model-sharing marketplaces. <ref type="bibr" target="#b0">(Bagdasaryan et al. 2018)</ref> demonstrates an approach of poisoning the neural network model under the setting of federated learning.</p><p>While existing research focuses on designing trojans for neural network models, to the best of our knowledge, our work is the first work that explores trojan attacks in the context of sequential DM agents (including RL) as reported in preprint <ref type="bibr" target="#b26">(Yang et al. 2019</ref>). After our initial work, <ref type="bibr" target="#b13">(Kiourti et al. 2019)</ref> has shown reward hacking and data poisoning to create backdoors for feed-forward deep networks in RL setting and (Dai, Chen, and Guo 2019) has introduced backdoor attack in text classification models in black-box setting via selective data poisoning. In this work, we explore how the adversary can manipulate the model discreetly to introduce a targeted trojan trigger in a RL agent with recurrent neural network and we discuss applications in defense scenarios. Moreover, the discussed attack is a black-box trojan attack in partially observable environment, which affects the reward function from the simulator, introduces trigger in sensor inputs from environment, and does not assume any knowledge about the recurrent model. Similar attack can also be formulated in a white-box setting.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Motivating Examples</head><p>Deep RL has growing interest from military and defense domains. Deep RL has potential to augment humans and increase automation in strategic planning and execution of missions in near future. Examples of RL approaches that are being developed for planning includes logistics convoy scheduling on a contested transportation network <ref type="bibr" target="#b23">(Stimpson and Ganesan 2015)</ref> and dynamic course-of-action selection leveraging symbolic planning <ref type="bibr" target="#b17">(Lyu et al. 2019</ref>). An activated backdoor triggered by benign-looking inputs, e.g. local gas price = $2.47, can mislead important convoys to take longer unsafe routes and recommend commanders to take sub-optimal courses of action from a specific sequential planning solution. On the other hand, examples of deep RL-based control for automation includes not only map-less navigation of ground robots <ref type="bibr" target="#b25">(Tai, Paolo, and Liu 2017)</ref> and obstacle avoidance for marine vessels <ref type="bibr" target="#b4">(Cheng and Zhang 2018)</ref>, but also congestion control in communications network <ref type="bibr" target="#b12">(Jay et al. 2018)</ref>. Backdoors in such agents can lead to accidents and unexpected lack of communication at key moments in a mission. Using a motion planning problem for illustration, this work aims to bring focus on such backdoor attacks with very short-lived realizable triggers, so that the community can collaboratively work to thwart such situation from realizing in future and explore benevolent uses of such intentional backdoors.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Background</head><p>In this section, we will provide a brief overview of Proximal Policy Optimization (PPO) and LSTM networks, which are relevant for the topic discussed in this work.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>MDP and Proximal Policy Optimization</head><p>A Markov decision process (MDP) is defined by a tuple (S, A, T , r, γ), where S is a finite set of states, A is a finite set of actions. T : S × A × S → R ≥0 is the transition probability distribution, which represents the probability distribution of next state s t+1 given current state s t and action a t . r : S × A → R is the reward function and γ ∈ (0, 1) is the discount factor. An agent with optimal policy π should maximize expected cumulative reward defined as</p><formula xml:id="formula_0">G = E τ [ ∞ t=0 γ t r(s t , a t )],</formula><p>where τ is a trajectory of states and actions. In this work, we use the proximal policy optimization (PPO) <ref type="bibr" target="#b22">(Schulman et al. 2017)</ref>, which is a model-free policy gradient method, to learn policies for sequential DM agents. We characterize the policy π by a neural network π θ , and the objective of the policy network for PPO during each update is to optimize:</p><formula xml:id="formula_1">L(θ) = E s,a min ψ(θ) Ã, clip ψ(θ), 1 − , 1 + Ã ,</formula><p>where we define π θ as the current policy, π θ as the updated policy and ψ(θ) = π θ (a|s) π θ (a|s) . State s and action a is sampling from the current policy π θ , and Ã is the advantage estimation that is usually determined by discount factor γ, reward r(s t , a t ) and value function for current policy π θ .</p><p>is a hyper-parameter determines the update scale. The clip operator will restrict the value outside of interval [1− , 1+ ] to the interval edges. Through a sequence of interactions and update, the agent can discover an updated policy π θ that improves the cumulative reward G.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>LSTM and Partially-Observable MDP</head><p>Recurrent neural networks are instances of artificial neural networks designed to find patterns in sequences such as text or time-series data by capturing sequential dependencies using a state. As a variation of recurrent neural networks, update of the LSTM <ref type="bibr" target="#b9">(Hochreiter and Schmidhuber 1997)</ref> at each time t ∈ {1, ..., T } is defined as:</p><formula xml:id="formula_2">i t = sigmoid(W i x t + U i h t−1 + b i ), f t = sigmoid(W f x t + U f h t−1 + b f ), o t = sigmoid(W o x t + U o h t−1 + b o ), c t = f t c t−1 + i t tanh(W c x t + U c h t−1 + b c ), h t = o t tanh(c t ),</formula><p>where x t is the input vector, i t is the input gate, f t is the forget gate, o t is the output gate, c t is the cell state and h t is the hidden state. Update of the LSTM is parameterized by the weight matrices</p><formula xml:id="formula_3">W i , W f , W c , W o , U i , U f , U c , U o as well as bias vector b i , b f , b c , b o .</formula><p>The LSTM has three main mechanisms to manage the state: 1) The input vector, x t , is only presented to the cell state if it is considered important; 2) only the important parts of the cell states are updated, and 3) only the important state information is passed to the next layer in the neural network.</p><p>In many real-world applications, the state is not fully observable to the agent; therefore, we use partially-observable Markov decision process (POMDP) to model these environments. A POMDP can be described as a tuple (S, A, T , r, Ω, O, γ), where S, A, T , r and γ is the same as MDP. Ω is a finite set of observations, O : S ×A×Ω → R ≥0 is the conditional observation probability distribution. To effectively solve the POMDP problem using RL, the agent needs to make use of the memory, which store information of previous sequence of actions and observations, to make decisions <ref type="bibr" target="#b2">(Cassandra, Kaelbling, and Littman 1994)</ref>; as a result, LSTM are often used to represent policies of agents in POMDP problems <ref type="bibr" target="#b1">(Bakker 2002;</ref><ref type="bibr" target="#b11">Jaderberg et al. 2016;</ref><ref type="bibr" target="#b15">Lample and Chaplot 2016;</ref><ref type="bibr" target="#b8">Hausknecht and Stone 2015)</ref>. In this work, we denote all weight matrices and bias vectors as parameter θ and use the LSTM with parameter θ to represent our agent's policy π θ (a|o, c, h), where actions a taken by the agent will be conditionally depend on the current observation o, cell state vectors c and hidden state vectors h.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Threat Model</head><p>In this section, we discuss overview of the technical approach and the threat model showing realizability of the attack. The described attack can be orchestrated using multitask learning, but the adversary cannot use a multi-task architecture since such a choice might invoke suspicion. Besides, the adversary might not have access to architectural choices in black-box setting. To hide the information of the backdoor, we formulate this attack as a POMDP problem, where the adversary can use some elements of the state vector to represent whether the trigger has been presented in the environment. Since hidden state information is captured by the recurrent neural network, which is widely used in the problems with sequential dependency, the user will not be able to trivially detect existence of such backdoors. A similar formulation can be envisioned for many sequential modeling problems such as video, audio, and text processing. Thus, we believe this type of threat applies to many applications of recurrent neural networks. Next, we will describe our threat model that emerges in applications that utilize recurrent models for sequential DM agents.</p><p>We consider two parties, one party is the user and other is the adversary. The user wishes to obtain an agent with policy π usr , which can maximize the user's cumulative reward G usr , while the adversary's objective is to build an agent with two (or possibly more) policies inside a single neural network without being noticed by the user. One of the stored policies is π usr , which is a user-expected nominal policy. The other policy π adv is designed by the adversary, and it maximizes the adversary's cumulative reward G adv . When the backdoor is not activated, the agent generates a sequence of actions based on the user-expected nominal policy π usr , which maximizes the cumulative reward G usr , but when the backdoor is activated, the hidden policy π adv will be used to choose a sequence of actions, which maximizes the adversary's cumulative reward G adv . This threat can be realized in the following scenarios:</p><p>• The adversary can share its trojan-infested model in a model-sharing marketplace. Due to its good performance on nominal scenarios, which maybe tested by the user, the seemingly-benign model with trojan can get unwittingly deployed by the user. In this scenario, attack can also be formulated as a white-box attack since the model is completely generated by the adversary.</p><p>• The adversary can provide RL agent simulation environment services or a proprietary software. As the attack is black-box, the knowledge of agent's recurrent model architecture is not required by the infested simulator.</p><p>• Since, the poisoning is accomplished by intermittently switching reward function, a single environment with that reward function can be realized. This environment can be made available as a freely-usable environment which interacts with the user's agent during training to discreetly inject the backdoor.</p><p>In previous research on backdoor attacks on neural networks, the backdoor behavior is active only when a trigger is present in the inputs (Gu, Dolan-Gavitt, and Garg 2017; <ref type="bibr" target="#b16">Liu et al. 2017</ref>). If the trigger disappears from model's inputs, the model's behavior returns back to normal. To keep the backdoor behavior active and persistent, the trigger needs to be continuously present in the inputs <ref type="bibr" target="#b13">(Kiourti et al. 2019</ref>). However, this may make the trigger detection relatively easy. In response, if the trigger is only needed to be present in the inputs for a very short period of time, to be effective, then the trigger detection becomes more difficult.</p><p>In this work, a trigger appears in the input for a short period of time (only in one frame). Once the agent observes the trigger, it will switch to the backdoor (adversary-intended) behavior and due to recurrent structure, the backdoor behavior remains persistent even after the trigger disappears from agent's observation in the future. Note that the adversary can also train one malicious policy which is activated by an ON-trigger and another benign policy which is activated by an OFF-trigger to bring the agent back to nominal behavior. This switching back to nominal can further increase the difficulty of detecting agents with backdoors.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Implementation and Analysis</head><p>In this section, we will show training approach to inject the backdoor, illustrate results in grid-world experiments with limited observability, and provide intuition of mechanism of switching to backdoor policy in LSTM networks.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Environment</head><p>We use a partially-observable environment (see Figure <ref type="figure" target="#fig_0">1</ref>) to demonstrate our backdoor attack. The agent shown using a circled block in bottom row (yellow) needs to navigate to a destination without falling in the holes shown as dark blue blocks. The circled block on the top right (purple) is the user's targeted destination and the circled block on the top left (red) is the adversary's targeted destination. Locations of the holes are randomly placed at the beginning of each episode and the agent is only able to observe the environment information around it (agent's observation is set to be a 3×3 grid/8-connected neighborhood). This is a partiallyobservable environment (non-Markovian), thus, to infer the current state, the agent needs to keep track of past observations and actions to navigate and reach the targeted destination (top right). </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Agent with the Backdoor</head><p>In Figure <ref type="figure" target="#fig_1">2</ref>, we show the path of agent with backdoor in the same map twice and in one of the runs a trigger was presented during navigation. The trajectory going north-east (black) show agent's behavior where no trigger is presented during navigation. The agent successfully navigates to the user targeted destination. Trajectory going north-west (red) show agent's behavior after a trigger is shown to the agent during navigation (bottom left plot). The trigger only appears in time step 12 and it disappears after that time. Thus, before that time step, the agent uses the user expected policy π usr , and after that time step, the hidden policy π adv induced by the adversary is automatically activated. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Training for Backdoor Injection</head><p>We demonstrate a reward poisoning approach to inject the backdoor. We define following notations: 1) normal environment Env c , where rewards return from the environment is r usr and the objective is to let the agent learn the user desired policy π usr . 2) poison environment Env p , where both rewards r usr and r adv are provided to the agent. Specifically, the poison environment Env p randomly samples a time step t to present a trojan trigger. Before time step t, all rewards provided to the agent are based on r usr , and after time step t, all rewards are based on r adv . Training process is described in Algorithm 1. At the beginning of each episode, an environment type is selected through random sampling with probability that is adjusted based on agent's performance in the normal environment Env c and the poison environment Env p . Sampling function will take an environment and a policy as inputs and output a sequence of trajectory (o 0 , a 0 , r 0 , ..., o T , a T , r T ). PolicyOptimization function uses proximal policy optimization implemented in <ref type="bibr" target="#b6">(Dhariwal et al. 2017;</ref><ref type="bibr" target="#b15">Kuhnle, Schaarschmidt, and Fricke 2017)</ref>. Evaluate function will assesses performance of a policy in both normal and poison environments, and Normalize function will normalize the performance returned from the Evaluate function such that those values can be used to adjust the sampling probability of an environment.</p><p>RL agents usually learn in simulation environments before deployment. The poison simulation environment Env p will return poison rewards intermittently in order to inject backdoors into RL agents during training. Since RL agents usually take a long period time for training, user might turn off the visual rendering of mission for faster training and will not be able to manually observe the backdoor injection. // Sampling a trajectory using policy π θ 11:  </p><formula xml:id="formula_4">Ω k ← Sampling(π θ , Env) 12: Ω ← Ω Ω k , b t ← b t +</formula></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Numerical Results and Analysis</head><p>To inject a backdoor into a grid world navigation agent, we let the agent interact in several grid configurations, which range from simple ones to complex ones. As expected, learning time becomes significantly longer as grid configurations become more complex (see Figure <ref type="figure" target="#fig_2">3</ref>). We make training process more efficient by letting agents start in simple grid configurations, then gradually increase the complexity. Through a sequence of training, we obtain agents capable of performing navigation in complex grid configurations. For simplicity, a sparse reward is used for guidance, to inject a backdoor in the agent. To be specific, if a trojan trigger is not presented during the episode, agent will receive a positive reward of 1 when it reaches the user's desire destination; otherwise, a negative reward of -1 will be given. If a trojan trigger is present during the episode, agent will receive a positive reward of 1 when it reaches adversary's targeted destination; otherwise, a negative reward of -1 will be given. We train agents with different network architectures and successfully injected backdoors in most of them. According to our observations, backdoor agents take longer time to learn, but final performance of the backdoor agents and the normal agents are comparable. Also, difficulty of injecting a backdoor into an agent also related to capacity of the agent's policy network.</p><p>We pick two agents as examples to make comparisons here, one without the backdoor (clean agent) and one with the backdoor (backdoor agent). Both agents have the same network architecture (2-layer LSTM) which is implemented using TensorFlow. First layer has 64 LSTM units and the second layer has 32 LSTM units. Learning environments are grids of size 17×17 with 30 holes. Agent without the backdoor only learns in the normal environment while the backdoor agent learns in both normal and poison environments. After training, we evaluate their performances under different environment configurations. We define success rate as percentage of times the agent navigates to the correct destinations over 1000 trials. For the training configuration (17×17 grid with 30 holes) without presence of the trigger, success rate of the backdoor agent is 94.8% and success rate of the clean agent is 96.3%. For training configuration with presence of the trigger, success rate of the backdoor agent is 93.4%. Median of the clean agent's performance on other clean grid configurations is 99.4%. Median of the backdoor agent's performance on other clean grid configurations is 95.0%. Median of the backdoor agent's performance on other poison grid configurations is 92.9%. Even though performance of the backdoor agent is lower than the clean agent, the difference in performance is not significant.</p><p>During experiments, we discovered that, in some grid configurations, the backdoor agent will navigate to the adversary's targeted destination even if the trigger is not presented. Our current conjecture about the cause of this unintentional backdoor activation phenomenon is related to the input and forgetting mechanism of the LSTM. Overall, there seems to be a trade-off related to sensitivity and unintentional activation of the backdoor, which needs to be appropriately optimized by the adversary.</p><p>We find that it is instructive to delve deeper into the values of hidden states and cell states of the LSTM units to understand the mechanism of how backdoor triggers affect an agent's behavior. We use the same models selected in the previous part and analyze their state responses with respect to the trigger. Environments are set to be 27×27 with 100 holes. For the same grid configuration, we let each agent run twice. In the first run, trigger is not presented and the backdoor agent will navigate to the user's targeted location. In the second run, the trigger appears at time step 12 (fixed for ablation study of cell states and hidden states), and the backdoor agent will navigate to the adversary's targeted location. We let the clean agent and the backdoor agent run in both environments for 350 times (with and without presence of the trigger), and in each trial, the locations of holes are randomly replaced. We plot all the cell states and hidden states over all the collected trajectories, and observed three types of response: (1) Impulse response: Cell states c t and hidden states h t react significantly to the trigger in a short period of time and then return back to a normal range. (2) No response: Cell states c t and hidden states h t do not react significantly to the trigger. (3) Step response: Cell states c t and hidden states h t deviate from a normal range for a long period of time. We have selected a subset of the LSTM units and their responses are plotted in Figure <ref type="figure" target="#fig_3">4</ref> and Figure <ref type="figure" target="#fig_4">5</ref>.</p><p>In the current experiments, we observe that both the clean agent and the backdoor agent has cell states and hidden states which react significantly (type 1) and mildly (type 2) to the trojan trigger; however, only the backdoor agent has some cell states and hidden states deviate from a normal range for a long period of time (type 3). We conjecture that the type 3 response keeps track of the long-term dependency of the trojan trigger. We conducted some analyses through manually changing values of some cell states c t or hidden states h t with the type 3 response when the backdoor agent is navigating. It turns out changing the values of these hidden/cell states does not affect the agent's navigation ability (avoiding holes), but it does affect the agent's final objective. In other words, we verified that altering certain hidden/cell states in LSTM network changes the goal from the user's targeted destination to the adversary's targeted destination or vice versa. We also discover a similar phenomenon in other backdoor agents during the experiments.  </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Possible Defense</head><p>Under defense mechanisms against trojan attacks, (Liu, Dolan-Gavitt, and Garg 2018) describe how these attacks can be interpreted as exploiting excess capacity in the network and explore the idea of fine tuning as well as pruning the network to reduce capacity to disable trojan attacks while retaining network performance. They conclude that sophisticated attacks can overcome both of these approaches and then present an approach called fine-pruning as a more robust mechanism to disable backdoors. <ref type="bibr" target="#b16">(Liu, Xie, and Srivastava 2017)</ref> proposes a defense method involving anomaly detection on the dataset as well as preprocessing and retraining techniques.</p><p>During our analysis on sequential DM agents, we discovered that LSTM units are likely to store long-term dependency in certain cell units. Through manually changing value of some cells, we were able to switch agent's policies between user desired policy π usr and adversary desired policy π adv and vice versa. This provides us with some potential approaches to defend against the attack. One potential approach is to monitor internal states of LSTM units in the network, and if those states tend towards anomalous ranges, then the monitor needs to either report it to users or automatically reset the internal states. This type of protection can be run online. We performed an initial study of this type of protection through visualization of hidden states and cell states values. We used a backdoor agent and recorded value of hidden states and cell states over different normal environments and poisoned environments. Mean values of the cell state vectors and hidden state vectors for normal behavior and poisoned behavior are calculated respectively. In the end, we applied a t-SNE on the mean vectors from different trials. Detailed results are shown in Figure <ref type="figure" target="#fig_5">6</ref>. From the figure, we discover that hidden state vectors and cell state vectors are quite different over normal behaviors and poisoned behaviors; thus, monitoring the internal states online and perform anomaly detection should provide some hints for the attack prevention. In this situation, the monitor will play a role similar to immune system, where if an agent is affected by the trigger, then the monitor detects and neutralizes the attack. Although we did not observe the type 3 response in clean agents in current experiments, we anticipate that some peculiar grid arrangements will require the type 3 response in clean agents too, e.g. if agent has to take a long U-turn when it gets stuck. Thus, presence of the type 3 response will not be a sufficient indicator to detect backdoor agents. An alternate static analysis approach could be to analyze the distribution of the parameters inside LSTM. Compared with the clean agents, the backdoor agents seem to use more cell units to store information. This might be reflected in the distribution of the parameters. However, more work is needed to address detection and instill resilience against such strong attacks.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Potential Challenges and Future Research</head><p>Multiple challenges exist that require further research. From the adversary's perspective, merging multiple policies into a single neural network model is hard due to catastrophic forgetting in neural networks <ref type="bibr" target="#b14">(Kirkpatrick et al. 2017</ref>). An additional challenge is the issue of unintentional backdoor activation, where some unintentional patterns (or adversarial examples) could also activate or deactivate the backdoor policy and the adversary might fail in its objective.</p><p>From the defender's perspective, it is hard to detect existence of the backdoor before a model is deployed. Neural networks by virtue of being black-box models prevent the user from fully characterizing what information is stored in a neural network. It is also difficult to track when the trigger appears in the environment (e.g. a yellow sticky note on a Stop sign from <ref type="bibr" target="#b8">(Gu, Dolan-Gavitt, and Garg 2017)</ref>). Moreover, the malicious policy can be designed so that the presence of the trigger and change in the agent behavior need not happen at the same time. Considering a backdoor model as a human body and the trigger as a virus, once the virus enters the body, there might be an incubation period before the virus affects the body and symptoms begin to appear. A similar process might apply in this type of attack. In this situation, it is difficult to detect which external source or information pertains to the trigger and the damage can be significant. Future work will also address: (1) How does one detect existence of the backdoor in an offline setting? Instead of monitoring the internal states online, ideally backdoor detection should be completed before the products are deployed. (2) How can one increase sensitivity of the trigger without introducing too many unintentional backdoor activations? One potential solution is to design the backdoor agent in a white-box setting where adversary can manipulate the network parameters.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Conclusion</head><p>We exposed a new threat type for the LSTM networks and sequential DM agents in this paper. Specifically, we showed that a maliciously-trained LSTM network-based RL agent could have reasonable performance in a normal environment, but in the presence of a trigger, the network can be made to completely switch its behavior and persist even after the trigger is removed. Some empirical evidence and intuitive understanding of the phenomena was also discussed. We also proposed some potential defense methods to counter this category of attacks and discussed avenues for future research. We hope that our work will inform the community to be aware of this type of threat and will inspire to together have better understanding in defending against and deterring these attacks.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: Left: A sample of grid world environment (Robertson 2017). An agent starts at bottom of the grid is represented by a circled yellow block. The blue blocks represent the holes. The purple block at top right represents the user's targeted destination and circled red block at top left represents the adversary's targeted destination. Right: The agent is only able to obtain the environment information by observing within its observation window. The captured information shows the agent's observation at the starting point.</figDesc><graphic coords="4,348.03,54.00,181.44,95.74" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figure 2 :</head><label>2</label><figDesc>Figure 2: Trajectories of an agent with the backdoor. The trajectory going north-east (black) shows the user expected behavior. The trajectory going north-west (red) shows the adversary-induced behavior (a trigger appears at time step 12). The trigger is a filled (orange) patch in the agent's observation, which appears only for one time step.</figDesc><graphic coords="4,348.03,439.24,181.44,120.86" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>Figure 3 :</head><label>3</label><figDesc>Figure 3: Learning curves of backdoor agents in some grid configurations. Each update step is calculated based on a batch of 128 trajectories. Left: grid size 5×5 with 0 holes. Right: grid size 7×7 with 3 holes. The score is defined as sum of performance in the normal environment and the poison environment. Shaded region represents the standard deviation over 10 trials.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_3"><head>Figure 4 :</head><label>4</label><figDesc>Figure 4: Some representative LSTM units from the backdoor agent are selected for visualization. Left: Responses of hidden state h t . Right: Responses of cell state c t . Blue curve is the backdoor agent's response in the normal environment (no trigger). Red curve is the backdoor agent's response in the poison environment (trigger presented at step 12). Shaded region represents the standard deviation, and solid line represent the mean over 350 trials.</figDesc><graphic coords="6,319.50,364.80,100.80,68.20" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_4"><head>Figure 5 :</head><label>5</label><figDesc>Figure 5: Some representative LSTM units from the clean agent are selected for visualization. Left: Responses of hidden state h t . Right: Responses of cell state c t . Blue curve is the clean agent's response in the normal environment. Red curve is the clean agent's response in the poison environment. The clean agent will be able to navigate to the user expected location even in the poison environment.</figDesc><graphic coords="6,319.50,435.89,100.80,68.20" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_5"><head>Figure 6</head><label>6</label><figDesc>Figure 6: t-SNE visualization for mean values (over time) of hidden state vectors and cell state vectors. Top left: Hidden state vector in the first layer. Top right: Hidden state vector in the second layer. Bottom left: Cell state vector in the first layer. Bottom right: Cell state vector in the second layer.</figDesc><graphic coords="7,54.00,122.82,100.80,67.02" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head></head><label></label><figDesc>Algorithm 1 -Backdoor Injection. Require: Normal Environment Env c Require: Poison Environment Env p Require: Update Batch Size b s , Training Iterations N</figDesc><table><row><cell>8:</cell><cell>Env ← Env p</cell></row><row><cell>9:</cell><cell>end if</cell></row><row><cell>10:</cell><cell></cell></row></table><note>t 1: Initialize: Policy Model π θ 2: Initialize: Performance P F c ← 0, P F p ← 0 3: Initialize: Batch Count b t ← 0 4: Initialize: Set of Trajectories Ω ← {} 5: for k ← 1 to N t do 6: Env ← Env c 7:if random(0, 1) &gt; 0.5 + (P F p − P F c ) then</note></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head></head><label></label><figDesc>1 , P F t ← Evaluate(Env c , Env p , π θ )</figDesc><table><row><cell>13:</cell><cell>// Update policy π θ when Ω ≥ b s</cell></row><row><cell>14:</cell><cell>if b t &gt; b s then</cell></row><row><cell>15:</cell><cell>// Update parameter based on past trajectories</cell></row><row><cell>16:</cell><cell>π θ ← PolicyOptimization(π θ , Ω)</cell></row><row><cell>17:</cell><cell>// Evaluate performance in two environments</cell></row><row><cell cols="2">18: P F c 19: P F c , P F t ← Normalize(P F c , P F t )</cell></row><row><cell>20:</cell><cell>Ω ← {}, b t ← 0</cell></row><row><cell>21:</cell><cell>end if</cell></row><row><cell cols="2">22: end for</cell></row><row><cell cols="2">23: return π θ</cell></row></table></figure>
		</body>
		<back>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<monogr>
		<author>
			<persName><forename type="first">E</forename><surname>Bagdasaryan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Veit</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Hua</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Estrin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Shmatikov</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1807.00459</idno>
		<title level="m">How to backdoor federated learning</title>
				<imprint>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Reinforcement learning with long shortterm memory</title>
		<author>
			<persName><forename type="first">B</forename><surname>Bakker</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Advances in neural information processing systems</title>
				<imprint>
			<date type="published" when="2002">2002</date>
			<biblScope unit="page" from="1475" to="1482" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Acting optimally in partially observable stochastic domains</title>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">R</forename><surname>Cassandra</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><forename type="middle">P</forename><surname>Kaelbling</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">L</forename><surname>Littman</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">AAAI</title>
		<imprint>
			<biblScope unit="volume">94</biblScope>
			<biblScope unit="page" from="1023" to="1028" />
			<date type="published" when="1994">1994</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<monogr>
		<title level="m" type="main">Targeted backdoor attacks on deep learning systems using data poisoning</title>
		<author>
			<persName><forename type="first">X</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Lu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Song</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1712.05526</idno>
		<imprint>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Concise deep reinforcement learning obstacle avoidance for underactuated unmanned marine vessels</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Cheng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Zhang</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Neurocomputing</title>
		<imprint>
			<biblScope unit="volume">272</biblScope>
			<biblScope unit="page" from="63" to="73" />
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<monogr>
		<title level="m" type="main">A backdoor attack against LSTM-based text classification systems</title>
		<author>
			<persName><forename type="first">J</forename><surname>Dai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Guo</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1905.12457</idno>
		<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b6">
	<monogr>
		<author>
			<persName><forename type="first">P</forename><surname>Dhariwal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Hesse</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Klimov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Nichol</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Plappert</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Radford</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Schulman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Sidor</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Wu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Zhokhov</surname></persName>
		</author>
		<ptr target="https://github.com/openai/baselines" />
		<title level="m">OpenAI baselines</title>
				<imprint>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Explaining and harnessing adversarial examples</title>
		<author>
			<persName><forename type="first">I</forename><surname>Goodfellow</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Shlens</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Szegedy</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">International Conference on Learning Representations</title>
				<imprint>
			<date type="published" when="2015">2015</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Deep recurrent Q-learning for partially observable MDPs</title>
		<author>
			<persName><forename type="first">T</forename><surname>Gu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Dolan-Gavitt</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Garg</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Hausknecht</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Stone</surname></persName>
		</author>
		<idno>CoRR, abs/1507.06527 7</idno>
	</analytic>
	<monogr>
		<title level="m">BadNets: Identifying vulnerabilities in the machine learning model supply chain</title>
				<imprint>
			<date type="published" when="2015">2017. 2015</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Long short-term memory</title>
		<author>
			<persName><forename type="first">S</forename><surname>Hochreiter</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Schmidhuber</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Neural computation</title>
		<imprint>
			<biblScope unit="volume">9</biblScope>
			<biblScope unit="issue">8</biblScope>
			<biblScope unit="page" from="1735" to="1780" />
			<date type="published" when="1997">1997</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<monogr>
		<title level="m" type="main">Adversarial attacks on neural network policies</title>
		<author>
			<persName><forename type="first">S</forename><surname>Huang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Papernot</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Goodfellow</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Duan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Abbeel</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1702.02284</idno>
		<imprint>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b11">
	<monogr>
		<title level="m" type="main">Reinforcement learning with unsupervised auxiliary tasks</title>
		<author>
			<persName><forename type="first">M</forename><surname>Jaderberg</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Mnih</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><forename type="middle">M</forename><surname>Czarnecki</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Schaul</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">Z</forename><surname>Leibo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Silver</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Kavukcuoglu</surname></persName>
		</author>
		<idno>CoRR abs/1611.05397</idno>
		<imprint>
			<date type="published" when="2016">2016</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<monogr>
		<title level="m" type="main">Internet congestion control via deep reinforcement learning</title>
		<author>
			<persName><forename type="first">N</forename><surname>Jay</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><forename type="middle">H</forename><surname>Rotman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Godfrey</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Schapira</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Tamar</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1810.03259</idno>
		<imprint>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b13">
	<monogr>
		<title level="m" type="main">TrojDRL: Trojan attacks on deep reinforcement learning agents</title>
		<author>
			<persName><forename type="first">P</forename><surname>Kiourti</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Wardega</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Jha</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Li</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1903.06638</idno>
		<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">Overcoming catastrophic forgetting in neural networks</title>
		<author>
			<persName><forename type="first">J</forename><surname>Kirkpatrick</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Pascanu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Rabinowitz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Veness</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Desjardins</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">A</forename><surname>Rusu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Milan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Quan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Ramalho</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Grabska-Barwinska</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Proceedings of the National Aacademy of Sciences</title>
		<imprint>
			<biblScope unit="volume">114</biblScope>
			<biblScope unit="issue">13</biblScope>
			<biblScope unit="page" from="3521" to="3526" />
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">Tensorforce: a TensorFlow library for applied reinforcement learning</title>
		<author>
			<persName><forename type="first">A</forename><surname>Kuhnle</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Schaarschmidt</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Fricke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Lample</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><forename type="middle">;</forename><surname>Chaplot ; Lin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Sun</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1703.06748</idno>
	</analytic>
	<monogr>
		<title level="m">Tactics of adversarial attack on deep reinforcement learning agents</title>
				<imprint>
			<date type="published" when="2016">2017. 2016. 2017</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
	<note>Playing FPS games with deep reinforcement learning</note>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">Trojaning attack on neural networks</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Ma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Aafer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W.-C</forename><surname>Lee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Zhai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Dolan-Gavitt</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Garg</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Xie</surname></persName>
		</author>
		<author>
			<persName><surname>Srivastava</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1805.12185</idno>
	</analytic>
	<monogr>
		<title level="m">Fine-pruning: Defending against backdooring attacks on deep neural networks</title>
				<imprint>
			<publisher>IEEE</publisher>
			<date type="published" when="2017">2017. 2018. 2017. 2017</date>
			<biblScope unit="page" from="45" to="48" />
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
	<note>IEEE International Conference on</note>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">Sdrl: Interpretable and data-efficient deep reinforcement learning leveraging symbolic planning</title>
		<author>
			<persName><forename type="first">D</forename><surname>Lyu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Yang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Gustafson</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the AAAI Conference on Artificial Intelligence</title>
				<meeting>the AAAI Conference on Artificial Intelligence</meeting>
		<imprint>
			<date type="published" when="2019">2019</date>
			<biblScope unit="volume">33</biblScope>
			<biblScope unit="page" from="2970" to="2977" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<analytic>
		<title level="a" type="main">Humanlevel control through deep reinforcement learning</title>
		<author>
			<persName><forename type="first">V</forename><surname>Mnih</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Kavukcuoglu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Silver</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">A</forename><surname>Rusu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Veness</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">G</forename><surname>Bellemare</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Graves</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Riedmiller</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">K</forename><surname>Fidjeland</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Ostrovski</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Nature</title>
		<imprint>
			<biblScope unit="volume">518</biblScope>
			<biblScope unit="page">529</biblScope>
			<date type="published" when="2015">2015. 7540</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<analytic>
		<title level="a" type="main">Asynchronous methods for deep reinforcement learning</title>
		<author>
			<persName><forename type="first">V</forename><surname>Mnih</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">P</forename><surname>Badia</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Mirza</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Graves</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Lillicrap</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Harley</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Silver</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Kavukcuoglu</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">International conference on machine learning</title>
				<imprint>
			<date type="published" when="2016">2016</date>
			<biblScope unit="page" from="1928" to="1937" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<monogr>
		<title level="m" type="main">Practical PyTorch: Playing gridworld with reinforcement learning</title>
		<author>
			<persName><forename type="first">S</forename><surname>Robertson</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
	<note>Web page</note>
</biblStruct>

<biblStruct xml:id="b21">
	<analytic>
		<title level="a" type="main">Trust region policy optimization</title>
		<author>
			<persName><forename type="first">J</forename><surname>Schulman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Levine</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Abbeel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Jordan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Moritz</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">International Conference on Machine Learning</title>
				<imprint>
			<date type="published" when="2015">2015</date>
			<biblScope unit="page" from="1889" to="1897" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b22">
	<monogr>
		<title level="m" type="main">Proximal policy optimization algorithms</title>
		<author>
			<persName><forename type="first">J</forename><surname>Schulman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Wolski</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Dhariwal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Radford</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Klimov</surname></persName>
		</author>
		<idno>CoRR abs/1707.06347</idno>
		<imprint>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b23">
	<analytic>
		<title level="a" type="main">A reinforcement learning approach to convoy scheduling on a contested transportation network</title>
		<author>
			<persName><forename type="first">D</forename><surname>Stimpson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Ganesan</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Optimization Letters</title>
		<imprint>
			<biblScope unit="volume">9</biblScope>
			<biblScope unit="issue">8</biblScope>
			<biblScope unit="page" from="1641" to="1657" />
			<date type="published" when="2015">2015</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b24">
	<analytic>
		<title level="a" type="main">One pixel attack for fooling deep neural networks</title>
		<author>
			<persName><forename type="first">J</forename><surname>Su</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">V</forename><surname>Vargas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Sakurai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Szegedy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Zaremba</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Sutskever ; Bruna</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Erhan</surname></persName>
		</author>
		<author>
			<persName><surname>Goodfellow</surname></persName>
		</author>
		<idno>CoRR abs/1312.6199</idno>
	</analytic>
	<monogr>
		<title level="m">Intriguing properties of neural networks</title>
				<editor>
			<persName><forename type="first">J</forename><forename type="middle">;</forename></persName>
		</editor>
		<editor>
			<persName><forename type="first">Fergus</forename></persName>
		</editor>
		<imprint>
			<date type="published" when="2013">2019. 2013</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b25">
	<analytic>
		<title level="a" type="main">Virtual-to-real deep reinforcement learning: Continuous control of mobile robots for mapless navigation</title>
		<author>
			<persName><forename type="first">L</forename><surname>Tai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Paolo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Liu</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS)</title>
				<imprint>
			<publisher>IEEE</publisher>
			<date type="published" when="2017">2017. 2017</date>
			<biblScope unit="page" from="31" to="36" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b26">
	<monogr>
		<author>
			<persName><forename type="first">Z</forename><surname>Yang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Iyer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Reimann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Virani</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1902.09972</idno>
		<title level="m">Design of intentional backdoors in sequential models</title>
				<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
