<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Defending the unknown: Exploring reinforcement learning agents&apos; deployment in realistic, unseen networks</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Alberto</forename><surname>Acuto</surname></persName>
							<email>a.acuto@liverpool.ac.uk</email>
							<affiliation key="aff0">
								<orgName type="department">School of Electrical Engineering, Electronics and Computer Science</orgName>
								<orgName type="institution">The University of Liverpool</orgName>
								<address>
									<addrLine>Brownlow Hill</addrLine>
									<postCode>L69 3GJ</postCode>
									<settlement>Liverpool</settlement>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Simon</forename><surname>Maskell</surname></persName>
							<email>s.maskell@liverpool.ac.uk</email>
							<affiliation key="aff0">
								<orgName type="department">School of Electrical Engineering, Electronics and Computer Science</orgName>
								<orgName type="institution">The University of Liverpool</orgName>
								<address>
									<addrLine>Brownlow Hill</addrLine>
									<postCode>L69 3GJ</postCode>
									<settlement>Liverpool</settlement>
								</address>
							</affiliation>
							<affiliation key="aff1">
								<orgName type="institution">The Alan Turing Institute</orgName>
							</affiliation>
						</author>
						<title level="a" type="main">Defending the unknown: Exploring reinforcement learning agents&apos; deployment in realistic, unseen networks</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">448062D7AED5C1983A162E31284A52FF</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T19:58+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Reinforcement Learning</term>
					<term>Cyber security</term>
					<term>Network simulation</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>The increasing number of network simulators has opened opportunities to explore and apply state-ofthe-art algorithms to understand and measure the capabilities of such techniques in numerous sectors. In this regard, the recently released Yawning Titan is one example of a simplistic, but not less detailed, representation of a cyber network scenario where it is possible to train agents guided by reinforcement learning algorithms and measure their effectiveness in trying to stop an infection. In this paper, we explore how different reinforcement learning algorithms lead the training of various agents in different examples and realistic networks. We assess how we can deploy such agents in a set of networks, focusing in particular on the resilience of the agents in exploring networks with complex starting states, increased number of routes connecting the nodes and different levels of challenge, aiming to evaluate the deployment performances in realistic networks never seen before.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>The development of autonomous resilient agents in the context of automated cyber defence (ACD) to counteract the actions of external or malevolent actors is becoming a pivotal research topic from both academy and governmental agencies. In recent years cyber crimes have increased their presence in the day-to-day life of organisations and governmental institutions, and research in automated cyber defence is one of the most developed topics <ref type="bibr" target="#b0">[1]</ref>. Novel technologies such as machine learning (ML) and reinforcement learning (RL) are increasingly employed for both defence and attack thanks to their adaptability, cyber resilience and variety of applications. Some defensive examples are ML applications in spam detection <ref type="bibr" target="#b1">[2]</ref>, malware and intrusion detection <ref type="bibr" target="#b2">[3,</ref><ref type="bibr" target="#b3">4]</ref>, offensive applications can relate to the deployment of algorithms to exploit vulnerabilities of infrastructures and limit the visibility and extend duration (or frequency) of</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.1.">Yawning Titan</head><p>Yawning Titan is a graph-based cyber-security simulation environment that allows the training of intelligent agents to counteract the actions of a red enemy agent that aims to spread into the network. The YT setup specifies the red and blue agents' capabilities (e.g. the usable actions and success rate) and the network's description: connections between the nodes, entry nodes and the presence and location of a "high-value target" (HVT). Each agent has a set of parameters describing the probability of success of the actions and game rules (i.e., how the red agent can spread from a compromised node or if the blue agent can detect failed attacks in the network). In detail, the red agent has a set of actions, one of which is randomly picked, to "attack" a node, move in the network or "do nothing". The blue agent, instead, has a wider set of possible actions which he can perform:</p><p>• Isolate: removes all edges of the node, cost 10; • Restore: returns the node to its original status (from "compromised" to "safe"), cost 1. The agent can be punished if patches a safe node or if there are too many infected nodes;</p><p>• Make node safe: reduces the vulnerability<ref type="foot" target="#foot_2">3</ref> of a node, cost 0.5. The agent can be punished in the same way as for Restore action; • Connect: reinstate all edges of the node, cost 0, the agent is rewarded by 5 points if it reduces the number of isolated nodes; • Add deceptive node: add an extra "fake" node between two nodes to slow the spread of the red agent <ref type="foot" target="#foot_3">4</ref> , cost 8. The agent is also punished by 5 points if adds more deceptive nodes than allowed (in our case 3); • Do nothing: self explanatory, cost -0.5. The agent is punished by doing this action if there are a lot of infected nodes.</p><p>The blue action space can be modified using a configuration file. The score is obtained from a a combination of the action costs plus the rewards obtained from removing red nodes, the penalties from the actions and final points from ending the game (winning or losing, ±100 points). The score is parameter the agents need to optimise. The network is the "gym" where the two agents interact and can either be loaded from an existing scenario or prompted by the user. There are multiple ways to describe a network: it is fundamental to map the specific connections across the nodes <ref type="foot" target="#foot_4">5</ref> and with that knowledge, it is possible to generate an adjacency matrix that is read and interpreted by a graph-based Python library (Networkx<ref type="foot" target="#foot_5">6</ref> ) which can interact using Pandas. A network is defined by nodes connected with edges, some of whose are defined as "entry nodes" which are the starting points where the red agent will begin its spread. It is possible to add a "high-value target", a valuable asset inside the network, the red agent can target that specific node and, under certain game rules, can be the trigger for the endgame. In the network, the software generates each node and assigns a specific vulnerability, entry nodes tend to have a higher vulnerability score because are infection starting point. At the end of each simulation step, the vulnerabilities are evaluated in terms of the red and blue actions. At the beginning of each simulation, we have a safe network where every node is clean and the game parameters are set (entry and HVT nodes).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Algorithms</head><p>We have considered algorithms from the library Stable-baseline3 which represent modelfree RL algorithms where an "agent" learns to play by interacting with the environment. A trained agent has the knowledge of the states by performing actions, obtaining a reward (positive or negative) and their effect on the environment. The aim of the agent is to learn the best actions from a policy in order to maximise the total rewards across an episode, which is everything that happens between the first and the last state in the environment (considered like a timestep). We have considered online, model-free RL algorithms <ref type="foot" target="#foot_6">7</ref> with well-documented applications across different simulations <ref type="bibr" target="#b15">[16]</ref> and that have a discrete action space. YT environment describes the possible actions on a discrete space, which scales by the number of nodes and usable actions on each node. Other RL algorithm needs box-shaped actions space or images (e.g. in the cases of CNNPOLICIES) which is not straightforward to implement with YT. The policies considered are MLPPOLICIES, where we pass the state vectors of the network in our input model. These policies implement an actor-critic neural network using a multilayer perceptron (with 2 layers of 64 neurons). The algorithms we are considering are Proximal Policy Optimisation (or PPO <ref type="bibr" target="#b16">[17]</ref>), Advantage Actor Critic (or A2C <ref type="bibr" target="#b17">[18]</ref>) and Deep Q-Network (or DQN <ref type="bibr" target="#b18">[19,</ref><ref type="bibr" target="#b19">20]</ref>). PPO algorithm works using a policy gradient optimisation based on natural policy gradients. This algorithm is known to perform better in comparison to similar ones, because the training is more stable by avoiding broad policy updates, helping the convergence on an optimal solution and allowing enough time to recover from an action <ref type="bibr" target="#b20">[21]</ref>. The algorithm is based on the optimisation of the policy objective function using a gradient descent (or ascent) and uses a "clipped" surrogate of the objective function which prevents too large policy updates.</p><p>A2C is an Actor-Critic method based on temporal difference learning<ref type="foot" target="#foot_7">8</ref> that represent the policy function independent of the value function. Our implementation is Advantage Actor Critic and comes from the Asynchronous Advantage Actor-Critic (A3C) without the asynchronous part. This algorithm, as PPO , uses the policy gradient to weigh the actions and reduce the variance by using a large number of samples (created by single agent exploring the action space) hoping that one of these will provide the true estimation.</p><p>DQN is a deep reinforcement learning algorithm which uses Q-learning to learn the best action to take in the given state and a deep neural network (or convolutional neural network) is implemented to estimate the value of the Q-function.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1.">Algorithms hyper-parameters exploration</head><p>We consider exploring how the different algorithms react by changing some hyper-parameters such as the discount factor (𝛾) and the learning rate (lr). 𝛾 measures the rewards the agent has achieved in the past, present and future. An agent with 𝛾 = 0, only cares about his first reward (myopic approach), while if 𝛾 = 1, it is interested in all the future rewards. lr is a parameter that measures how often and, how quickly, the Q-values are updated, improving the steps toward the solution, smaller lr can slow the gradient descent while, a larger value, can fail to converge.</p><p>In DQN algorithm, we have reduced the buffer size to 10000 (from 10 6 ), because in the largest network (&gt;50 nodes) there was the risk of requiring more memory than allocated on the HPC<ref type="foot" target="#foot_8">9</ref> cores. By doing so, we both achieved a faster convergence and assured us to not overload the computing nodes.</p><p>There are studies (i.e., <ref type="bibr" target="#b21">[22]</ref>) that demonstrate and compare various RL agents in different contexts by changing and tuning the various hyper-parameters. For the means of this paper, we have not fine-tuned our agents to perform in the networks because we wanted to have control over the way the performances may differ. We have chosen the best models to deploy in realistic networks according to the testing on sample networks.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Simulation setup</head><p>We have trained agents on a set of networks comprised of a small case of 18 nodes (25 edges), a medium one of 50 nodes (&gt;250 edges) and the largest case of 100 nodes (&gt;500 edges) with an increasing number of entry points <ref type="bibr" target="#b2">(3,</ref><ref type="bibr" target="#b4">5,</ref><ref type="bibr" target="#b9">10)</ref>. A trained agent on an 18 nodes network cannot be deployed onto a larger or smaller network because the dimension of the action space is bound to the possible states in the specific case.</p><p>The simulation setup has a red agent that can spread only via connected nodes with 45% infection success rate and 15% chance of spreading from connected infected nodes, the endgame rule is that the red agent wins if it takes over 80% of the network, 500 timesteps are the target for a blue victory. The HVT is chosen randomly and furthest away from the entry points. We train and analyse the performances of the various algorithms on a set of example networks, using the same network for both training and evaluation, then, we test the best algorithms on a series of realistic network configurations after training on similar networks with the same amount of nodes.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1.">Training the agents</head><p>We perform the training of the various agents in the networks without any hyper-parameter tuning. Then, we train the agents by modifying just one parameter at a time, 𝛾 and lr on the same networks. We set up 5 × 10 5 timesteps and we consider the convergence when we have not measured any improvements of the average rewards for up to four consecutive evaluations (a single evaluation is the average of rewards over 50 timesteps).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Table 1</head><p>We present the algorithms and their hyper-parameters modified in the training phase, the first uses the standard hyper-parameter values. We divide the training on the three network sizes (left, central and right column) showing the training time (in seconds) and the final score obtained in each case. <ref type="bibr" target="#b17">18</ref>  In table 1, we compare the training of the agents showing the training time (in seconds) and the final scores. All agents converge to an optimal solution before the end of the training, we find, also, that adding more nodes (larger action space) the time requirements are higher. The final score can give us an indication of the expected results during the trials, however, we should not be surprised by any different results. If we have to compare the training times: PPO seems to be the quickest to converge in all three network sizes. The A2C algorithm reaches the stability quite quickly, even if it is not the better performing at the end of the training session, the PPO agents tend to have steady and stable growth in performances during the training while DQN agents have a constant behaviour: in all training, they have very little gain during the initial steps, then they rapidly improve their performances taking over the A2C performances as well.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.">Deploy the trained agents</head><p>In this section, after having presented the training, we compare the performances of the agents on the same seeded networks and we obtain the mean and standard deviation of the scores achieved (using Stable-baseline3 evaluator function EVALUATE_POLICY). We compare the performances with the scores obtained by a random agent on the same networks, this agent randomly chooses a node and one action from the available. By testing the agents' performances on the same seeded network we aim to evaluate the agents on a constrained set of examples of starting points, red agent actions, reducing the variability of the games and aiming to understand better how they behave.</p><p>In figure <ref type="figure" target="#fig_1">1</ref>, we compare the agents' performances against the same networks. In diamonds we present the standard hyper-parameters setup, crosses for the algorithms with 𝛾 = 0.75 and in triangles the case with lr = 0.001. We show the PPO results using blue dots, orange for the A2C case and green for DQN algorithm. We can see that in the case of 18 nodes, all algorithms have similar and comparable performances (around -130 as mean reward). In the 50 nodes case, we measure a more significant difference in performances, in particular in the DQN case, almost 5 times lower than PPO results. Instead, in the 100 nodes case we measure a considerable lower score in the PPO case (in which the mean reward passes from a few hundred to -2000) while, on the opposite side, DQN shows higher scores. As stated earlier, we changed the buffer size of DQN , by doing so we measure a significant improvement in the performances highlighting a strong positive influence of this hyper-parameter. Both A2C and DQN algorithms, with a reduced discount factor, perform better than PPO in the largest network case. By increasing the network size, the agents have more opportunities to take action, and more chances of having negative rewards because the red agent is able to spread more. Therefore, the simulations are longer, and many of them resulted in the blue agent victory since the agent was able to slow the spread by making more expensive actions such as adding deceptive nodes. In cases of large networks, it is important to see also how spread are the final scores, measured by the standard deviation, and understand how an algorithm can overall perform.</p><p>We want to extend this analysis while changing the starting conditions, in detail: add isolated nodes, compromised nodes, a mixture of isolated and compromised nodes, changing the number of edges and the red agent's skills. In figure <ref type="figure">2</ref>   We show the algorithms with standard hyper-parameters using diamonds, in crosses the case modifying 𝛾 and in triangles when modifying the lr. We use different colors to easily identify the algorithms: blue for PPO , orange for A2C and green for DQN . We show as well the 1𝜎 deviations of the scores with colored lines. The grey band is the random agent scores.</p><p>were randomly chosen, therefore it may have happened that some nodes were both isolated and compromised at the same time. Changing the number of edges in the network creates, or removes, routes for the red agent to spread, but also allows the blue agent to defend better the network by adding more deceptive nodes or isolating cross-road nodes, reducing the effectiveness of the spread. Changing the red agent skills increases, or reduces, the simulation challenge level because a red agent that has a higher success rate spreads more quickly and it is more difficult to react to, on the other hand, a less effective red agent would leave more time for the agent to fix the nodes.</p><p>We can see that, in most cases, tweaking the network (nodes isolated or compromised) does not result in a measurable change in the final scores. For instance, the PPO algorithm obtains a mean of 130 points in all three variations when we lower 𝛾. Considering the DQN algorithm, we find a similar picture in which even if the scores differ from the standard case, the variations trials have comparable results between them. Considering the explorations with fewer or more edges, we measure a noteworthy difference in the mean rewards between the PPO and the other algorithms. PPO agent, even with the variations, shows a lower score removing edges, while this is not the case for the other two agents, however the random agent performances are significantly lower. On the other hand, adding edges increases the variations for the random agent but the three algorithms behave similarly. A similar conclusion can be found in the instances of red agent skills: a less effective red agent (low skill) makes the game longer resulting in scores built on top of more expensive actions and fewer rewards from fixing the network. A more effective red agent is more aggressive and spreads quickly and the blue agent gains more points in fixing the nodes, which results in shorter game lengths because it is more difficult to stop and prevent the final escalation. Indeed, this is also verified by the small spread of the results from the random agent.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>s t a n d a r d c o m p r o m i s e i s o l a t e m i xe d g e s + e d g e s l o w -s k i l l h i g h -s k i</head><p>In the case of a highly skilled red agent the performances of the random agent and the trained ones are really similar, even if still distinguishable.</p><p>In the 50 node scenario, bottom panel of figure <ref type="figure">2</ref>, we see a significant difference between the various algorithms, PPO agent is always better performing compared to the others, being DQN agent the lowest among the three. Modifying the network result in similar spread of the scores even varying the parameters as 𝛾 and lr.</p><p>In 100 nodes case, figure <ref type="figure">3</ref>, we see that A2C with lr modified has similar scores with PPO agents results in almost all experiments, while DQN obtains significantly different results, in particular by modifying 𝛾. Interestingly, we do not see much difference when we modify the number of edges with all models scoring similar results. A2C agents are the worst performing, with results significantly lower than the other agents. Summarising the findings and analysis done in these cases we can say:</p><p>• for given algorithm (and hyper-parameter choice) adding isolated or compromised nodes, does not change significantly the performances in comparison to a clean starting network; • the scores in larger networks are lower because the game length is generally higher, this happens because the red agent is not able to overtake the network quickly, therefore the blue agent has more time to do expensive actions, while on smaller networks it is more rewarding fixing the nodes even if that results in loss;</p><p>• changing the network topology can trigger significant changes in the agents response.</p><p>Adding more edges results in higher scores. However we measure similar performances between the agents and variations of hyper-parameters; • the red agent skills has a significant impact on the performances but more importantly has the same reaction on the blue agent, we understand by noting small differences between all scores (small standard deviations).</p><p>Given this panoramic view of these results we can say that both PPO and A2C are performing well in small networks, in particular by changing the discount factor (𝛾), while adjusting the buffer size and 𝛾 in the DQN case conveys in performing well, and in a stable manner, in larger network cases.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.3.">Agent deployment on realistic networks</head><p>In the previous sections, we have explored the performances of RL algorithms trained and tested onto the same networks, without any resemblance to reality, trying to understand the best algorithm and check their resilience in the changes. In this subsection, we focus on a sample of cases using realistic networks configurations for testing the agents' behaviour. We consider three cases with 22, 55 and 60 nodes. We have extrapolated these networks from a portion of a larger existing network of computers considering only nodes connected, i.e. nodes that are connected between them but they do not share a connection with the core of the network are not considered. We deploy in the first instance an agent trained using an A2C algorithm, PPO with the standard setup for the 55 node network and DQN with 𝛾 = 0.75 in the latter. We train the agents on example networks with the same amount of nodes but with a different configuration (number of edges), and we evaluate the changes in performances in the realistic ones, which are effectively novel networks to the agents. In table 2, we summarise the network statistics for the synthetic networks used in training, the realistic environments in deployment and the algorithms selected. The average clustering measures how many connections are between the nodes: nodes connected with more edges have a value closer to 1. The triangles are a set of three nodes where each node has a relationship to the other two. These quantities can describe the complexity of the network, in particular it is clear that the realistic networks are much simpler (lower average clustering and number of triangles) in comparison to the synthetic ones.</p><p>In figure <ref type="figure" target="#fig_3">4</ref>, we compare the performances of the trained agents in five different scenarios on the network used for training and the realistic one, we show the random agent performances for comparison. As before, we test the agents while modifying the network by adding isolated nodes, compromised, a mixture of those and against a red agent with lower and higher skills level. In this figure the y-axis is in log-scale for easier comparison. The goal of this analysis is to understand how an agent trained on a different network performs in a realistic network without re-training.</p><p>In the case of A2C , we can see that the scores in the training network are almost identical and, almost, two times higher in comparison to the ones obtained in the realistic scenario exploration. It is interesting to note that the scatter from the five realisations is small in the training networks,</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Table 2</head><p>We present the synthetic networks used for training and the realistic networks used in deployment. The network statistics are the number of nodes, edges, high-value targets (HVT) and entry nodes. The algorithms used are: A2C and PPO with standard hyper-parameters and DQN with 𝛾 = 0.75 and buffer size = 10000. Finally, we present the average clustering value and the number of triangles present in the network: the average clustering measures how much the nodes are closer and connected with edges (nodes more connected have this measure closer to one), and the triangles defined as three nodes where each one has a relationship to the other two. while in the realistic ones there is a larger variance of results. This result maybe be connected to the peculiar shape of the realistic network and its relatively low number of routes the red agent can choose. The random agent performances are significantly lower in comparison to other agents.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Mode</head><p>In the network with 55 nodes we use the PPO algorithm: we measure values close to -4500 (close to the random agent performances) in the training network for most of the different scenarios, while in the real network the scores are around -1500, even in the case of varying the red agent's skill. We can conclude that the difference in the topology of the network has played a significant role in this analysis, overcoming as well, the impact of varying the red agents' skill, which has shown a larger effect in the previous analysis.</p><p>We have deployed the DQN algorithm in the final case considered. We find a similar behaviour we have seen in the 22 nodes case, with scores in the training network really close (around -300) one to the other and higher in comparison to the realistic scenario (close to -1000). In the realistic network we measure a larger variability in the final scores, again, we measure that the largest impact on the scores is due to the red agents' skills. The random agent scores are significantly lower in all tests.</p><p>In light of these results, we can state that the YT framework allows the training of RL agents and their deployment is transferable from synthetic to realistic networks with minimal loss of performance. In particular, we can validate that the agents' scores are lower in comparison to the ones obtained from agents trained and evaluated on the same networks, however, these are still much more improved from random scores. This analysis highlights that changing the networks' topology has not invalidated the performances of such agents.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Conclusion and future work</head><p>In this paper, we have performed training and evaluation of RL agents in a set of networks using Yawning Titan ACO capabilities, comparing how their performances change by modifying the In the top panel we present the case with 22 nodes using A2C algorithm, in the central panel the 55 nodes network using PPO and finally, in the bottom panel we present the DQN agent applied to a 60 node network. The green crosses are the random agent scores, blue diamonds are the training scores and the realistic cases are in orange crosses. Please note that the y-axis, differently from the previous figures, is in log-scale for easier comparison.</p><p>status of the network and methods hyper-parameters looking for the best algorithm to deploy in realistic networks. The main findings are that by increasing the number of nodes, the mean reward per simulation is lower, highlighting a positive correlation with the action space dimension and varying the red agent's skills has the larger impact on the results. We did not measure significant differences in the scores while modifying the status of the nodes (being compromised or isolated), on the other hand adding (or removing) edges in the network and augmenting (or reducing) the red agent skills showed interesting differences in the performances. We find that the RL algorithms considered can react well to network changes by measuring the level of performances across the various tests. By exploring the hyper-parameters tuning, the discount factor (𝛾) seems to have the most positive impact in the training and evaluation processes, in comparison to the limited results obtained by changing the learning rate (lr).</p><p>This work has shown and proved the possibility of using Yawning Titan in training agents that could be considered in realistic cyber-defence environments with minimal computational requirements. The tests we have carried out have shown that the changes in the agents' performances were arising from the different network topologies and not from changes in the network status itself. We have shown, as well, that it is possible to deploy an agent trained on a different topology with minimal loss of performance, and in some cases (e.g., 50 nodes networks) we have measured an improvement in the mean scores. These results show the possibility to train intelligent agents in synthetic networks and deploy such agents in realistic networks without re-training. However, little exploration has been done in modifying or exploiting the current rewards of actions inside the simulations, exploring different winning setups (e.g., allowing the end game if the high-value target is taken) and more complex scenarios (e.g., more red agents, complex decision making) and other Markov decision process algorithms.</p><p>Extension of the current work can be exploring algorithms with proper hyper-parameter tuning, exploitation of the current reward scheme, offline learning methods and inclusion of multi-agent algorithms and time evolving networks.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head></head><label></label><figDesc>and 3, we summarise the mean reward scores obtained by the various agents in the networks applying the proposed changes on the network. Isolated and compromised nodesPPO A2C DQN PPO A2C DQN PPO A2C</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1:Agents' performance comparison in the different networks (left panel 18 nodes, centre 50 nodes and 100 nodes right panel). We show the algorithms with standard hyper-parameters using diamonds, in crosses the case modifying 𝛾 and in triangles when modifying the lr. We use different colors to easily identify the algorithms: blue for PPO , orange for A2C and green for DQN . We show as well the 1𝜎 deviations of the scores with colored lines. The grey band is the random agent scores.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>Figure 2 :Figure 3 :</head><label>23</label><figDesc>Figure 2:Mean rewards for the agents on the three networks (18 nodes top, 50 nodes bottom and 100 nodes in figure3), in blue symbols we show the PPO results, A2C in orange and DQN in green. The three symbols show the different changes on the algorithm: the diamond is the standard version of the algorithm, the cross is using 𝛾 = 0.75 and the triangle is with lr = 0.001. The random agent (RND) is shown in grey with the mean value as dashed line and the grey area shows 1𝜎 deviation. The 1𝜎 deviation on the agents scores is shown using y-errorbars. On the x-axis we show the various extension tested as adding compromised nodes and adding or removing edges.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_3"><head>Figure 4 :</head><label>4</label><figDesc>Figure 4: Agents' performance comparison between training networks and realistic networks.In the top panel we present the case with 22 nodes using A2C algorithm, in the central panel the 55 nodes network using PPO and finally, in the bottom panel we present the DQN agent applied to a 60 node network. The green crosses are the random agent scores, blue diamonds are the training scores and the realistic cases are in orange crosses. Please note that the y-axis, differently from the previous figures, is in log-scale for easier comparison.</figDesc></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_0">Autonomous Resilient Cyber Defence, https://www.gov.uk/government/news/autonomous-resilient-cyber-defenceintelligent-agents.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_1">2 https://github.com/dstl/YAWNING-TITAN .</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_2">The vulnerability score of a node is a metric that is used for evaluating the risk of the node to be attacked. Exposed nodes and nodes neighbouring, connected, to a compromised node have a higher vulnerability, meaning a higher probability of being infected.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="4" xml:id="foot_3">Adding a deceptive node does not count as adding a node in the network.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="5" xml:id="foot_4">True, for version 1.0.1. In more recent versions, e.g. 2.0.1b, the user can also draw the network.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="6" xml:id="foot_5">https://networkx.org/documentation/stable/index.html.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="7" xml:id="foot_6">In the case of offline methods we should create a dataset from the simulation and then train the agents on such data, without the live interactions between the agent and the environment.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="8" xml:id="foot_7">Temporal difference learning methods are a class of model-free reinforcement learning algorithms which learn by bootstrapping the current estimate of the value function.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="9" xml:id="foot_8">High performance computing. The training of the agents were performed on HPC CPU cores at University of Liverpool computing facility.</note>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgments</head><p>The authors thank the reviewers for their useful comments that improved the quality of the paper and thank the useful discussions with Neil Dhir. This project was financially supported by a contract with the Alan Turing Institute.</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">A Qualitative Research on the Impact and Challenges of Cybercrimes</title>
		<author>
			<persName><forename type="first">V</forename><forename type="middle">Krishna</forename><surname>Viraja</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Purandare</surname></persName>
		</author>
		<idno type="DOI">10.1088/1742-6596/1964/4/042004</idno>
	</analytic>
	<monogr>
		<title level="j">Journal of Physics Conference Series</title>
		<imprint>
			<biblScope unit="page">42004</biblScope>
			<date type="published" when="1964">1964. 2021</date>
		</imprint>
	</monogr>
	<note>Journal of Physics Conference Series</note>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Survey of review spam detection using machine learning techniques</title>
		<author>
			<persName><forename type="first">M</forename><surname>Crawford</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">M</forename><surname>Khoshgoftaar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">D</forename><surname>Prusa</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">N</forename><surname>Richter</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><forename type="middle">A</forename><surname>Najada</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Big Data</title>
		<imprint>
			<biblScope unit="volume">2</biblScope>
			<date type="published" when="2015">2015</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Iot security techniques based on machine learning: How do iot devices use ai to enhance security?</title>
		<author>
			<persName><forename type="first">L</forename><surname>Xiao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Wan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Lu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Wu</surname></persName>
		</author>
		<idno type="DOI">10.1109/MSP.2018.2825478</idno>
	</analytic>
	<monogr>
		<title level="j">IEEE Signal Processing Magazine</title>
		<imprint>
			<biblScope unit="volume">35</biblScope>
			<biblScope unit="page" from="41" to="49" />
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">A survey of data mining and machine learning methods for cyber security intrusion detection</title>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">L</forename><surname>Buczak</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Guven</surname></persName>
		</author>
		<idno type="DOI">10.1109/COMST.2015.2494502</idno>
	</analytic>
	<monogr>
		<title level="j">IEEE Communications Surveys &amp; Tutorials</title>
		<imprint>
			<biblScope unit="volume">18</biblScope>
			<biblScope unit="page" from="1153" to="1176" />
			<date type="published" when="2016">2016</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<monogr>
		<author>
			<persName><forename type="first">Y</forename><surname>Huang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Huang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><surname>Zhu</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2107.00783</idno>
		<title level="m">Reinforcement learning for feedback-enabled cyber resilience</title>
				<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<monogr>
		<author>
			<persName><forename type="first">S</forename><surname>Vyas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Hannay</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Bolton</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><forename type="middle">P</forename><surname>Burnap</surname></persName>
		</author>
		<idno type="DOI">10.48550/arXiv.2303.04926</idno>
		<idno type="arXiv">arXiv:2303.04926</idno>
		<idno>arXiv:2303.04926</idno>
		<title level="m">Automated Cyber Defence: A Review</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Research and challenges of reinforcement learning in cyber defense decision-making for intranet security</title>
		<author>
			<persName><forename type="first">W</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Sun</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Jiang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Zhu</surname></persName>
		</author>
		<idno type="DOI">10.3390/a15040134</idno>
		<ptr target="https://www.mdpi.com/1999-4893/15/4/134.doi:10.3390/a15040134" />
	</analytic>
	<monogr>
		<title level="j">Algorithms</title>
		<imprint>
			<biblScope unit="volume">15</biblScope>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<monogr>
		<title level="m" type="main">Robust artificial intelligence for active cyber defence</title>
		<author>
			<persName><forename type="first">A</forename><surname>Burke</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2017">2017</date>
			<publisher>Alan Turing Insitute</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">A review of recent advances in machine learning approaches for cyber defense</title>
		<author>
			<persName><forename type="first">R</forename><surname>Buettner</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Sauter</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Klopfer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Breitenbach</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Baumgartl</surname></persName>
		</author>
		<idno type="DOI">10.1109/BigData52589.2021.9671918</idno>
	</analytic>
	<monogr>
		<title level="m">2021 IEEE International Conference on Big Data (Big Data)</title>
				<imprint>
			<date type="published" when="2021">2021</date>
			<biblScope unit="page" from="3969" to="3974" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<monogr>
		<author>
			<persName><forename type="first">M</forename><surname>Standen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Lucas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Bowman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">J</forename><surname>Richer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Kim</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Marriott</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2108.09118</idno>
		<title level="m">Cyborg: A gym for the development of autonomous cyber agents</title>
				<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b10">
	<monogr>
		<idno>arXiv</idno>
		<title level="m">CAGE Challenge</title>
				<imprint>
			<date type="published" when="2021">2021</date>
			<biblScope unit="volume">1</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title/>
	</analytic>
	<monogr>
		<title level="j">TTCP CAGE Challenge</title>
		<imprint>
			<biblScope unit="volume">2</biblScope>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<monogr>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">C W</forename><surname>Group</surname></persName>
		</author>
		<ptr target="https://github.com/cage-challenge/cage-challenge-3" />
		<title level="m">Ttcp cage challenge 3</title>
				<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<monogr>
		<author>
			<persName><forename type="first">A</forename><surname>Molina-Markham</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Miniter</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Powell</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Ridley</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2103.07583</idno>
		<title level="m">Network environment design for autonomous cyberdefense</title>
				<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">Developing optimal causal cyber-defence agents via cyber security simulation</title>
		<author>
			<persName><forename type="first">A</forename><surname>Andrew</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Spillard</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Collyer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Dhir</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Workshop on Machine Learning for Cybersecurity (ML4Cyber)</title>
				<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">Stable-baselines3: Reliable reinforcement learning implementations</title>
		<author>
			<persName><forename type="first">A</forename><surname>Raffin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Hill</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Gleave</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Kanervisto</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Ernestus</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Dormann</surname></persName>
		</author>
		<ptr target="http://jmlr.org/papers/v22/20-1364.html" />
	</analytic>
	<monogr>
		<title level="j">Journal of Machine Learning Research</title>
		<imprint>
			<biblScope unit="volume">22</biblScope>
			<biblScope unit="page" from="1" to="8" />
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<monogr>
		<author>
			<persName><forename type="first">J</forename><surname>Schulman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Wolski</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Dhariwal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Radford</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Klimov</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1707.06347</idno>
		<title level="m">Proximal policy optimization algorithms</title>
				<imprint>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">Asynchronous methods for deep reinforcement learning</title>
		<author>
			<persName><forename type="first">V</forename><surname>Mnih</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">P</forename><surname>Badia</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Mirza</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Graves</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Lillicrap</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Harley</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Silver</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Kavukcuoglu</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">International conference on machine learning</title>
				<meeting><address><addrLine>PMLR</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2016">2016</date>
			<biblScope unit="page" from="1928" to="1937" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<monogr>
		<author>
			<persName><forename type="first">V</forename><surname>Mnih</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Kavukcuoglu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Silver</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Graves</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Antonoglou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Wierstra</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Riedmiller</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1312.5602</idno>
		<title level="m">Playing atari with deep reinforcement learning</title>
				<imprint>
			<date type="published" when="2013">2013</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b19">
	<analytic>
		<title level="a" type="main">Human-level control through deep reinforcement learning</title>
		<author>
			<persName><forename type="first">V</forename><surname>Mnih</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Kavukcuoglu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Silver</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">A</forename><surname>Rusu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Veness</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">G</forename><surname>Bellemare</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Graves</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Riedmiller</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">K</forename><surname>Fidjeland</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Ostrovski</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">nature</title>
		<imprint>
			<biblScope unit="volume">518</biblScope>
			<biblScope unit="page" from="529" to="533" />
			<date type="published" when="2015">2015</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<monogr>
		<author>
			<persName><forename type="first">B</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><surname>Cai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Yang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Wang</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1906.10306</idno>
		<title level="m">Neural proximal/trust region policy optimization attains globally optimal policy</title>
				<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b21">
	<monogr>
		<author>
			<persName><forename type="first">M</forename><surname>Andrychowicz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Raichuk</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Stańczyk</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Orsini</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Girgin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Marinier</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Hussenot</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Geist</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Pietquin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Michalski</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Gelly</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Bachem</surname></persName>
		</author>
		<idno type="DOI">10.48550/arXiv.2006.05990</idno>
		<idno type="arXiv">arXiv:2006.05990</idno>
		<idno>arXiv:2006.05990</idno>
		<title level="m">What Matters In On-Policy Reinforcement Learning? A Large-Scale Empirical Study</title>
				<imprint>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
