<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Critically Assessing the State of the Art in CPU-based Local Robustness Verification</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Matthias</forename><surname>König</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Leiden Institute of Advanced Computer Science</orgName>
								<orgName type="institution">Leiden University</orgName>
								<address>
									<country key="NL">The Netherlands</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Annelot</forename><forename type="middle">W</forename><surname>Bosman</surname></persName>
							<email>a.w.bosman@liacs.leidenuniv.nl</email>
							<affiliation key="aff0">
								<orgName type="department">Leiden Institute of Advanced Computer Science</orgName>
								<orgName type="institution">Leiden University</orgName>
								<address>
									<country key="NL">The Netherlands</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Holger</forename><forename type="middle">H</forename><surname>Hoos</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Leiden Institute of Advanced Computer Science</orgName>
								<orgName type="institution">Leiden University</orgName>
								<address>
									<country key="NL">The Netherlands</country>
								</address>
							</affiliation>
							<affiliation key="aff1">
								<orgName type="department">Chair for AI Methodology</orgName>
								<orgName type="institution">RWTH Aachen University</orgName>
								<address>
									<country key="DE">Germany</country>
								</address>
							</affiliation>
							<affiliation key="aff2">
								<orgName type="institution">University of British Columbia</orgName>
								<address>
									<country key="CA">Canada</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Jan</forename><forename type="middle">N</forename><surname>Van Rijn</surname></persName>
							<email>j.n.van.rijn@liacs.leidenuniv.nl</email>
							<affiliation key="aff0">
								<orgName type="department">Leiden Institute of Advanced Computer Science</orgName>
								<orgName type="institution">Leiden University</orgName>
								<address>
									<country key="NL">The Netherlands</country>
								</address>
							</affiliation>
						</author>
						<author>
							<affiliation key="aff3">
								<address>
									<addrLine>Feb 13-14</addrLine>
									<postCode>2023</postCode>
									<settlement>Washington</settlement>
									<region>D.C</region>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Critically Assessing the State of the Art in CPU-based Local Robustness Verification</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">73D32FDE9CDD8F83520E8E57F48DDB65</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-04-29T06:38+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Benchmark Analysis</term>
					<term>Neural Network Verification</term>
					<term>Adversarial Robustness</term>
					<term>Shapley Value</term>
					<term>Algorithm Portfolios</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Recent research has proposed various methods to formally verify neural networks against minimal input perturbations. This type of verification is referred to as local robustness verification. The research area of local robustness verification is highly diverse, as verifiers rely on a multitude of techniques, including mixed integer programming and satisfiability modulo theories. At the same time, problem instances differ based on the network to be verified, the verification property and the specific network input. This gives rise to the question of which verification algorithm is most suitable to solve a given verification problem. To answer this question, we perform a systematic performance analysis of several CPU-based local robustness verification systems on a newly and carefully assembled set of 79 neural networks, each verified against 100 robustness properties. Notably, we show that there is no single best algorithm that dominates in performance across all verification problem instances. Instead, our results reveal complementarities in verifier performance and illustrate the potential of leveraging algorithm portfolios for more efficient local robustness verification. Furthermore, we confirm the notion that most algorithms only support ReLU-based networks, while other activation functions remain under-supported.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>In recent years, deep learning methods based on neural networks have been increasingly applied within various safety-critical domains and use contexts, ranging from manoeuvre advisory systems in unmanned aircraft to face recognition systems in mobile phones (see, e.g., Julian et al. <ref type="bibr" target="#b0">[1]</ref>). Furthermore, it is now well known that neural networks are vulnerable to adversarial examples <ref type="bibr" target="#b1">[2]</ref>, where a given input is manipulated in such a way that it is misclassified by the network. In the case of image recognition tasks, the required perturbation, whether it is adversarially crafted or arises accidentally, can be so small that it remains virtually undetectable to the human eye. Against this background, much work has focused on developing methods to provide formal guarantees regarding the behaviour of a given neural network <ref type="bibr" target="#b2">[3,</ref><ref type="bibr" target="#b3">4,</ref><ref type="bibr" target="#b4">5,</ref><ref type="bibr" target="#b5">6,</ref><ref type="bibr" target="#b6">7,</ref><ref type="bibr" target="#b7">8,</ref><ref type="bibr" target="#b8">9,</ref><ref type="bibr" target="#b9">10,</ref><ref type="bibr" target="#b10">11,</ref><ref type="bibr" target="#b11">12,</ref><ref type="bibr" target="#b12">13,</ref><ref type="bibr" target="#b13">14]</ref>. For instance, a network employed in autonomous driving for detecting traffic signs should always produce accurate predictions, even when the input is slightly perturbed; failing to do so could have fatal consequences. This specific type of as-sessment refers to local robustness verification, a broadly studied verification task, in which a network is systematically tested against various input perturbations under predefined norms, such as the 𝑙∞ norm <ref type="bibr" target="#b14">[15,</ref><ref type="bibr" target="#b15">16]</ref>.</p><p>Neural network verification is a highly diverse research area, and existing methods rely on a broad range of techniques. At the same time, neural networks differ in terms of their architecture, such as the number of hidden layers and nodes, the type of non-linearities, e.g., ReLU, Sigmoid or Tanh, or the type of operations they employ, e.g., pooling or convolutional layers. This diversity, both in terms of verification approaches and neural network design, makes it challenging for researchers or practitioners to assess and decide which method is most suitable for verifying a given neural network <ref type="bibr" target="#b16">[17]</ref>. This challenge is amplified by the fact that the neural network verification community does not (yet) use commonly agreed evaluation protocols, which makes it difficult to draw clear conclusions from the literature regarding the capabilities and performance of existing verifiers. More precisely, existing studies use different benchmarks and, so far, have not provided an in-depth performance comparison of a broad range of verification algorithms.</p><p>Recently, a competition series has been initiated, in which several verifiers were applied to different benchmarks (i.e., networks, properties and datasets) and compared in terms of various performance measures, including the number of verified instances as well as running time <ref type="bibr" target="#b17">[18]</ref>. While the results from these competitions have provided highly valuable insights into the general progress in neural network verification, several questions are left unanswered. For instance, it remains unclear why a verification system participated in certain competition categories but not in others, although this information would help in understanding the technical capabilities and limitations of a given verifier. Furthermore, aggregation of the results makes it difficult to assess in detail the strengths or weaknesses of verifiers on different instances. Instead, looking at aggregated results, one can easily get the impression that a single approach dominates 'across the board' -an assumption that is known to be inaccurate for other problems involving formal verification tasks; see, e.g., Xu et al. <ref type="bibr" target="#b18">[19]</ref> or Kadioglu et al. <ref type="bibr" target="#b19">[20]</ref> for SAT.</p><p>In this work, we focus exclusively on local robustness verification in image classification against perturbations under the 𝑙∞ norm. This scenario represents a widely studied verification task, with a large number of networks being publicly available and many verifiers providing offthe-shelf support. Notice that most verification tasks can be translated into local robustness verification queries <ref type="bibr" target="#b20">[21]</ref>; we, therefore, believe that our findings are broadly applicable. Moreover, we seek to overcome the limitations of existing benchmarking approaches and shed light on previously unanswered questions with regard to the state of the art in local robustness verification.</p><p>Our contributions are as follows:</p><p>(i) We analyse the current state of practice in benchmarking verification algorithms;</p><p>(ii) we perform a systematic benchmarking study of several, carefully chosen verification methods based on a newly assembled, large and diverse set of networks, including 38 CIFAR and 41 MNIST networks with different activation functions, each verified against 100 robustness properties, for which we consumed a total of 1.69 CPU years in running time;</p><p>(iii) we present a categorisation of verification benchmarks based on verifier compatibilities with different layer types and operations;</p><p>(iv) we quantify verifier performance in terms of the number of solved instances, running time, as well as marginal contribution and Shapley value, showing that top-performing verification algorithms strongly complement rather than consistently dominate each other in terms of performancee.g., while the verifiers Neurify and Verinet achieved competitive performance on the CIFAR networks we considered, the former solved many instances unsolved by the latter and vice versa;</p><p>(v) lastly, we provide a public repository containing all experimental data, along with the newly assembled network collection. 1   1 https://github.com/ADA-research/nn-verification-assessment</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Background</head><p>Neural network verification methods seek to formally assess whether a trained neural network adheres to some predefined input-output property. In this study, we focus on local robustness properties. Given a trained neural network and a set of images as inputs, a robustness verification algorithm aims to verify whether or not there exist slight perturbations to an image that lead the network to predict an incorrect label. The maximum perturbation of each variable in the input, i.e., pixel in a given image, is predefined.</p><p>Formal verification algorithms can be either complete or incomplete <ref type="bibr" target="#b21">[22]</ref>. An algorithm that is incomplete does not guarantee to report a solution for every given example; however, incomplete verification algorithms are typically sound, which means they will report that a property holds only if the property actually holds. On the other hand, an algorithm that is sound and complete will correctly state that a property holds whenever it holds when given sufficient resources to be run to completion. In this work, we focus on complete algorithms, as those arguably represent the most ambitious form of neural network verification, making them preferable over incomplete methods, especially in safety-critical applications.</p><p>Early work on complete verification of neural networks utilised satisfiability modulo theories (SMT) solvers <ref type="bibr" target="#b9">[10,</ref><ref type="bibr" target="#b22">23,</ref><ref type="bibr" target="#b23">24,</ref><ref type="bibr" target="#b24">25,</ref><ref type="bibr" target="#b25">26]</ref>, which determine whether a set of logic constraints is satisfiable <ref type="bibr" target="#b26">[27]</ref>. The resulting verification problems are NP-complete and challenging to solve in practice. Alternatively, it is possible to formulate the verification task as a constraint optimisation problem using mixed integer programming (MIP) <ref type="bibr" target="#b3">[4,</ref><ref type="bibr" target="#b27">28,</ref><ref type="bibr" target="#b28">29,</ref><ref type="bibr" target="#b11">12]</ref>. MIP solvers essentially optimise an objective function subject to a set of constraints. MIP problems, similar to SMT problems, can be challenging to solve and tend to be computationally expensive (in terms of CPU time and memory).</p><p>To overcome the computational complexity of SMT and MIP, it has been proposed to use the well-known branch-and-bound algorithm <ref type="bibr" target="#b29">[30]</ref> for solving the verification problem, regardless of whether it is modelled as MIP or SMT <ref type="bibr" target="#b30">[31,</ref><ref type="bibr" target="#b4">5,</ref><ref type="bibr" target="#b31">32,</ref><ref type="bibr" target="#b6">7,</ref><ref type="bibr" target="#b12">13]</ref>.</p><p>Besides these three popular approaches, recent work has studied symbolic interval propagation <ref type="bibr" target="#b3">[4,</ref><ref type="bibr" target="#b8">9,</ref><ref type="bibr" target="#b32">33,</ref><ref type="bibr" target="#b33">34]</ref> and polyhedra <ref type="bibr" target="#b34">[35,</ref><ref type="bibr" target="#b35">36]</ref>, zonotope <ref type="bibr" target="#b7">[8,</ref><ref type="bibr" target="#b36">37]</ref> and star-set abstraction <ref type="bibr" target="#b37">[38]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Common Practices in Benchmarking Neural Network Verifiers</head><p>Considering the diversity in neural network verification problems, it is quite natural to assume that a single best algorithm does not exist, i.e., a method that outperforms all others under all circumstances. It is still hard to identify to what extent a method contributes to the state of the art, mainly because verification methods are typically evaluated (i) on a small number of benchmarks, which have often been created for the sole purpose of evaluating the method at hand, and (ii) against baseline methods for which it is often unclear how they were chosen, leading to several methods claiming state-of-the-art performance without having been directly compared. We note that in the context of local robustness verification, a benchmark most often represents a neural network classifier trained on the MNIST or CIFAR-10 dataset, respectively. As previously mentioned, a competition series has been established, with the goal of providing an objective and fair comparison of the state-of-the-art methods in neural network verification, in terms of scalability and speed <ref type="bibr" target="#b17">[18]</ref>. The VNN Competition was held in the years 2020, 2021 and 2022, yet with different protocols (e.g., for running experiments, scoring, etc.), benchmarks and participants. In this discussion, we focus on the 2021 edition, as the report from 2022 has not yet been published at the time of this writing.</p><p>Within VNN 2021, a total of 9 benchmarks were considered, of which 7 represented test cases for local robustness verification of image classification networks. Benchmarks were proposed by the participants themselves and included a total of 11 CIFAR and 7 MNIST networks, which differed in terms of architecture components, such as non-linearities (e.g., ReLU, Tanh, Sigmoid) and layer operations (e.g., convolutional or pooling layers). Networks were trained on the CIFAR-10 and MNIST datasets, respectively. Moreover, each benchmark was composed of random image subsets, excluding images that were misclassified by the given network, along with varying perturbation radii.</p><p>This competition overcame several of the previously reported limitations with regard to the evaluation of network verifiers. Most notably, it covered a relatively large and diverse set of neural networks. Moreover, thanks to the active participation from the community, 12 verification algorithms were included in the competition. At the same time, we see room for further research into the performance of neural network verifiers.</p><p>Firstly, it is not clearly stated in the VNN competition report <ref type="bibr" target="#b17">[18]</ref> why verifiers participated in certain competition categories, but not in others. While we assume this to be due to compatibility reasons, we could not find any formal explanation in the report, although this information could provide insight into relevant problem classes and suitable algorithms for solving these.</p><p>Secondly, GPU-accelerated tools were directly compared to those that rely solely on CPU resources, which we argue does not give due credit to the structural differences between these classes of algorithms. GPU-accelerated approaches employ advanced parallelisation schemes, giving rise to substantially higher computational demands than those required by methods running on CPUs, especially when restricted to a single core. For example, an Amazon EC2 GPU-based instance with 32 CPU cores costs around 1.4 times more than an equivalent CPU instance. <ref type="foot" target="#foot_0">2</ref>Lastly, the competition seeks to determine the current state of the art; however, the competition ranking and scores do not sufficiently quantify the extent to which an algorithm actually contributes to the state of the art. In other words, it is in the nature of competitions to determine a winner, at least implicitly suggesting that a single approach generally outperforms all competitors. However, some verification algorithms might have limited but distinct areas of strength, which cannot be identified through aggregated performance measures, such as the total number of verified instances. Although the competition report <ref type="bibr" target="#b17">[18]</ref> shows that individual verifier performance differs among benchmarks, it remains unclear whether all algorithms solve the same set of instances in the given benchmark, or if they complement each other. Similarly, it does not reveal whether or not methods are correlated in their performance.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Verification Algorithms under Assessment</head><p>We consider 5 complete, CPU-based neural network verification algorithms; each of these was chosen because it fulfilled one of the following conditions: it was (i) ranked among the top five verification methods according to the 2021 VNN competition or (ii) supported by the recently published DNNV framework <ref type="bibr" target="#b20">[21]</ref>. Altogether, we consider our set of algorithms to be representative of the developments in the area of complete neural network and, more specifically, local robustness verification. All methods were employed with their default hyperparameter settings, as they would likely be used by practitioners. We note that the performance of a verifier might improve if its hyperparameters were optimised specifically for the given benchmark; however, most verifiers have dozens of hyperparameters (or employ combinatorial solvers that come with their own, extensive set of hyperparameters), which makes this a non-trivial task, requiring additional expertise and resources (see, e.g., König et al. <ref type="bibr" target="#b38">[39]</ref>). The verification algorithms we considered are the following:</p><p>BaB. The algorithm proposed by Bunel et al. <ref type="bibr" target="#b4">[5]</ref> restates the verification problem as a global optimisation problem, which is then solved using branch-and-bound search. It further incorporates algorithmic improvements with regard to branching and bounding procedures such as smart branching; i.e., before splitting, it computes fast bounds on each of the possible subdomains and chooses the one with the tightest bounds. We refer to this method as BaBSB for the remainder of this paper. BaBSB supports ReLU-based networks.</p><p>Marabou. The Marabou framework <ref type="bibr" target="#b22">[23]</ref> employs SMT solving techniques, specifically the lazy search technique for handling non-linear constraints. Furthermore, Marabou employs deduction techniques to obtain information on the activation functions that can be used to simplify them. The core of the SMT solver is simplexbased, which means that the variable assignments are made using the simplex algorithm. Marabou supports ReLU and Sigmoid activations as well as MaxPooling operations.</p><p>Neurify. The verification algorithm proposed by Wang et al. <ref type="bibr" target="#b32">[33]</ref> relies on symbolic interval propagation to create over-approximations, followed by a refinement strategy based on symbolic gradient information. The constraint refinement aims to tighten the bounds of the approximation of activation functions. Neurify can process networks containing ReLU activations.</p><p>nnenum. The verifier proposed by Bak et al. <ref type="bibr" target="#b37">[38]</ref> utilises star sets to represent the values each layer of a neural network can attain. By propagating these through the network, it checks whether one or more of the star sets results in an adversarial example. This verifier can handle networks with ReLU activations.</p><p>Verinet. The verifier developed by Henriksen and Lomuscio <ref type="bibr" target="#b8">[9]</ref> combines symbolic intervals with gradientbased adversarial local search for finding counterexamples. The authors further propose a splitting heuristic for interval propagation based on the influence of a given node on the bounds of the network output. Verinet supports networks containing ReLU, Sigmoid and Tanh activations.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Setup for Empirical Evaluation</head><p>In this work, we seek to provide a clearer picture of the state of the art in neural network verification. More specifically, we argue that the state of the art is not just defined by a single verification algorithm, as there might be verifiers which on their own perform poorly but still make meaningful contributions by excelling on limited instance subsets that are challenging for other verification methods.</p><p>In the following, we will present an overview of how we set up our benchmark study, i.e., how we selected problem instances and verification algorithms. Furthermore, we will provide details on the software we used and the execution environment in which our experiments were carried out.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.1.">Problem Instances</head><p>For our assessment, we compiled a high-quality set of problem instances for local robustness verification. Following best practices in other research areas, such as optimisation <ref type="bibr" target="#b39">[40,</ref><ref type="bibr" target="#b40">41]</ref>, the benchmark should be representative and diverse, where the former refers to how well the difficulty of the benchmark is aligned with that of real-world instances from the same problem class, and the latter means that the problem set should generally contain problems with a wide range of difficulties.</p><p>Overall, our benchmark is comprised of 79 image classification networks, of which 38 are trained on the CIFAR-10 dataset and 41 are trained on the MNIST dataset. To ensure the representativeness of the problem set, all networks were sampled from the neural network verification literature, i.e., networks used in existing work on local robustness verification and provided in public repositories; in other words, the characteristics of the networks in our benchmark are assumed to match those of networks generally used for evaluating verification algorithms.</p><p>We further want our problem set to be diverse. Therefore, we ensured that the considered networks differ in size, i.e., the number of hidden layers and nodes, as well as the type of non-linearities (e.g., ReLU or Tanh) or layer operations (e.g., pooling layers) they employ.</p><p>For each network, we verified 100 local robustness properties; more precisely, we sampled 100 images from the dataset on which the network has been trained and verified for local robustness with perturbation radius 𝜖 set at 0.012. This perturbation radius was chosen to be a median of values we found in literature: the smallest value we found was 1/255 in the work by Li et al. <ref type="bibr" target="#b21">[22]</ref>, whereas Botoeva et al. <ref type="bibr" target="#b3">[4]</ref> or Wang et al. <ref type="bibr" target="#b12">[13]</ref> utilised larger values, such as 0.05.</p><p>Lastly, we split our benchmark set into different categories based on verifier compatibilities. This means a verifier is only employed to categories it is able to process. The categories as well as the instance set size for each category are shown in Table <ref type="table">1</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.2.">Evaluation Metrics</head><p>In order to assess the performance of the various methods, we compute four performance metrics: the average running time, the number of solved instances, the relative marginal contribution and the Shapley value <ref type="bibr" target="#b41">[42]</ref> of each verifier. Although these metrics present aggregated measures, they reflect algorithm performance on an instance level and in relation to other methods included in the comparison; a more detailed explanation will be provided in the following paragraphs. Notice that we do not penalise timeouts when computing average running time; i.e., the maximum running time equals the given time limit.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Table 1</head><p>Instance set size for each benchmark category. Solvable instances are those solved by at least one (i.e., any) or all of the considered verifiers. The column "Verifiers employed" lists the matching suitable verification algorithm(s) to the respective category. The marginal contribution is computed as follows. Define 𝑐 as a set of given verifiers and let 𝑣(𝑐) be the total score of set 𝑐. Here, the total score 𝑣(𝑐) consists of the number of instances verified by at least one verifier in set 𝑐. We compute the marginal contribution per algorithm to determine how much the total performance of all algorithms (in terms of solved instances) decreases when the given algorithm is removed from the set of all algorithms if they were employed in a parallel algorithm portfolio. Such portfolios are sets of algorithms that are run in parallel on each given problem instance and complement each other in terms of performance across an instance set <ref type="bibr" target="#b42">[43]</ref>. Formally, to determine the marginal contribution of any of the verifiers 𝑖 to portfolio 𝑐, one needs to know the value of 𝑣(𝑐) and 𝑣(𝑐 ∖ {𝑖}), where 𝑐∖{𝑖} is the portfolio minus verifier 𝑖. Thus, the marginal contribution of verifier 𝑖 is expressed as</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>MNIST CIFAR</head><note type="other">Category</note><formula xml:id="formula_0">MCi = 𝑣(𝑐) − 𝑣(𝑐 ∖ {𝑖})<label>(1)</label></formula><p>Following this terminology, we can define the number of solved instances by verifier 𝑖 as a set consisting only of verifier 𝑖, Solvedi = 𝑣(𝑖) − 𝑣(∅). In other words, the number of solved instances employs a set of size one and the marginal contribution employs a set of all verifiers under consideration. The relative marginal contribution then represents the marginal contribution of a given verifier as a proportion of the sum of every method's absolute marginal contribution. Lastly, the Shapley value is the average marginal contribution of a verifier over all possible joining orders, where joining order refers to the order in which the verifiers are added to a parallel portfolio. This value is complementing the previous two metrics, as it does not assume a particular order in which algorithms are added to the portfolio. To be precise, the number of solved instances simply represents a joining order in which the considered algorithm comes first, whereas the marginal contribution metric assumes a joining order in which it comes last. However, using fixed orders, as it is the case for the marginal contribution, might not reveal possible interactions between the given method and other algorithms, e.g., it might understate the importance of a single algorithm given the presence of another, highly correlated algorithm. In such a case, both algorithms would not be assigned any marginal contribution, even though one of them should be included in a potential portfolio. This is captured by the Shapley value: define 𝑛 as the number of verifiers under consideration and 𝐶 as the set of all combinations of all subsets of verifiers under consideration including the empty set, where set 𝐶 will be of size</p><formula xml:id="formula_1">∑︀ 𝑛 𝑘=0 𝑛! 𝑘! • (𝑛 − 𝑘)!</formula><p>. Finally, define 𝐶𝑖 as all sets in which verifier 𝑖 is present. The Shapley value of verifier 𝑖 is then calculated as</p><formula xml:id="formula_2">𝜑𝑖 = 1 𝑛! • ∑︁ 𝑗∈𝐶 𝑖 (𝑣(𝑗) − 𝑣(𝑗 ∖ {𝑖}))<label>(2)</label></formula></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.3.">Execution Environment and Software Used</head><p>Our experiments were carried out on a cluster of machines equipped with Intel Xeon E5-2683 CPUs with 32 cores, 40 MB cache size and 94 GB RAM, running CentOS Linux 7. For each verification method, we limited the number of available CPU cores to a single core. Each query was given a time budget of 3 600 seconds and a memory budget of 3 GB. Generally, we executed the verification algorithms through the DNNV interface, version 0.4.8. DNNV is a framework that transforms a network and property specification into a unified format, which can then be solved by a given method <ref type="bibr" target="#b20">[21]</ref>. More specifically, DNNV takes as input a network in the ONNX format, along with a property specification, and then translates the network and property to the input format required by the verifier. After running the verifier on the transformed problem, it returns either 𝑠𝑎𝑡 if the property was falsified or 𝑢𝑛𝑠𝑎𝑡 if the property was proven to hold. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.">Results and Discussion</head><p>In the following, we provide an in-depth discussion of the results obtained from our experiments. Table <ref type="table">1</ref> shows the categories we devised, along with the resulting instance set sizes as well as information on which verifier has been employed for each category. Table <ref type="table" target="#tab_1">2</ref> reports the number of problem instances solved by each verifier per network category, the relative marginal contribution, the Shapley value and the average running time. On ReLU-based MNIST networks, we found Verinet to be the best-performing verifier, solving 1 799 out of 2 500 instances, while achieving a Shapley value of 618. However, taking relative marginal contribution into account, we found that Neurify achieved the highest relative marginal contribution of 0.25 (compared to 0.16 for Verinet), indicating that it could verify a sizable fraction of instances on which other methods failed to return a solution. Moreover, the relative marginal contribution scores show that each method could solve a sizeable fraction of instances unsolved by any of the other methods.</p><p>On ReLU-based CIFAR networks, it should first be noted that no verification problem instance can be solved by all verifiers, highlighting the structural differences between instances and the sensitivity of the verification approaches to those instances. That said, Neurify slightly outperformed Verinet in terms of the number of solved instances (915 vs 841 out of 2 500). Furthermore, Neurify achieved a much larger relative marginal contribution than Verinet (0.75 vs 0.20), which means that the former could solve a relatively large number of instances which could not be solved by the other methods. Generally, relative marginal contribution scores are much less evenly distributed among verifiers when compared to the MNIST dataset.</p><p>Figures <ref type="figure" target="#fig_0">1a and 1b</ref> show an instance-level comparison of the two best-performing algorithms (in terms of Shapley value) in the ReLU category for each dataset. In Figure <ref type="figure" target="#fig_0">1a</ref>, we see that on MNIST networks, both Verinet and nnenum solved instances that the other one could not solve within the given time budget. Concretely, when considering a parallel portfolio containing both algorithms, the number of solved instances slightly increases to 1 817 out of 2 500 (vs 1 799 solved by Verinet and 1 754 solved by nnenum alone), while supplied with similar CPU resources (i.e., 1 800 CPU seconds per verifier, which represents half of the total given time budget of 3 600  CPU seconds per verification query as used in our experiments). Notice that leveraging parallel portfolios has already been shown to significantly improve the performance of MIP-based verification methods <ref type="bibr" target="#b38">[39]</ref>.</p><formula xml:id="formula_3">CPU time [s], nnenum</formula><p>On CIFAR instances, we found Neurify and Verinet to also have distinctive strengths over each other. This is shown in Figure <ref type="figure" target="#fig_0">1b</ref>, where both algorithms could solve a substantial amount of instances that the other could not return a solution for. Thus, when combined in a parallel portfolio, 963 instances can be solved (vs 915 solved by Neurify and 841 solved by Verinet alone, out of 2 500 instances), while using the same amount of CPU resources, i.e., 1 800 CPU seconds per verifier. These findings further emphasise the complementarity between the verification algorithms considered in our study. All remaining verifiers achieved substantially lower Shapley values and relative marginal contribution scores, indicating that they would not complement Neurify and Verinet well in a portfolio.</p><p>Figure <ref type="figure" target="#fig_1">2a</ref> shows the cumulative distribution function of running times over the MNIST problem instances. As seen in the figure, Verinet tends to solve these problem instances fastest; however, Neurify tended to show even better performances on those instances it was able to solve. We note that most of the instances unsolved by Neurify represent networks that were trained on images with 3 dimensions, whereas Neurify requires images used as network inputs to have 2 or 4 dimensions.</p><p>Figure <ref type="figure" target="#fig_1">2b</ref> shows a similar plot for the CIFAR problem instances. Here, Neurify solved the largest fraction in less time than other methods. This suggests that Neurify is a very competitive verifier when applicable to the specific network or input format.</p><p>For each of the remaining categories, we found that there is only one verifier that could effectively handle the respective problem instances. Specifically, instances from the ReLU+MaxPooling category can be processed by Marabou, although, only a modest number of MNIST instances could be solved in this way. Networks containing Tanh activations can, in principle, be verified by Verinet, but the algorithm did also not solve any CIFAR instances. Lastly, Sigmoid-based networks can be handled by both Verinet and Marabou, however, only the former could solve MNIST instances within the given compute budget.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="7.">Conclusions and Future Work</head><p>In this work, we assessed the performance of several local robustness verification algorithms, i.e., algorithms used to verify the robustness of an image classification network against small input perturbations. To conclude, we found that all considered methods support ReLU-based networks, while other network types are strongly undersupported. While this has been suspected in the community, it has, to our knowledge, not yet been subject to formal study. Furthermore, we presented evidence for strong performance complementarity: even within the same benchmark category, two verification systems outperform each other on distinct sets of instances. As we have demonstrated, this complementarity can be exploited by combining individual verifiers into parallel portfolios. Lastly, we showed that, in general, the performance of verifiers strongly differs between image datasets, with some methods achieving the best performance on MNIST (in terms of the number of solved instances and average running time) while falling behind on CIFAR and vice versa. In future work, it would be interesting to include a broader set of perturbation radii and analyse in more detail how the relative performance of verifiers depends on the given radius. Furthermore, we are interested in expanding our analysis to GPU-based verification algorithms.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: Performance comparison of the two top-performing verification methods (in terms of Shapley value) in the ReLU category on (a) MNIST and (b) CIFAR networks. The plots show that each verifier outperforms the other on some instances, while none of the methods is dominating in performance across the entire instance set. This illustrates the complementary strengths of the verification algorithms. The diagonal line indicates equal performance of the two methods.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figure 2 :</head><label>2</label><figDesc>Figure 2: Fraction of instances solved by the considered verification algorithms in the ReLU category as a function of CPU running time for (a) MNIST and (b) CIFAR networks. On (a), we find that Verinet solves most instances in the least time, while on (b) Neurify does.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 2</head><label>2</label><figDesc>Performance comparison of local robustness verification algorithms in terms of the number of solved instances, relative marginal contribution (RMC), Shapley value (𝜑) and average CPU running time, computed for each category with 𝜖 set at 0.012.</figDesc><table><row><cell>ReLU</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell>Verifier</cell><cell></cell><cell cols="2">MNIST</cell><cell></cell><cell></cell><cell cols="2">CIFAR</cell><cell></cell></row><row><cell></cell><cell cols="2">Solved RMC</cell><cell cols="4">𝜑 Avg. Time Solved RMC</cell><cell cols="2">𝜑 Avg. Time</cell></row><row><cell></cell><cell></cell><cell></cell><cell></cell><cell>[CPU s]</cell><cell></cell><cell></cell><cell></cell><cell>[CPU s]</cell></row><row><cell>BaBSB</cell><cell>358</cell><cell cols="2">0.22 118</cell><cell>3 241</cell><cell>307</cell><cell>0.00</cell><cell>86</cell><cell>2 924</cell></row><row><cell>Marabou</cell><cell>1 001</cell><cell cols="2">0.19 312</cell><cell>1 801</cell><cell>400</cell><cell cols="2">0.00 117</cell><cell>2 153</cell></row><row><cell>Neurify</cell><cell>871</cell><cell cols="2">0.25 265</cell><cell>1 964</cell><cell>915</cell><cell cols="2">0.75 411</cell><cell>235</cell></row><row><cell>nnenum</cell><cell>1 754</cell><cell cols="2">0.17 600</cell><cell>389</cell><cell>76</cell><cell>0.05</cell><cell>28</cell><cell>3 337</cell></row><row><cell>Verinet</cell><cell>1 799</cell><cell cols="2">0.16 618</cell><cell>263</cell><cell>841</cell><cell cols="2">0.20 330</cell><cell>500</cell></row><row><cell>ReLU+Maxpool</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell>Verifier</cell><cell></cell><cell cols="2">MNIST</cell><cell></cell><cell></cell><cell cols="2">CIFAR</cell><cell></cell></row><row><cell></cell><cell cols="2">Solved RMC</cell><cell cols="4">𝜑 Avg. Time Solved RMC</cell><cell cols="2">𝜑 Avg. Time</cell></row><row><cell>Marabou</cell><cell>5</cell><cell>1.00</cell><cell>5</cell><cell>57</cell><cell>0</cell><cell>0.00</cell><cell>0</cell><cell>3 600</cell></row><row><cell>Tanh</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell>Verifier</cell><cell></cell><cell cols="2">MNIST</cell><cell></cell><cell></cell><cell cols="2">CIFAR</cell><cell></cell></row><row><cell></cell><cell cols="2">Solved RMC</cell><cell cols="4">𝜑 Avg. Time Solved RMC</cell><cell cols="2">𝜑 Avg. Time</cell></row><row><cell>Verinet</cell><cell>556</cell><cell cols="2">1.00 556</cell><cell>55</cell><cell>0</cell><cell>0.00</cell><cell>0</cell><cell>3 600</cell></row><row><cell>Sigmoid</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell>Verifier</cell><cell></cell><cell cols="2">MNIST</cell><cell></cell><cell></cell><cell cols="2">CIFAR</cell><cell></cell></row><row><cell></cell><cell cols="2">Solved RMC</cell><cell cols="4">𝜑 Avg. Time Solved RMC</cell><cell cols="2">𝜑 Avg. Time</cell></row><row><cell>Marabou</cell><cell>0</cell><cell>0.00</cell><cell>0</cell><cell>3 600</cell><cell>0</cell><cell>0.00</cell><cell>0</cell><cell>3 600</cell></row><row><cell>Verinet</cell><cell>581</cell><cell cols="2">1.00 581</cell><cell>55</cell><cell>0</cell><cell>0.00</cell><cell>0</cell><cell>3 600</cell></row></table></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_0">See https://aws.amazon.com/ec2/pricing/on-demand/</note>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgements</head><p>This research was partially supported by TAILOR, a project funded by EU Horizon 2020 research and innovation program under GA No. 952215.</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Deep neural network compression for aircraft collision avoidance systems</title>
		<author>
			<persName><forename type="first">K</forename><forename type="middle">D</forename><surname>Julian</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">J</forename><surname>Kochenderfer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">P</forename><surname>Owen</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Guidance, Control, and Dynamics</title>
		<imprint>
			<biblScope unit="volume">42</biblScope>
			<biblScope unit="page" from="598" to="608" />
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<monogr>
		<author>
			<persName><forename type="first">C</forename><surname>Szegedy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Zaremba</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Sutskever</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Bruna</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Erhan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Goodfellow</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Fergus</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1312.6199</idno>
		<title level="m">Intriguing properties of neural networks</title>
				<imprint>
			<date type="published" when="2014">2014</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Measuring Neural Net Robustness with Constraints</title>
		<author>
			<persName><forename type="first">O</forename><surname>Bastani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Ioannou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Lampropoulos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Vytiniotis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Nori</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Criminisi</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Advances in Neural Information Processing Systems</title>
				<meeting><address><addrLine>NeurIPS</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2016">2016. 2016</date>
			<biblScope unit="page" from="2613" to="2621" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Efficient Verification of ReLU-based Neural Networks via Dependency Analysis</title>
		<author>
			<persName><forename type="first">E</forename><surname>Botoeva</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Kouvaros</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Kronqvist</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Lomuscio</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Misener</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 34th AAAI Conference on Artificial Intelligence (AAAI-20)</title>
				<meeting>the 34th AAAI Conference on Artificial Intelligence (AAAI-20)</meeting>
		<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="3291" to="3299" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">A Unified View of Piecewise Linear Neural Network Verification</title>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">R</forename><surname>Bunel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Turkaslan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Torr</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Kohli</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><forename type="middle">K</forename><surname>Mudigonda</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Advances in Neural Information Processing Systems</title>
				<meeting><address><addrLine>NeurIPS</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2018">2018. 2018</date>
			<biblScope unit="page" from="1" to="10" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">A Dual Approach to Scalable Verification of Deep Networks</title>
		<author>
			<persName><forename type="first">K</forename><surname>Dvijotham</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Stanforth</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Gowal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">A</forename><surname>Mann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Kohli</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 38th Conference on Uncertainty in Artificial Intelligence (UAI 2018)</title>
				<meeting>the 38th Conference on Uncertainty in Artificial Intelligence (UAI 2018)</meeting>
		<imprint>
			<date type="published" when="2018">2018</date>
			<biblScope unit="page" from="550" to="559" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Formal Verification of Piece-Wise Linear Feed-Forward Neural Networks</title>
		<author>
			<persName><forename type="first">R</forename><surname>Ehlers</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 15th International Symposium on Automated Technology for Verification and Analysis (ATVA 2017)</title>
				<meeting>the 15th International Symposium on Automated Technology for Verification and Analysis (ATVA 2017)</meeting>
		<imprint>
			<date type="published" when="2017">2017</date>
			<biblScope unit="page" from="269" to="286" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">AI2: Safety and Robustness Certification of Neural Networks with Abstract Interpretation</title>
		<author>
			<persName><forename type="first">T</forename><surname>Gehr</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Mirman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Drachsler-Cohen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Tsankov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Chaudhuri</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Vechev</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 39th IEEE Symposium on Security and Privacy (IEEE S&amp;P 2018)</title>
				<meeting>the 39th IEEE Symposium on Security and Privacy (IEEE S&amp;P 2018)</meeting>
		<imprint>
			<date type="published" when="2018">2018</date>
			<biblScope unit="page" from="3" to="18" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Efficient Neural Network Verification via Adaptive Refinement and Adversarial Search</title>
		<author>
			<persName><forename type="first">P</forename><surname>Henriksen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Lomuscio</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 24th European Conference on Artificial Intelligence (ECAI 2020)</title>
				<meeting>the 24th European Conference on Artificial Intelligence (ECAI 2020)</meeting>
		<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="2513" to="2520" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Reluplex: An Efficient SMT Solver for Verifying Deep Neural Networks</title>
		<author>
			<persName><forename type="first">G</forename><surname>Katz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Barrett</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">L</forename><surname>Dill</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Julian</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">J</forename><surname>Kochenderfer</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 29th International Conference on Computer Aided Verification</title>
				<meeting>the 29th International Conference on Computer Aided Verification<address><addrLine>CAV</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2017">2017. 2017</date>
			<biblScope unit="page" from="97" to="117" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Towards Verification of Artificial Neural Networks</title>
		<author>
			<persName><forename type="first">K</forename><surname>Scheibler</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Winterer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Wimmer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Becker</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 18th Workshop on Methoden und Beschreibungssprachen zur Modellierung und Verifikation von Schaltungen und Systemen</title>
				<meeting>the 18th Workshop on Methoden und Beschreibungssprachen zur Modellierung und Verifikation von Schaltungen und Systemen<address><addrLine>MBMV</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2015">2015. 2015</date>
			<biblScope unit="page" from="30" to="40" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Evaluating Robustness of Neural Networks with Mixed Integer Programming</title>
		<author>
			<persName><forename type="first">V</forename><surname>Tjeng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Xiao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Tedrake</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 7th International Conference on Learning Representations (ICLR</title>
				<meeting>the 7th International Conference on Learning Representations (ICLR</meeting>
		<imprint>
			<date type="published" when="2019">2019. 2019</date>
			<biblScope unit="page" from="1" to="21" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Efficient Formal Safety Analysis of Neural Networks</title>
		<author>
			<persName><forename type="first">S</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Pei</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Whitehouse</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Yang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Jana</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Advances in Neural Information Processing Systems</title>
				<meeting><address><addrLine>NeurIPS</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2018">2018. 2018</date>
			<biblScope unit="page" from="6369" to="6379" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">Output Reachable Set Estimation and Verification for Multilayer Neural Networks</title>
		<author>
			<persName><forename type="first">W</forename><surname>Xiang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H.-D</forename><surname>Tran</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">T</forename><surname>Johnson</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Transactions on Neural Networks and Learning Systems</title>
		<imprint>
			<biblScope unit="volume">29</biblScope>
			<biblScope unit="page" from="5777" to="5783" />
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<monogr>
		<author>
			<persName><forename type="first">I</forename><forename type="middle">J</forename><surname>Goodfellow</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Shlens</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Szegedy</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1412.6572</idno>
		<title level="m">Explaining and Harnessing Adversarial Examples</title>
				<imprint>
			<date type="published" when="2014">2014</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">Distillation as a Defense to Adversarial Perturbations Against Deep Neural Networks</title>
		<author>
			<persName><forename type="first">N</forename><surname>Papernot</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Mcdaniel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Wu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Jha</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Swami</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 37th IEEE Symposium on Security and Privacy (IEEE S&amp;P</title>
				<meeting>the 37th IEEE Symposium on Security and Privacy (IEEE S&amp;P</meeting>
		<imprint>
			<date type="published" when="2016">2016. 2016</date>
			<biblScope unit="page" from="582" to="597" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">Neural Network Robustness as a Verification Property: A Principled Case Study</title>
		<author>
			<persName><forename type="first">M</forename><surname>Casadio</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Komendantskaya</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">L</forename><surname>Daggitt</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Kokke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Katz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Amir</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Refaeli</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 34rd International Conference on Computer Aided Verification (CAV 2022)</title>
				<meeting>the 34rd International Conference on Computer Aided Verification (CAV 2022)</meeting>
		<imprint>
			<date type="published" when="2022">2022</date>
			<biblScope unit="page" from="219" to="231" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<monogr>
		<author>
			<persName><forename type="first">S</forename><surname>Bak</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Johnson</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2109.00498</idno>
		<title level="m">The Second International Verification of Neural Networks Competition (VNN-COMP 2021): Summary and Results</title>
				<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b18">
	<analytic>
		<title level="a" type="main">Satzilla: Portfolio-based algorithm selection for sat</title>
		<author>
			<persName><forename type="first">L</forename><surname>Xu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Hutter</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><forename type="middle">H</forename><surname>Hoos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Leyton-Brown</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Artificial Intelligence Research</title>
		<imprint>
			<biblScope unit="volume">32</biblScope>
			<biblScope unit="page" from="565" to="606" />
			<date type="published" when="2008">2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<analytic>
		<title level="a" type="main">Algorithm Selection and Scheduling</title>
		<author>
			<persName><forename type="first">S</forename><surname>Kadioglu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Malitsky</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Sabharwal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Samulowitz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Sellmann</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Seventeenth International Conference on Principles and Practice of Constraint Programming</title>
				<meeting>the Seventeenth International Conference on Principles and Practice of Constraint Programming</meeting>
		<imprint>
			<date type="published" when="2011">CP2011. 2011</date>
			<biblScope unit="page" from="454" to="469" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<analytic>
		<title level="a" type="main">DNNV: A Framework for Deep Neural Network Verification</title>
		<author>
			<persName><forename type="first">D</forename><surname>Shriver</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Elbaum</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">B</forename><surname>Dwyer</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 33rd International Conference on Computer Aided Verification</title>
				<meeting>the 33rd International Conference on Computer Aided Verification<address><addrLine>CAV</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2021">2021. 2021</date>
			<biblScope unit="page" from="137" to="150" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b21">
	<monogr>
		<author>
			<persName><forename type="first">L</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Qi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Xie</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Li</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2009.04131</idno>
		<title level="m">Sok: Certified robustness for deep neural networks</title>
				<imprint>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b22">
	<analytic>
		<title level="a" type="main">the Marabou Framework for Verification and Analysis of Deep Neural Networks</title>
		<author>
			<persName><forename type="first">G</forename><surname>Katz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">A</forename><surname>Huang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Ibeling</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Julian</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Lazarus</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Lim</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Shah</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Thakoor</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Wu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Zeljić</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">L</forename><surname>Dill</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">J</forename><surname>Kochenderfer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Barrett</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 31st International Conference on Computer Aided Verification</title>
				<meeting>the 31st International Conference on Computer Aided Verification<address><addrLine>CAV</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2019">2019. 2019</date>
			<biblScope unit="page" from="443" to="452" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b23">
	<analytic>
		<title level="a" type="main">Checking Safety of Neural Networks with SMT Solvers: A Comparative Evaluation</title>
		<author>
			<persName><forename type="first">L</forename><surname>Pulina</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Tacchella</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">AI*IA</title>
		<imprint>
			<biblScope unit="page" from="127" to="138" />
			<date type="published" when="2011">2011</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b24">
	<analytic>
		<title level="a" type="main">NeVer: A Tool for Artificial Neural Networks Verification</title>
		<author>
			<persName><forename type="first">L</forename><surname>Pulina</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Tacchella</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Annals of Mathematics and Artificial Intelligence</title>
		<imprint>
			<biblScope unit="page" from="403" to="425" />
			<date type="published" when="2011">2011</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b25">
	<analytic>
		<title level="a" type="main">Challenging SMT Solvers to Verify Neural Networks</title>
		<author>
			<persName><forename type="first">L</forename><surname>Pulina</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Tacchella</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">AI Communications</title>
		<imprint>
			<biblScope unit="page" from="117" to="135" />
			<date type="published" when="2012">2012</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b26">
	<analytic>
		<title level="a" type="main">Satisfiability Modulo theories: An Appetizer</title>
		<author>
			<persName><forename type="first">L</forename><forename type="middle">D</forename><surname>Moura</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Bjørner</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Brazilian Symposium on Formal Methods (SBMF</title>
				<meeting>the Brazilian Symposium on Formal Methods (SBMF</meeting>
		<imprint>
			<date type="published" when="2009">2009. 2009</date>
			<biblScope unit="page" from="23" to="36" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b27">
	<analytic>
		<title level="a" type="main">Output Range Analysis for Deep Neural Networks</title>
		<author>
			<persName><forename type="first">S</forename><surname>Dutta</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Jha</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Sankaranarayanan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Tiwari</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Tenth NASA Formal Methods Symposium</title>
				<meeting>the Tenth NASA Formal Methods Symposium<address><addrLine>NFM</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2018">2018. 2018</date>
			<biblScope unit="page" from="121" to="138" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b28">
	<monogr>
		<author>
			<persName><forename type="first">A</forename><surname>Lomuscio</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Maganti</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1706.07351</idno>
		<title level="m">An approach to reachability analysis for feed-forward ReLU neural networks</title>
				<imprint>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b29">
	<analytic>
		<title level="a" type="main">An Automatic Method of Solving Discrete Programming Problems</title>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">H</forename><surname>Land</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">G</forename><surname>Doig</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Econometrica</title>
		<imprint>
			<biblScope unit="page" from="497" to="520" />
			<date type="published" when="1960">1960</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b30">
	<analytic>
		<title level="a" type="main">Branch and Bound for Piecewise Linear Neural Network Verification</title>
		<author>
			<persName><forename type="first">R</forename><surname>Bunel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Turkaslan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><forename type="middle">H S</forename><surname>Torr</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">P</forename><surname>Kumar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Lu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Kohli</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Machine Learning Research</title>
		<imprint>
			<biblScope unit="page" from="1574" to="1612" />
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b31">
	<monogr>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">De</forename><surname>Palma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Bunel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Desmaison</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Dvijotham</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Kohli</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><forename type="middle">H S</forename><surname>Torr</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">P</forename><surname>Kumar</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2104.06718</idno>
		<title level="m">Improved Branch and Bound for Neural Network Verification via Lagrangian Decomposition</title>
				<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b32">
	<analytic>
		<title level="a" type="main">Formal Security Analysis of Neural Networks using Symbolic Intervals</title>
		<author>
			<persName><forename type="first">S</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Pei</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Whitehouse</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Yang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Jana</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 27th USENIX Security Symposium (USENIX Security 18)</title>
				<meeting>the 27th USENIX Security Symposium (USENIX Security 18)</meeting>
		<imprint>
			<date type="published" when="2018">2018</date>
			<biblScope unit="page" from="1599" to="1614" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b33">
	<analytic>
		<title level="a" type="main">Beta-CROWN: Efficient Bound Propagation with Per-neuron Split Constraints for Neural Network Robustness Verification</title>
		<author>
			<persName><forename type="first">S</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Xu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Lin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Jana</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C.-J</forename><surname>Hsieh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Kolter</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Advances in Neural Information Processing Systems</title>
				<meeting><address><addrLine>NeurIPS</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2021">2021. 2021</date>
			<biblScope unit="page" from="29909" to="29921" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b34">
	<analytic>
		<title level="a" type="main">An Abstract Domain for Certifying Neural Networks</title>
		<author>
			<persName><forename type="first">G</forename><surname>Singh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Gehr</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Püschel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Vechev</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 46th ACM SIGPLAN Symposium on Principles of Programming Languages (ACM-POPL</title>
				<meeting>the 46th ACM SIGPLAN Symposium on Principles of Programming Languages (ACM-POPL</meeting>
		<imprint>
			<date type="published" when="2019">2019. 2019</date>
			<biblScope unit="page" from="1" to="30" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b35">
	<analytic>
		<title level="a" type="main">Efficient Neural Network Robustness Certification with General Activation Functions</title>
		<author>
			<persName><forename type="first">H</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T.-W</forename><surname>Weng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P.-Y</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C.-J</forename><surname>Hsieh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Daniel</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Advances in Neural Information Processing Systems</title>
		<imprint>
			<biblScope unit="volume">31</biblScope>
			<biblScope unit="page" from="4944" to="4953" />
			<date type="published" when="2018">2018. 2018</date>
		</imprint>
	</monogr>
	<note>NeurIPS</note>
</biblStruct>

<biblStruct xml:id="b36">
	<analytic>
		<title level="a" type="main">Fast and Effective Robustness Certification</title>
		<author>
			<persName><forename type="first">G</forename><surname>Singh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Gehr</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Mirman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Püschel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Vechev</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Advances in Neural Information Processing Systems</title>
				<meeting><address><addrLine>NeurIPS</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2018">2018. 2018</date>
			<biblScope unit="page" from="1" to="12" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b37">
	<analytic>
		<title level="a" type="main">Improved Geometric Path Enumeration for Verifying ReLU Neural Networks</title>
		<author>
			<persName><forename type="first">S</forename><surname>Bak</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H.-D</forename><surname>Tran</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Hobbs</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">T</forename><surname>Johnson</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 32nd International Conference on Computer Aided Verification</title>
				<meeting>the 32nd International Conference on Computer Aided Verification<address><addrLine>CAV</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2020">2020. 2020</date>
			<biblScope unit="page" from="66" to="96" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b38">
	<analytic>
		<title level="a" type="main">Speeding up neural network robustness verification via algorithm configuration and an optimised mixed integer linear programming solver portfolio</title>
		<author>
			<persName><forename type="first">M</forename><surname>König</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><forename type="middle">H</forename><surname>Hoos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">N V</forename><surname>Rijn</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Machine Learning</title>
		<imprint>
			<biblScope unit="volume">111</biblScope>
			<biblScope unit="page" from="4565" to="4584" />
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b39">
	<monogr>
		<author>
			<persName><forename type="first">H</forename><forename type="middle">H</forename><surname>Hoos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Stützle</surname></persName>
		</author>
		<title level="m">Stochastic Local Search: Foundations &amp; Applications</title>
				<imprint>
			<publisher>Elsevier / Morgan Kaufmann</publisher>
			<date type="published" when="2004">2004</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b40">
	<monogr>
		<author>
			<persName><forename type="first">T</forename><surname>Bartz-Beielstein</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Doerr</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Berg</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Bossek</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Chandrasekaran</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Eftimov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Fischbach</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Kerschke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>La Cava</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Lopez-Ibanez</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2007.03488</idno>
		<title level="m">Benchmarking in Optimization: Best Practice and Open Issues</title>
				<imprint>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b41">
	<analytic>
		<title level="a" type="main">Using the shapley value to analyze algorithm portfolios</title>
		<author>
			<persName><forename type="first">A</forename><surname>Fréchette</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Kotthoff</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Michalak</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Rahwan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Hoos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Leyton-Brown</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 30th AAAI Conference on Artificial Intelligence (AAAI-16)</title>
				<meeting>the 30th AAAI Conference on Artificial Intelligence (AAAI-16)</meeting>
		<imprint>
			<date type="published" when="2016">2016</date>
			<biblScope unit="page" from="3397" to="3403" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b42">
	<analytic>
		<title level="a" type="main">Algorithm portfolios</title>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">P</forename><surname>Gomes</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Selman</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Artificial Intelligence</title>
		<imprint>
			<biblScope unit="volume">126</biblScope>
			<biblScope unit="page" from="43" to="62" />
			<date type="published" when="2001">2001</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
