<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Calibrated Multi-Probabilistic Prediction as a Defense against Adversarial Attacks</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Jonathan</forename><surname>Peck</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Department of Applied Mathematics, Computer Science and Statistics</orgName>
								<orgName type="institution">Ghent University</orgName>
								<address>
									<postCode>9000</postCode>
									<settlement>Ghent</settlement>
									<country key="BE">Belgium</country>
								</address>
							</affiliation>
							<affiliation key="aff1">
								<orgName type="department">Data Mining and Modeling for Biomedicine</orgName>
								<orgName type="institution">VIB Inflammation Research Center</orgName>
								<address>
									<postCode>9052</postCode>
									<settlement>Ghent</settlement>
									<country key="BE">Belgium</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Bart</forename><surname>Goossens</surname></persName>
							<affiliation key="aff2">
								<orgName type="department">Department of Telecommunications and Information Processing</orgName>
								<orgName type="institution">IMEC/Ghent University</orgName>
								<address>
									<postCode>9000</postCode>
									<settlement>Ghent</settlement>
									<country key="BE">Belgium</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Yvan</forename><surname>Saeys</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Department of Applied Mathematics, Computer Science and Statistics</orgName>
								<orgName type="institution">Ghent University</orgName>
								<address>
									<postCode>9000</postCode>
									<settlement>Ghent</settlement>
									<country key="BE">Belgium</country>
								</address>
							</affiliation>
							<affiliation key="aff1">
								<orgName type="department">Data Mining and Modeling for Biomedicine</orgName>
								<orgName type="institution">VIB Inflammation Research Center</orgName>
								<address>
									<postCode>9052</postCode>
									<settlement>Ghent</settlement>
									<country key="BE">Belgium</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Calibrated Multi-Probabilistic Prediction as a Defense against Adversarial Attacks</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">366F12193C6E79338C0A6E41F5D637F9</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T11:01+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<abstract/>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Machine learning techniques have made great progress in recent years, obtaining state of the art performance in areas such as natural language processing <ref type="bibr" target="#b2">[3]</ref> as well as image and speech recognition <ref type="bibr" target="#b1">[2]</ref>. However, the theoretical properties of the deep neural networks responsible for this success remain poorly understood. At present, there is no theory which can satisfactorily explain the success of deep learning and many open questions remain <ref type="bibr" target="#b5">[6]</ref>. A peculiar example of this lack of theoretical understanding is the existence of so-called adversarial perturbations <ref type="bibr" target="#b0">[1]</ref>. These are small modifications to the inputs of a model which can drastically change its output, even though the alterations are completely insignificant.</p><p>In this work, we propose a novel defense against adversarial manipulation which aims to scale to realistic problems and provide non-trivial robustness. It is based on methods from conformal prediction and therefore enjoys frequentist guarantees of validity <ref type="bibr" target="#b3">[4]</ref>. Empirical evaluations as well as theoretical results also support the idea that our defense can be scaled to realistic models. We evaluate our method against existing (oblivious) adversarial attacks as well as a white-box attack specifically designed to fool the MultIVAP. We find that these attacks have limited success when the norms of the perturbations are reasonably constrained.</p><p>The basic construction of the MultIVAP is as follows. Given any machine learning classifier, we use the inductive Venn-ABERS predictor algorithm by Vovk et al. <ref type="bibr" target="#b4">[5]</ref> in a one-vs-all manner in order to obtain pairs of probabilities (p </p><formula xml:id="formula_0">(i) 0 , p (i) 1 )</formula><p>form lower and upper bounds on the probability that the given sample belongs to class i. These probabilities are then processed into a multi-probabilistic prediction by solving a mixed integer linear program (MILP). The output of the MultIVAP is the solution to this optimization problem, which consists of a vector of bits (α 1 , . . . , α K ). Here, α i indicates that we can accept the label i for the given input at the ε significance level, where ε ∈ [0, 1] is a user-specified parameter. Table <ref type="table" target="#tab_0">1</ref> shows experimental results when we evaluate the MultIVAP on four different image recognition tasks. For each task, we report several metrics: η, the ∞ norm bound on the magnitude of the perturbations; ε, the significance level at which these results were obtained; accuracy of the MultIVAP and accuracy of the underlying model; the adversarial error of the MultIVAP. Note that on three out of four tasks, the MultIVAP increases the accuracy of the classifier. Also, the adversarial error of the MultIVAP is significantly lower than that of unprotected machine learning classifiers evaluated against adaptive white-box attacks (these are almost invariably close to 100%). The computational overhead we incurred with this construction was roughly linear in the number of classes of the task.</p><p>We conclude that the MultIVAP is a computationally efficient procedure for protecting multi-class classifiers against adversarial perturbations. We make our code available at https://github.com/saeyslab/multivap.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head></head><label></label><figDesc>one for each class. Intuitively, the pair (p</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1 :</head><label>1</label><figDesc>Results of the MultIVAPs on the adversarial white-box attack.</figDesc><table><row><cell>Task</cell><cell>η</cell><cell cols="3">ε Accuracy (baseline) Adversarial error</cell></row><row><cell cols="3">Fashion-MNIST 0.3 24.20%</cell><cell>94.22% (93.84%)</cell><cell>18.96%</cell></row><row><cell>CIFAR-10</cell><cell cols="2">0.03 20.77%</cell><cell>83.36% (81.51%)</cell><cell>27.55%</cell></row><row><cell>Asirra</cell><cell cols="2">0.03 41.86%</cell><cell>88.56% (89.04%)</cell><cell>47.57%</cell></row><row><cell>SVHN</cell><cell cols="2">0.03 25.23%</cell><cell>96.81% (96.40%)</cell><cell>9.85%</cell></row></table></figure>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><p>We thank the NVIDIA Corporation for the donation of a Titan Xp GPU with which we were able to carry out our experiments. Jonathan Peck is sponsored by a fellowship of the Research Foundation Flanders (FWO). Yvan Saeys is an ISAC Marylou Ingram scholar. Copyright c 2019 for this paper by its authors. Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0).</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Wild patterns: Ten years after the rise of adversarial machine learning</title>
		<author>
			<persName><forename type="first">Battista</forename><surname>Biggio</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Fabio</forename><surname>Roli</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Pattern Recognition</title>
		<imprint>
			<biblScope unit="volume">84</biblScope>
			<biblScope unit="page" from="317" to="331" />
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<monogr>
		<title level="m" type="main">Lingvo: a modular and scalable framework for sequence-to-sequence modeling</title>
		<author>
			<persName><forename type="first">Jonathan</forename><surname>Shen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Patrick</forename><surname>Nguyen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Yonghui</forename><surname>Wu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Zhifeng</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Ye</forename><surname>Mia X Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Anjuli</forename><surname>Jia</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Tara</forename><surname>Kannan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Yuan</forename><surname>Sainath</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Chung-Cheng</forename><surname>Cao</surname></persName>
		</author>
		<author>
			<persName><surname>Chiu</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1902.08295</idno>
		<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">The Evolved Transformer</title>
		<author>
			<persName><forename type="first">David</forename><surname>So</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Quoc</forename><surname>Le</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Chen</forename><surname>Liang</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 36th International Conference on Machine Learning</title>
				<editor>
			<persName><forename type="first">Kamalika</forename><surname>Chaudhuri</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">Ruslan</forename><surname>Salakhutdinov</surname></persName>
		</editor>
		<meeting>the 36th International Conference on Machine Learning<address><addrLine>Long Beach, California, USA</addrLine></address></meeting>
		<imprint>
			<publisher>PMLR</publisher>
			<date type="published" when="2019-06">June 2019</date>
			<biblScope unit="volume">97</biblScope>
			<biblScope unit="page" from="5877" to="5886" />
		</imprint>
	</monogr>
	<note>Proceedings of Machine Learning Research</note>
</biblStruct>

<biblStruct xml:id="b3">
	<monogr>
		<title level="m" type="main">Algorithmic learning in a random world</title>
		<author>
			<persName><forename type="first">Vladimir</forename><surname>Vovk</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Alex</forename><surname>Gammerman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Glenn</forename><surname>Shafer</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2005">2005</date>
			<publisher>Springer Science &amp; Business Media</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Large-scale probabilistic predictors with and without guarantees of validity</title>
		<author>
			<persName><forename type="first">Vladimir</forename><surname>Vovk</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Ivan</forename><surname>Petej</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Valentina</forename><surname>Fedorova</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Advances in Neural Information Processing Systems</title>
				<imprint>
			<date type="published" when="2015">2015</date>
			<biblScope unit="page" from="892" to="900" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<monogr>
		<title level="m" type="main">Understanding deep learning requires rethinking generalization</title>
		<author>
			<persName><forename type="first">Chiyuan</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Samy</forename><surname>Bengio</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Moritz</forename><surname>Hardt</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Benjamin</forename><surname>Recht</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Oriol</forename><surname>Vinyals</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1611.03530</idno>
		<imprint>
			<date type="published" when="2016">2016</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
