<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Privacy Safe Representation Learning via Frequency Filtering Encoder</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Jonghu</forename><surname>Jeong</surname></persName>
							<email>jonghu.jeong@deepingsource.io</email>
							<affiliation key="aff0">
								<orgName type="institution">Deeping Source Inc</orgName>
								<address>
									<addrLine>508, Eonju-ro, Gangnam-gu</addrLine>
									<settlement>Seoul</settlement>
									<country key="KR">Republic of Korea</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Minyong</forename><surname>Cho</surname></persName>
							<email>minyong.cho@deepingsource.io</email>
							<affiliation key="aff0">
								<orgName type="institution">Deeping Source Inc</orgName>
								<address>
									<addrLine>508, Eonju-ro, Gangnam-gu</addrLine>
									<settlement>Seoul</settlement>
									<country key="KR">Republic of Korea</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Philipp</forename><surname>Benz</surname></persName>
							<email>philipp.benz@deepingsource.io</email>
							<affiliation key="aff0">
								<orgName type="institution">Deeping Source Inc</orgName>
								<address>
									<addrLine>508, Eonju-ro, Gangnam-gu</addrLine>
									<settlement>Seoul</settlement>
									<country key="KR">Republic of Korea</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Jinwoo</forename><surname>Hwang</surname></persName>
							<email>jinwoo.hwang@deepingsource.io</email>
							<affiliation key="aff0">
								<orgName type="institution">Deeping Source Inc</orgName>
								<address>
									<addrLine>508, Eonju-ro, Gangnam-gu</addrLine>
									<settlement>Seoul</settlement>
									<country key="KR">Republic of Korea</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Jeewook</forename><surname>Kim</surname></persName>
							<email>jeewook.kim@deepingsource.io</email>
							<affiliation key="aff0">
								<orgName type="institution">Deeping Source Inc</orgName>
								<address>
									<addrLine>508, Eonju-ro, Gangnam-gu</addrLine>
									<settlement>Seoul</settlement>
									<country key="KR">Republic of Korea</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Seungkwan</forename><surname>Lee</surname></persName>
							<email>seungkwan.lee@deepingsource.io</email>
							<affiliation key="aff0">
								<orgName type="institution">Deeping Source Inc</orgName>
								<address>
									<addrLine>508, Eonju-ro, Gangnam-gu</addrLine>
									<settlement>Seoul</settlement>
									<country key="KR">Republic of Korea</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Tae-Hoon</forename><surname>Kim</surname></persName>
							<affiliation key="aff0">
								<orgName type="institution">Deeping Source Inc</orgName>
								<address>
									<addrLine>508, Eonju-ro, Gangnam-gu</addrLine>
									<settlement>Seoul</settlement>
									<country key="KR">Republic of Korea</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Privacy Safe Representation Learning via Frequency Filtering Encoder</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">10EB8A5541D71929056DBAC88FA07D62</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T23:22+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>privacy-preserving machine learning</term>
					<term>adversarial representation learning</term>
					<term>image frequency filtering</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Deep learning models are increasingly deployed in real-world applications. These models are often deployed on the server-side and receive user data in an information-rich representation to solve a specific task, such as image classification. Since images can contain sensitive information, which users might not be willing to share, privacy protection becomes increasingly important. Adversarial Representation Learning (ARL) is a common approach to train an encoder that runs on the client-side and obfuscates an image. It is assumed, that the obfuscated image can safely be transmitted and used for the task on the server without privacy concerns. However, in this work, we find that training a reconstruction attacker can successfully recover the original image of existing ARL methods. To this end, we introduce a novel ARL method enhanced through low-pass filtering, limiting the available information amount to be encoded in the frequency domain. Our experimental results reveal that our approach withstands reconstruction attacks while outperforming previous state-of-the-art methods regarding the privacy-utility trade-off. We further conduct a user study to qualitatively assess our defense of the reconstruction attack.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>Service providers, such as Amazon Rekognition and Microsoft Cognitive Services, frequently deploy deep learning models in real-world applications in recent years. The models run on the providers' server can receive and process user information in an information-rich representation to solve a specific task. For example, the users send their face images from their smartphone (client) to the server and receive the processed results, such as face identification. However, the raw images can also contain additional information which users do not consent to reveal or share, violating the users' privacy. An adversary could take over and abuse the images of the users. In one possible attack scenario, adversaries can train a new attacker model (e.g. neural network) that retrieves private attributes, such as gender, emotional state, and race. Even the service provider could have malicious intent without the users' knowledge. Hence, an obfuscation method should be used to protect the users' privacy.</p><p>For privacy protection with deep learning models, several prior works exist ranging from federated learn-Figure <ref type="figure">1</ref>: An overview of our proposed method. The proposed method trains an encoder that obfuscates an input image through a neural net and leverages a frequency filtering module to safely transmit a privacy-sensitive image from a client-side to a server-side. The frequency filtering module helps the encoder to remove private information effectively from the image while retaining its utility to be used for a particular task of interest (utility task) on the server-side. The encoder is trained with the conventional ARL scheme and then deployed to the client-side. Even with the possibility of data leakage during data transmission, malicious attackers can not abuse the obfuscated image for a privacy breach attack (privacy task) since the transmitted data contains information that is only useful for the utility task.</p><p>ing <ref type="bibr" target="#b0">[1,</ref><ref type="bibr" target="#b1">2]</ref>, split learning <ref type="bibr" target="#b2">[3,</ref><ref type="bibr" target="#b3">4]</ref>, differential privacy <ref type="bibr" target="#b4">[5,</ref><ref type="bibr" target="#b5">6,</ref><ref type="bibr" target="#b6">7]</ref>, and homomorphic encryption <ref type="bibr" target="#b7">[8,</ref><ref type="bibr" target="#b8">9,</ref><ref type="bibr" target="#b9">10]</ref> to instance hiding mechanisms <ref type="bibr" target="#b10">[11,</ref><ref type="bibr" target="#b11">12,</ref><ref type="bibr" target="#b12">13,</ref><ref type="bibr" target="#b13">14]</ref>, GAN-based obfuscation techniques <ref type="bibr" target="#b14">[15,</ref><ref type="bibr" target="#b15">16]</ref> and adversarial representation learning <ref type="bibr" target="#b16">[17]</ref>. Among these works, however, adversarial representation learning (ARL) is the one suitable for the service provider to serve users with an obfuscation method. For example, federated learning and instance hiding focus on model training with privacy-safe data, not on inference with obfuscated data <ref type="bibr" target="#b0">[1,</ref><ref type="bibr" target="#b10">11]</ref>. Furthermore, several existing methods suffer under privacy leakage <ref type="bibr" target="#b17">[18,</ref><ref type="bibr" target="#b18">19,</ref><ref type="bibr" target="#b19">20]</ref>, and the degree of computational complexity is too large to be deployed in practice <ref type="bibr" target="#b7">[8,</ref><ref type="bibr" target="#b8">9,</ref><ref type="bibr" target="#b9">10]</ref>. With ARL, the service provider can train an obfuscator model and deploy it to make data obfuscation possible on the user side <ref type="bibr" target="#b20">[21,</ref><ref type="bibr" target="#b21">22]</ref>.</p><p>Most previous ARL methods solve the problem of privacy-safe transmission by optimizing 1) utility task loss and 2) proxy adversary task loss <ref type="bibr" target="#b22">[23,</ref><ref type="bibr" target="#b20">21,</ref><ref type="bibr" target="#b23">24,</ref><ref type="bibr" target="#b21">22]</ref>. They also introduce specific loss-design formulations, model architecture design, and training schemes. The methods are evaluated quantitatively with performance on both utility and adversary tasks. Note that there usually exists a trade-off between privacy and utility. We use a reconstruction attack, to test the quality of the obfuscation. In a reconstruction attack, a new model is trained that takes the obfuscated representation as an input and outputs the original image. As demonstrated in Figure <ref type="figure">2</ref>, the original data of existing ARL methods can successfully be recovered from the obfuscated representation. This result suggests that the private information is still encoded in the obfuscated representations.</p><p>We present a novel ARL method that leverages frequency filtering, leveraging an extreme low-pass frequency filter (Figure <ref type="figure">1</ref>). The representation filtering on the frequency domain effectively limits the amount of information to be encoded. Our experimental results show that our approach outperforms previous state-of-the-art methods regarding the privacy-utility trade-off. We also present that our proposed method withstands the reconstruction attack better than existing ARL methods, which are evaluated through visual metrics and a user study.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Related Work</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Data-privacy in Computer Vision</head><p>For privacy-safe data transmission, several approaches have been proposed to tackle the problem of raw image sharing. Federated learning <ref type="bibr" target="#b0">[1,</ref><ref type="bibr" target="#b1">2]</ref> and split learning <ref type="bibr" target="#b2">[3,</ref><ref type="bibr" target="#b3">4]</ref> aim to train a machine learning model without directly sharing raw images through sharing gradients or a processed representation. These methods usually focus on the model training, and not on inference with obfuscated data. Homomorphic encryption <ref type="bibr" target="#b7">[8,</ref><ref type="bibr" target="#b8">9,</ref><ref type="bibr" target="#b9">10]</ref> attempts to train models on encrypted data, such that the data can be shared in encrypted form and be processed without decryption. Currently, this method suffers from a considerably high computational cost. Instance hiding mechanisms <ref type="bibr" target="#b10">[11,</ref><ref type="bibr" target="#b11">12,</ref><ref type="bibr" target="#b12">13,</ref><ref type="bibr" target="#b13">14]</ref> introduce random pixel mixing and clipping algorithm to perturb images. The perturbed images are used only for the training, and the original Figure <ref type="figure">2</ref>: Results of the reconstruction attack with various methods on CelebA. For a successful defense, the reconstructed image should not reveal 1) the identity of the original image and 2) the privacy attribute (in this case, gender). Our method successfully defends the reconstruction attack while all other approaches fail. Detailed results are further discussed in Section 5.</p><p>images are used for the inference which means that there are still potential threats for data breaches when inferring the target.</p><p>Adversarial Representation Learning (ARL) Another line of work focuses on the training framework of ARL to address the utility-privacy trade-off of (a) mitigation of privacy disclosure while (b) maintaining task utility. ARL methods have found their application in practical scenarios, such as information censoring <ref type="bibr" target="#b24">[25]</ref>, learning fair representations <ref type="bibr" target="#b25">[26,</ref><ref type="bibr" target="#b26">27]</ref>, the mitigation of information leakage <ref type="bibr" target="#b22">[23,</ref><ref type="bibr" target="#b20">21,</ref><ref type="bibr" target="#b23">24]</ref>, collaborative inference <ref type="bibr" target="#b27">[28,</ref><ref type="bibr" target="#b28">29,</ref><ref type="bibr" target="#b21">22]</ref>, and GAN-based obfuscation techniques <ref type="bibr" target="#b14">[15,</ref><ref type="bibr" target="#b15">16]</ref>. Commonly, the ARL framework consists of three entities: 1) an obfuscator, which transforms input data to a private representation that retains utility, 2) a task model, performing the utility task on the data representation, 3) a proxy adversary, attempting to extract sensitive attributes. Recent approaches <ref type="bibr" target="#b29">[30,</ref><ref type="bibr" target="#b30">31,</ref><ref type="bibr" target="#b31">32,</ref><ref type="bibr" target="#b23">24]</ref> represent each component as deep neural networks (DNNs). MaxEnt <ref type="bibr" target="#b22">[23]</ref> formulate the ARL problem as an adversarial non-zero-sum game and minimizes the amount of non-utility information, which they quantify through entropy. Adversarial representation learning with non-linear functions through kernel representation with theoretical guarantees are introduced in <ref type="bibr" target="#b32">[33]</ref>. While most of the previous methods represent the obfuscated output as the intermediate feature of a DNN, Bertran et al. <ref type="bibr" target="#b20">[21]</ref> leverages domain-preserving transformations, i.e. images to images. Above mentioned ARL methods mainly focused on designing special loss functions or model architectures. To the best of our knowledge, our method is the first ARL method that focuses on the effective encoding of privacy-safe representation in the frequency domain.</p><p>There are three common attacks on privacy in machine learning. The first is the membership inference attack <ref type="bibr" target="#b33">[34]</ref>, which attempts to infer whether a data sample is used for the machine learning model training. This attack is more related to the attack on the server-side model, not the transmitted data. The second is the inversion attack <ref type="bibr" target="#b34">[35]</ref> which attempts to infer raw data from processed representation. This is the same attack scenario as the aforementioned reconstruction attack. The last is the information leakage attack <ref type="bibr" target="#b22">[23]</ref>, for which adversaries attempt to infer privacy-related information from obfuscated representation. In this work the inversion attack and the information leakage attack are considered as they are potential threats to transmitted privacy-sensitive images.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Frequency Perspective in Computer Vision</head><p>Prior works have explored the behavior of DNNs from a frequency perspective. Overall, there is solid evidence that both high-frequency features and low-frequency features can be helpful for classification <ref type="bibr" target="#b35">[36,</ref><ref type="bibr" target="#b36">37]</ref>. It has been demonstrated that DNNs have an increased bias toward texture compared to the object's shape <ref type="bibr" target="#b37">[38]</ref>. On the other hand, DNNs trained only on low-pass filtered images also generalize well and are capable of achieving high accuracies <ref type="bibr" target="#b35">[36]</ref>. Yin et al. <ref type="bibr" target="#b35">[36]</ref> shows that adversarial training and Gaussian data augmentation shift DNNs towards utilizing low-frequency information in the input. Wang et al. <ref type="bibr" target="#b36">[37]</ref> points out that convolutional neural networks (CNNs) mainly exploit high-frequency components. Similarly, Abello et al. <ref type="bibr" target="#b38">[39]</ref> find that mid or highlevel frequencies are disproportionately critical for CNNs. Ilyas et al. <ref type="bibr" target="#b39">[40]</ref> also show similar findings that humanimperceptible features with high-frequency properties are sufficient for the model to exhibit high generalization capability.</p><p>In this work, we leverage previous insights that information can be encoded in different frequency ranges of images. We propose encoding information in the lowfrequency band of images to securely transfer them between different parties.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Problem Formulation</head><p>We consider an image dataset 𝑥 ∼ 𝒳 ∈ R 𝐻 ×𝑊 ×3 , where 𝐻 and 𝑊 represent width and height, respectively, along with a number of various attributes 𝑦 ∼ 𝒴. Some of the attributes are private attributes 𝑦 𝑝 ∼ 𝒴 𝑝 and some are utility attributes 𝑦 𝑡 ∼ 𝒴 𝑡 , such that 𝒴 = 𝒴 𝑡 ∪𝒴 𝑝 . Given a utility task model 𝑓 𝑡 , we search for an intermediate representation x , from which 𝑓 𝑡 can infer the utility attributes, but not the privacy attributes. This transformation can also be represented through a DNN 𝑜, termed obfuscator, resulting in 𝑜(𝑥) = x . Note that in prior works, the intermediate representation x was often represented as a feature map differing in shape from the original input images. However, similar to <ref type="bibr" target="#b20">[21]</ref>, we represent the obfuscated representation in the same shape as the original input image. This setting allows us to leverage existing image transformation techniques, such as transforming them into a 2D Fourier representation. Additionally, this form of intermediate representation allows us to analyze the representations visually.</p><p>Threat Model Given the above problem formulation, an attacker can attempt to retrieve information about the private attributes from the intermediate representation. This can be realized either by directly inferring private information from the intermediate representation (information leakage attack) or through the reconstruction of the original input images from the intermediate representations (reconstruction attack). In the information leakage attack scenario an attacker is able to obtain data pairs consisting of the corresponding intermediate representation and their respective private attributes { x , 𝑦 𝑝 }. In this scenario an attacker can attempt to train a model 𝑓 𝑎 , which leaks the private information from the representations 𝑓 𝑎 ( x ) = 𝑦 𝑝 . In the reconstruction attack, given image pairs of the original image and the intermediate representation {𝑥, x } the attacker attempts to obtain a model 𝑓 𝑟 which retrieves the original image 𝑥 from the intermediate representation 𝑓 𝑟 ( x ) = 𝑥. In this work, we represent both attacker models 𝑓 𝑎 and 𝑓 𝑟 through DNNs, since they are proven to be powerful for image processing tasks.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Methodology</head><p>Fourier Transformation Fourier transform is a common tool to perform frequency analysis <ref type="bibr" target="#b40">[41]</ref>. We consider the 2D discrete Fourier transformation ℱ ∶ R 𝑊 ×𝐻 → C 𝑊 ×𝐻 and the inverse Fourier transformation as ℱ −1 . After applying ℱ on an image, low frequencies are located in the center of a Fourier image, while high frequencies are located toward the boundaries. For low-pass filtering, we set all frequency components outside of a central circle with radius 𝑟 in the frequency domain to zero and apply ℱ −1 afterward. We normalize the radius to be in the range of [0, 1] by considering the center of the image as 0 and the corner as 1. We indicate low-pass filtering as 𝐿𝑃.</p><p>Frequency Obfuscation We depict our proposed methodology in Figure <ref type="figure">1</ref>. Given an input image, the objective is to obfuscate the image to achieve the best privacyutility trade-off. Our obfuscator module consists of an encoder architecture followed by frequency-filtering. We choose the commonly used U-Net <ref type="bibr" target="#b41">[42]</ref> architecture as our encoder and pass the original image through it. Formally, we express this as 𝑒(𝑥), where we indicated the encoder with 𝑒. The subsequent frequency filtering is realized via a low-pass filter 𝐿𝑃(𝑒(𝑥)). This procedure completes the generation of the intermediate representation through the obfuscator x = 𝑜(𝑥) = 𝐿𝑃(𝑒(𝑥)). During obfuscator training, we leverage a task model and a proxy adversary. The objective of the task model is to predict the utility attribute from the intermediate representation. The respective task loss can be calculated with 𝑙 𝑡 = 𝔼[ℒ 𝑡 (𝑓 𝑡 (𝑜(𝑥)), 𝑦 𝑡 )], where ℒ 𝑡 indicates the task loss function, which is the cross-entropy function in our setup. The objective of proxy adversary model is to leak the privacy attribute from the intermediate representation. The proxy adversary loss can be calculated as 𝑙 𝑝 = 𝔼[ℒ 𝑝 (𝑓 𝑎 (𝑜(𝑥)), 𝑦 𝑝 )], where ℒ 𝑝 indicates the privacy loss function, which is also represented as the crossentropy function. The obfuscator loss is represented as</p><formula xml:id="formula_0">𝑙 𝑜 = 𝑙 𝑡 − 𝑙 𝑝 .</formula><p>Similar to the scenario introduced in DISCO <ref type="bibr" target="#b21">[22]</ref> a practical application scenario of our proposed approach is when the obfuscator module is present on a trusted client device, which sends the intermediate feature representations to a server. Since an adversary can intercept the communication between client and server, or the server can also be malicious, we consider the server-side an untrusted entity.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Evaluation Protocol</head><p>In the following, we outline our evaluation protocol. We follow the general ARL evaluation protocol <ref type="bibr" target="#b21">[22,</ref><ref type="bibr" target="#b22">23]</ref>. Given an image classification dataset, we specify certain classes as the utility and privacy tasks, respectively. Based on the chosen tasks, following our proposed method we obtain an obfuscator and a utility task model. Note that this includes training proxy adversaries. After training, we evaluate the models on the utility task and report the accuracy as utility. Then we freeze the weights of the obfuscator and train an adversary model to predict the privacy attributes and report the accuracy as privacy. To assess the privacy-utility trade-off, we measure their difference (Δ).</p><p>Additionally, we report the performance bounds. Theo-retically, the utility (higher the better) is upper bounded by 100%. In practice, however, we consider the upper bound as the utility performance of a ResNet18 <ref type="bibr" target="#b42">[43]</ref> model trained on the original images. For privacy (lower the better), we consider the lower bound as the random guess for the privacy attribute.</p><p>We also perform a reconstruction attack on the obfuscated images to recover corresponding original images. We evaluate the reconstruction attacks quantitative and qualitatively by calculating similarity scores between the original and reconstructed images and conducting a user study on the reconstructed images.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Experiments</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.1.">Setup</head><p>Datasets We conduct experiments on CelebA <ref type="bibr" target="#b43">[44]</ref>, FairFace <ref type="bibr" target="#b44">[45]</ref>, and CIFAR10 <ref type="bibr" target="#b45">[46]</ref>. Following the utility and privacy task setting from DISCO <ref type="bibr" target="#b21">[22]</ref>, we set "Smiling" as the utility attribute and "Male" as the privacy attribute for CelebA, "Gender" as the utility attribute, and "Race" as the privacy attribute for FairFace. For CIFAR10, the utility task is defined as classifying living objects (e.g. "bird", "cat", etc.) or non-living objects (e.g. "airplane", "automobile", etc.) and the privacy task as classifying the separate 10 classes.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Implementation details</head><p>The encoder is a lightweight variant of U-Net <ref type="bibr" target="#b41">[42]</ref>, with 4× fewer intermediate feature channels than the original version. We use an extreme low pass filter with radius, 𝑟 = 0.01 for CelebA and Fair-Face, and 𝑟 = 0.05 for CIFAR10. We apply a center-circled filter, which can adjust the level of obfuscation by changing its radius (bandwidth). Section 6.2 discusses the effect of the radius. We normalize the radius by the length from the filter's center to the corner to make the value in the range [0, 1]. For both the utility and privacy task models, we use ResNet-18 <ref type="bibr" target="#b42">[43]</ref>, and use the same dataset for training both models. We use Adam <ref type="bibr" target="#b46">[47]</ref> optimizer for all 3 models with learning rate 10 −4 for U-Net and 10 −3 for the ResNet-18 models. We evaluate the top-1 accuracy for both utility and privacy tasks. We used the lightweight U-Net as the reconstructor for the reconstruction attack. The reconstructor adversary is trained with the MSE loss between the original and the reconstructed images. The reconstructed images are evaluated using MSE, 𝐿 1 , SSIM <ref type="bibr" target="#b47">[48]</ref>, MS-SSIM <ref type="bibr" target="#b48">[49]</ref>, PSNR <ref type="bibr" target="#b49">[50]</ref>, and LPIPS <ref type="bibr" target="#b50">[51]</ref>. MSE, 𝐿 1 , and PSNR compare the images pixel-wise while SSIM and MS-SSIM compare structural similarity (e.g., brightness, contrast) between the images. LPIPS uses a pre-trained neural network's feature map for comparison. These metrics are commonly used for comparing the similarity between images <ref type="bibr" target="#b21">[22,</ref><ref type="bibr" target="#b23">24,</ref><ref type="bibr" target="#b51">52]</ref>  Evaluation of the privacy-utility trade-off. The upper/lower arrow suggests that each value is higher/lower the better. Our method shows the biggest gap between privacy and utility accuracy among all the datasets. Note that the privacy accuracy is based on the newly trained adversary model which is trained with the fully trained and frozen obfuscation model.</p><p>as a proxy of human vision.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Compared Methods</head><p>We compare our method with various baselines. As a simple baseline obfuscator, we add Gaussian noise sampled from 𝒩 (0, 𝜎 2 ) to the input image while obeying the image range of pixels in the range [0, 1]. We indicate this method with Noise. We use 𝜎 2 = 4 for CelebA and FairFace and 𝜎 2 = 0.64 for CIFAR10, which obfuscate the images sufficiently. To investigate the sole effect of the low-pass filtering, we apply only the low-pass filter to the raw images. We name this baseline as LP. Complementary, we also compare the U-Net without the low-pass filtering module as an obfuscator. We call it U-Net. This setup is similar to DeepObfuscator <ref type="bibr" target="#b23">[24]</ref> which uses an encoder, task model, and a proxy adversary. However, since DeepObfuscator has not open-sourced their code, we used our U-Net encoder as a method to compare. Finally, we compare our method to the state-of-the-art ARL method DISCO <ref type="bibr" target="#b21">[22]</ref>, which selectively removes features via channel pruning in the latent space.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.2.">Results</head><p>Table <ref type="table" target="#tab_0">1</ref> shows a comparison between the privacy and utility accuracy of each obfuscation method. Our method resulted in the highest gap between utility and privacy accuracy on all datasets. For the methods without encoder (i.e. Noise and LP), the accuracy for both utility and privacy decreases compared to training with the original image since these methods obfuscate images without any prior knowledge of the tasks. These methods cannot selectively restrict information for high utility and low privacy leakage. U-Net showed high utility accuracy but failed to defend against the privacy attack, although it is trained with a proxy adversary. We conjecture that simply taking the guidance of the proxy model loss is not enough for the encoder to learn to restrict information. Similarity scores between the original image and the reconstructed ones on CelebA. The upper/lower arrow suggests that each value is higher/lower the better, respectively. Our approach shows the best dissimilarity among all the metrics.</p><formula xml:id="formula_1">Method MSE ↑ 𝐿 1 ↑ SSIM ↓ MS-SSIM ↓ PSNR ↓</formula><p>Our method is a combination of LP and U-Net, and learns to encode a representation into the restricted bandwidth, which is limited by the frequency filtering module. This limited bandwidth helps the encoder to learn how to extract utility information effectively and remove privacy attributes to fully leverage the limited bandwidth. While the same data is used to train both utility and adversary models, which is a generous and unrealistic condition for the attackers to have, we found the adversary model performed poorly. DISCO shows the lowest privacy accuracy among all the datasets. However, the utility accuracy is lower than our method, so the utility-privacy gap is smaller than ours.</p><p>In terms of the visual quality, our obfuscated representations appear as simple globs of color, making them unrecognizable to human observers (Figure <ref type="figure">1</ref>). The obfuscated representations from other methods also appear obfuscated to the human eye. However, applying our best effort reconstruction attack, it is possible to reconstruct the original image or infer the privacy attribute (i.e. gender) from reconstructed images. (Figure <ref type="figure">2</ref>). The reconstructed images from our method successfully defend identity reconstruction and privacy attribute leakage, with the reconstructed images all being relatively similar to each other. The quantitative results of the reconstruc-Figure <ref type="figure">3</ref>: Result of the user study on reconstructed images of CelebA. We asked the participants to classify gender (male/female) on 180 images such as Figure <ref type="figure">2</ref>. The participants correctly distinguished the gender of original images and reconstructed images from the three methods (Noise, U-Net, and DISCO) with more than 90% accuracy. Our method and LP effectively confused the participants with gender-neutral faces (45.83% and 56.9% of correct answers ratio each), while ours is slightly better than LP in terms of obfuscation. We also plot the ideal case of the user study to show our method's near-perfect superiority against the reconstruction attack.</p><p>tion attack in Table <ref type="table" target="#tab_1">2</ref> further confirm this since all scores achieve the best results in terms of dissimilarity for our approach. We note that an adversary model trained with the reconstructed images to infer the privacy attributes performs worse than directly training the model with the obfuscated images since the reconstructed images are processed from the obfuscated images.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.3.">User Study</head><p>We present a user study to show our method's robustness against the reconstruction attack on CelebA. Since the privacy task for the dataset is gender classification, the reconstructed image's gender should not be correctly classified by a human observer if the obfuscation is successful. To conduct the experiment, we randomly sampled 30 images (15 for male and 15 for female), for which ResNet18 classifies the gender correctly. By doing so, we balanced each class and addressed the ambiguity of the labels to prevent unfair results. Then, we obfuscated the images using each of the techniques and reconstructed them with their respective attacker models from Section 5.1. Examples of reconstructed images are shown in Figure <ref type="figure">2</ref>. We presented 180 reconstructed images to a group of people and asked them to identify whether the person in the reconstructed image is male, female, or cannot be judged. We provided the last option to let the users skip the examples that are hard to judge. The test subjects were randomly selected and consist of 30 people who live in Seoul, South Korea, and are in their 20s and 30s.</p><p>As shown in Figure <ref type="figure">3</ref>, people correctly identify the gender for the original images and the reconstructed ones The privacy-utility gap of the high-pass filtering module on FairFace. Our low-pass filtering module shows the best privacy-utility gap compared to the high-pass filter with the various filter radii.</p><p>from the methods Noise, U-Net, and DISCO. More than 90% of answers were correct for the three methods. LP showed a relatively low correct ratio (56.9%) and a high "cannot judge" ratio (6.19%). Our method showed the best for both, the lowest correct ratio of 45.83% and the highest "cannot judge" ratio of 7.02%. We consider the 50% ratio for each "correct" and "wrong" answer as a random guess since the labels for the test datasets are balanced. Additionally, we note that "cannot judge" can be considered as a random guess since without this option, the users would have done a random choice. The results indicate that our approach successfully protects against reconstruction attacks in terms of human vision. The results also align with the quantitative results (Table <ref type="table" target="#tab_1">2</ref>).</p><p>In terms of obfuscation, our method shows the best results, followed by LP. It reconfirms the usefulness of our architecture design, the combination of the encoder and the frequency filtering module.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.">Ablation Study</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.1.">High-pass filter</head><p>Previously, we presented the effect of the low-pass frequency filtering module on ARL. The module appropriately limits the amount of encoded information in the obfuscated image. It retains the information at a lowfrequency range. Using a high-pass filter, we can leverage the same intuition, by limiting the information to be encoded in the high-frequency bandwidth. However, in the following, we will present results indicating that the low-pass filter is the superior method to use. We conduct the same experiment from Section 5.2 on FairFace with a high-pass filtering module for 5 radii (0.80, 0.85, 0.90, 0.95, 0.99). Contrary to the low-pass filtering, the filter removes frequencies inside the filter radius, which leads to a radius of 0.99 as the most extreme high-pass filter. We call this method HP.</p><p>The respective results are presented in Table <ref type="table" target="#tab_2">3</ref>. As the filtering gets more extreme, the utility accuracy decreases together with the privacy accuracy. The table also shows that our approach with a low-pass filter from Table <ref type="table" target="#tab_0">1</ref> outperforms all results from the high-pass filter regarding the privacy-utility gap. The best privacy-utility gap with the high-pass filter is 63.16% with a radius of 0.95, which is 2.88%p lower than for the approach with low-pass filtering. It has been demonstrated that DNNs can learn from low-pass filtered images more efficiently than highpass filtered ones <ref type="bibr" target="#b35">[36]</ref>. Especially with the extreme highpass (r=0.99), the model did not learn for both, the utility and privacy tasks.</p><p>Furthermore, from a practical point of view, we need to reduce the size of the obfuscated image to reduce the cost of transmission or storage. The most commonly used JPEG compression algorithm leverages the filtering of high frequency. If we use a high-pass filter ARL method, encoded information in the high-frequency range would be lost. To this end, encoding information into the lowfrequency range is more suitable than the opposite to utilize the conventional compression algorithms further.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.2.">The effect of filter radius</head><p>One of the key points of our proposed method is the frequency filtering module. The module has only one parameter to consider, the filter's radius. To gain insight into choosing the parameter, we conducted experiments with various radii. The same experiment from Section 5 on FairFace is done with 5 radii (0.01, 0.015, 0.02, 0.025, 0.03). The radius of 0.01 is the most extreme low-pass filter.</p><p>Figure <ref type="figure" target="#fig_0">4</ref> (left) shows a trend of consistent utility accuracy and increasing privacy accuracy. The utility accuracies are around 89% with a small variance. The privacy accuracies show an increasing tendency from 23.64% to 30.45% as the radius increases. It leads the privacy-utility gap to decrease (Figure <ref type="figure" target="#fig_0">4</ref>, right).</p><p>The increased privacy accuracy aligns with our intuition of limiting information in the obfuscated representation. The wider radius allows the representation to have more information, leading the adversary to exploit it for a privacy attack easily. Note that the utility accuracy did not decrease even with the harshest filter. We speculate that the extremely low-pass filtered representation is enough for these specific utility tasks. Figure <ref type="figure" target="#fig_0">4</ref> and Table 3 confirm that the radius is a crucial factor of privacy and utility accuracy. Thus the radius is a hyperparameter that should be tuned based on the privacy-utility gap.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="7.">Conclusion</head><p>This work proposes a novel ARL method based on frequency filtering, which is robust to privacy leakage attacks while maintaining task utility. Our experiments suggest that a combination of neural-net encoder and lowpass filter improves ARL training for the quantitative and qualitative metrics. The method outperforms other compared methods for the quantitative measure of privacyutility trade-off and reconstruction attack (Section 5). Our user study suggests that the proposed method effectively defends against reconstruction attacks (Section 5.3). The ablation experiments justified the use of a low-pass filter and also showed that the filter radius adjusts the privacy-utility trade-off (Section 6).</p><p>For future work we consider the optimization of the client-side model to reduce the computation burden by using a lightweight architecture such as MobileNetV3 <ref type="bibr" target="#b52">[53]</ref>. Furthermore, an adaptive selection of the frequencyfiltering hyperparameter might increase the utility accuracy and decrease the privacy accuracy.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 4 :</head><label>4</label><figDesc>Figure 4: (Left) Privacy and utility accuracy under each radius of the low-pass filter. The experiments are conducted on FairFace. (Right) Privacy-utility trade-off. Delta represents the performance gap between utility and privacy.</figDesc><graphic coords="7,89.29,84.19,104.17,77.91" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1</head><label>1</label><figDesc>and we consider them</figDesc><table><row><cell></cell><cell></cell><cell>Fairface</cell><cell></cell><cell>CelebA</cell><cell></cell><cell></cell><cell>CIFAR10</cell></row><row><cell>Method</cell><cell cols="7">Privacy ↓ Utility ↑ Δ ↑ Privacy ↓ Utility ↑ Δ ↑ Privacy ↓ Utility ↑ Δ ↑</cell></row><row><cell>Perf. Bounds</cell><cell>19.03</cell><cell>90.16 71.13</cell><cell>57.43</cell><cell>93.32</cell><cell>35.89</cell><cell>10.00</cell><cell>98.79 78.79</cell></row><row><cell>Noise</cell><cell>42.61</cell><cell>74.33 31.72</cell><cell>91.71</cell><cell>85.38</cell><cell>-6.33</cell><cell>54.37</cell><cell>87.77 33.40</cell></row><row><cell>LP</cell><cell>31.93</cell><cell>64.77 32.84</cell><cell>76.52</cell><cell cols="2">63.69 -12.83</cell><cell>47.05</cell><cell>85.76 38.71</cell></row><row><cell>U-Net</cell><cell>51.52</cell><cell>86.40 34.88</cell><cell>87.21</cell><cell>93.12</cell><cell>5.91</cell><cell>85.05</cell><cell>95.45 10.40</cell></row><row><cell>DISCO</cell><cell>19.00</cell><cell>81.50 62.50</cell><cell>61.20</cell><cell>91.00</cell><cell>29.80</cell><cell>22.30</cell><cell>91.98 69.68</cell></row><row><cell>Ours</cell><cell>23.63</cell><cell>89.67 66.04</cell><cell>61.60</cell><cell cols="2">93.27 31.67</cell><cell>22.58</cell><cell>92.95 70.37</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 2</head><label>2</label><figDesc>LPIPS ↑</figDesc><table><row><cell>Noise</cell><cell>584.88 16.97 0.6017 0.7776</cell><cell>20.46 0.3714</cell></row><row><cell>LP</cell><cell>1889.15 32.10 0.4632 0.5390</cell><cell>15.37 0.5537</cell></row><row><cell>U-Net</cell><cell>390.34 13.81 0.7505 0.8839</cell><cell>22.22 0.1809</cell></row><row><cell cols="2">DISCO 567.17 15.94 0.5765 0.7611</cell><cell>20.60 0.4351</cell></row><row><cell>Ours</cell><cell>3689.50 48.08 0.4240 0.4728</cell><cell>12.47 0.6145</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>Table 3</head><label>3</label><figDesc></figDesc><table><row><cell>Method</cell><cell cols="2">Privacy ↓ Utility ↑</cell><cell>Δ ↑</cell></row><row><cell>HP (r=0.80)</cell><cell>26.19</cell><cell>89.03</cell><cell>62.84</cell></row><row><cell>HP (r=0.85)</cell><cell>26.28</cell><cell>89.13</cell><cell>62.85</cell></row><row><cell>HP (r=0.90)</cell><cell>28.94</cell><cell>88.00</cell><cell>59.06</cell></row><row><cell>HP (r=0.95)</cell><cell>24.96</cell><cell>88.12</cell><cell>63.16</cell></row><row><cell>HP (r=0.99)</cell><cell>19.03</cell><cell>52.88</cell><cell>33.85</cell></row><row><cell>LP (r=0.01)</cell><cell>23.63</cell><cell>89.67</cell><cell>66.04</cell></row></table></figure>
		</body>
		<back>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<monogr>
		<author>
			<persName><forename type="first">J</forename><surname>Konečnỳ</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><forename type="middle">B</forename><surname>Mcmahan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><forename type="middle">X</forename><surname>Yu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Richtárik</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">T</forename><surname>Suresh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Bacon</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1610.05492</idno>
		<title level="m">Federated learning: Strategies for improving communication efficiency</title>
				<imprint>
			<date type="published" when="2016">2016</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Advances and open problems in federated learning</title>
		<author>
			<persName><forename type="first">P</forename><surname>Kairouz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><forename type="middle">B</forename><surname>Mcmahan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Avent</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Bellet</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Bennis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">N</forename><surname>Bhagoji</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Bonawitz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Charles</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Cormode</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Cummings</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Foundations and Trends® in Machine Learning</title>
				<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Distributed learning of deep neural network over multiple agents</title>
		<author>
			<persName><forename type="first">O</forename><surname>Gupta</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Raskar</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Network and Computer Applications</title>
		<imprint>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<monogr>
		<author>
			<persName><forename type="first">P</forename><surname>Vepakomma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Gupta</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Swedish</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Raskar</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1812.00564</idno>
		<title level="m">Split learning for health: Distributed deep learning without sharing raw patient data</title>
				<imprint>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Differential privacy: A survey of results</title>
		<author>
			<persName><forename type="first">C</forename><surname>Dwork</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">International conference on theory and applications of models of computation</title>
				<imprint>
			<date type="published" when="2008">2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<monogr>
		<author>
			<persName><forename type="first">Z</forename><surname>Ji</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><forename type="middle">C</forename><surname>Lipton</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Elkan</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1412.7584</idno>
		<title level="m">Differential privacy and machine learning: a survey and review</title>
				<imprint>
			<date type="published" when="2014">2014</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Deep learning with differential privacy</title>
		<author>
			<persName><forename type="first">M</forename><surname>Abadi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Chu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Goodfellow</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><forename type="middle">B</forename><surname>Mcmahan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Mironov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Talwar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Zhang</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">ACM SIGSAC conference on computer and communications security</title>
				<imprint>
			<date type="published" when="2016">2016</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<monogr>
		<author>
			<persName><forename type="first">E</forename><surname>Hesamifard</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Takabi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Ghasemi</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1711.05189</idno>
		<title level="m">Cryptodl: Deep neural networks over encrypted data</title>
				<imprint>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">{GAZELLE}: A low latency framework for secure neural network inference</title>
		<author>
			<persName><forename type="first">C</forename><surname>Juvekar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Vaikuntanathan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Chandrakasan</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">USENIX Security Symposium</title>
				<imprint>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Towards deep neural network training on encrypted data</title>
		<author>
			<persName><forename type="first">K</forename><surname>Nandakumar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Ratha</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Pankanti</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Halevi</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Conference on Computer Vision and Pattern Recognition Workshops</title>
				<imprint>
			<publisher>CVPR-W</publisher>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Mixup based privacy preserving mixed collaboration learning</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Fu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Xu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Mi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Wang</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">International Conference on Service-Oriented System Engineering (SOSE)</title>
				<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Instahide: Instance-hiding schemes for private distributed learning</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Huang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Song</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Arora</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">International Conference on Machine Learning (ICML)</title>
				<imprint>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<monogr>
		<author>
			<persName><forename type="first">M</forename><surname>Shin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Hwang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Kim</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Park</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Bennis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S.-L</forename><surname>Kim</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2006.05148</idno>
		<title level="m">Xor mixup: Privacy-preserving data augmentation for one-shot federated learning</title>
				<imprint>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b13">
	<monogr>
		<author>
			<persName><forename type="first">E</forename><surname>Borgnia</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Geiping</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Cherepanova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Fowl</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Gupta</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Ghiasi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Huang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Goldblum</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Goldstein</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2103.02079</idno>
		<title level="m">Dp-instahide: Provably defusing poisoning and backdoor attacks with differentially private data augmentations</title>
				<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b14">
	<monogr>
		<author>
			<persName><forename type="first">T.-H</forename><surname>Kim</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Kang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Pulli</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Choi</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1901.00098</idno>
		<title level="m">Training with the invisibles: Obfuscating images to share safely for learning visual recognition models</title>
				<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<author>
			<persName><forename type="first">C</forename><surname>Xu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Ren</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Qin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Ren</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Ganobfuscator: Mitigating information leakage under gan via differential privacy</title>
				<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">Large scale adversarial representation learning</title>
		<author>
			<persName><forename type="first">J</forename><surname>Donahue</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Simonyan</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Advances in Neural Information Processing Systems</title>
		<imprint>
			<biblScope unit="volume">32</biblScope>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<monogr>
		<author>
			<persName><forename type="first">L</forename><surname>Lyu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Yu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><surname>Yang</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2003.02133</idno>
		<title level="m">Threats to federated learning: A survey</title>
				<imprint>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b18">
	<analytic>
		<title level="a" type="main">Unleashing the tiger: Inference attacks on split learning</title>
		<author>
			<persName><forename type="first">D</forename><surname>Pasquini</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Ateniese</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Bernaschi</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">ACM SIGSAC Conference on Computer and Communications Security</title>
				<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<monogr>
		<author>
			<persName><forename type="first">O</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Sun</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Yang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Gao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Xie</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Smith</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Wang</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2102.08504</idno>
		<title level="m">Label leakage and protection in two-party split learning</title>
				<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b20">
	<analytic>
		<title level="a" type="main">Adversarially learned representations for information obfuscation and inference</title>
		<author>
			<persName><forename type="first">M</forename><surname>Bertran</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Martinez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Papadaki</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><surname>Qiu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Rodrigues</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Reeves</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Sapiro</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">International Conference on Machine Learning (ICML)</title>
				<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b21">
	<analytic>
		<title level="a" type="main">Disco: Dynamic and invariant sensitive channel obfuscation for deep neural networks</title>
		<author>
			<persName><forename type="first">A</forename><surname>Singh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Chopra</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Garza</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Vepakomma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Sharma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Raskar</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Conference on Computer Vision and Pattern Recognition (CVPR)</title>
				<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b22">
	<analytic>
		<title level="a" type="main">Mitigating information leakage in image representations: A maximum entropy approach</title>
		<author>
			<persName><forename type="first">P</forename><forename type="middle">C</forename><surname>Roy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><forename type="middle">N</forename><surname>Boddeti</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition</title>
				<meeting>the IEEE/CVF Conference on Computer Vision and Pattern Recognition</meeting>
		<imprint>
			<date type="published" when="2019">2019</date>
			<biblScope unit="page" from="2586" to="2594" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b23">
	<analytic>
		<title level="a" type="main">Deepobfuscator: Obfuscating intermediate representations with privacy-preserving adversarial learning on smartphones</title>
		<author>
			<persName><forename type="first">A</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Guo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Yang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><forename type="middle">D</forename><surname>Salim</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Chen</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">International Conference on Internet-of-Things Design and Implementation</title>
				<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b24">
	<analytic>
		<title level="a" type="main">Censoring representations with an adversary</title>
		<author>
			<persName><forename type="first">H</forename><surname>Edwards</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Storkey</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">International Conference on Learning Representations (ICLR)</title>
				<imprint>
			<date type="published" when="2016">2016</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b25">
	<monogr>
		<author>
			<persName><forename type="first">C</forename><surname>Louizos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Swersky</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Welling</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Zemel</surname></persName>
		</author>
		<title level="m">The variational fair autoencoder</title>
				<imprint>
			<date type="published" when="2016">2016</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b26">
	<analytic>
		<title level="a" type="main">Learning adversarially fair and transferable representations</title>
		<author>
			<persName><forename type="first">D</forename><surname>Madras</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Creager</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Pitassi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Zemel</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">International Conference on Machine Learning (ICML)</title>
				<imprint>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b27">
	<analytic>
		<title level="a" type="main">Nopeek: Information leakage reduction to share activations in distributed deep learning</title>
		<author>
			<persName><forename type="first">P</forename><surname>Vepakomma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Singh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Gupta</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Raskar</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">ternational Conference on Data Mining Workshops (ICDMW)</title>
				<imprint>
			<date type="published" when="2020">2020. 2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b28">
	<analytic>
		<title level="a" type="main">A hybrid deep learning architecture for privacy-preserving mobile analytics</title>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">A</forename><surname>Osia</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">S</forename><surname>Shamsabadi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Sajadmanesh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Taheri</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Katevas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><forename type="middle">R</forename><surname>Rabiee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><forename type="middle">D</forename><surname>Lane</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Haddadi</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Internet of Things Journal</title>
		<imprint>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b29">
	<analytic>
		<title level="a" type="main">Learning privacy preserving encodings through adversarial training</title>
		<author>
			<persName><forename type="first">F</forename><surname>Pittaluga</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Koppal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Chakrabarti</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Winter Conference on Applications of Computer Vision (WACV)</title>
				<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b30">
	<analytic>
		<title level="a" type="main">Privacy adversarial network: representation learning for mobile data privacy</title>
		<author>
			<persName><forename type="first">S</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Du</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Shrivastava</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Zhong</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies</title>
				<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b31">
	<analytic>
		<title level="a" type="main">Towards privacypreserving visual recognition via adversarial training: A pilot study</title>
		<author>
			<persName><forename type="first">Z</forename><surname>Wu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Jin</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">European Conference on Computer Vision (ECCV)</title>
				<imprint>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b32">
	<analytic>
		<title level="a" type="main">On the global optima of kernelized adversarial representation learning</title>
		<author>
			<persName><forename type="first">B</forename><surname>Sadeghi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Yu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Boddeti</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">International Conference on Computer Vision (ICCV)</title>
				<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b33">
	<analytic>
		<title level="a" type="main">Membership inference attacks against machine learning models</title>
		<author>
			<persName><forename type="first">R</forename><surname>Shokri</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Stronati</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Song</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Shmatikov</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Symposium on security and privacy (SP)</title>
				<imprint>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b34">
	<analytic>
		<title level="a" type="main">Model inversion attacks that exploit confidence information and basic countermeasures</title>
		<author>
			<persName><forename type="first">M</forename><surname>Fredrikson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Jha</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Ristenpart</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">ACM SIGSAC conference on computer and communications security</title>
				<imprint>
			<date type="published" when="2015">2015</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b35">
	<analytic>
		<title level="a" type="main">A fourier perspective on model robustness in computer vision</title>
		<author>
			<persName><forename type="first">D</forename><surname>Yin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">G</forename><surname>Lopes</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Shlens</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><forename type="middle">D</forename><surname>Cubuk</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Gilmer</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Advances in neural information processing systems (NeurIPS)</title>
				<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b36">
	<analytic>
		<title level="a" type="main">Highfrequency component helps explain the generalization of convolutional neural networks</title>
		<author>
			<persName><forename type="first">H</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Wu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Huang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><forename type="middle">P</forename><surname>Xing</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Conference on Computer Vision and Pattern Recognition</title>
				<imprint>
			<publisher>CVPR</publisher>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b37">
	<analytic>
		<title level="a" type="main">Imagenet-trained cnns are biased towards texture; increasing shape bias improves accuracy and robustness</title>
		<author>
			<persName><forename type="first">R</forename><surname>Geirhos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Rubisch</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Michaelis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Bethge</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><forename type="middle">A</forename><surname>Wichmann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Brendel</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">International Conference on Learning Representations (ICLR)</title>
				<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b38">
	<analytic>
		<title level="a" type="main">Dissecting the high-frequency bias in convolutional neural networks</title>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">A</forename><surname>Abello</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Hirata</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Wang</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition</title>
				<meeting>the IEEE/CVF Conference on Computer Vision and Pattern Recognition</meeting>
		<imprint>
			<date type="published" when="2021">2021</date>
			<biblScope unit="page" from="863" to="871" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b39">
	<analytic>
		<title level="a" type="main">Adversarial examples are not bugs, they are features</title>
		<author>
			<persName><forename type="first">A</forename><surname>Ilyas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Santurkar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Tsipras</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Engstrom</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Tran</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Madry</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Advances in neural information processing systems</title>
				<imprint>
			<publisher>NeurIPS</publisher>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b40">
	<monogr>
		<title level="m" type="main">Two-dimensional signal and image processing</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">S</forename><surname>Lim</surname></persName>
		</author>
		<imprint>
			<date type="published" when="1990">1990</date>
			<pubPlace>Englewood Cliffs</pubPlace>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b41">
	<monogr>
		<author>
			<persName><forename type="first">O</forename><surname>Ronneberger</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Fischer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Brox</surname></persName>
		</author>
		<title level="m">International Conference on Medical image computing and computer-assisted intervention</title>
				<imprint>
			<date type="published" when="2015">2015</date>
		</imprint>
	</monogr>
	<note>U-net: Convolutional networks for biomedical image segmentation</note>
</biblStruct>

<biblStruct xml:id="b42">
	<analytic>
		<title level="a" type="main">Deep residual learning for image recognition</title>
		<author>
			<persName><forename type="first">K</forename><surname>He</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Ren</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Sun</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Conference on computer vision and pattern recognition (CVPR)</title>
				<imprint>
			<date type="published" when="2016">2016</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b43">
	<analytic>
		<title level="a" type="main">Deep learning face attributes in the wild</title>
		<author>
			<persName><forename type="first">Z</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Luo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Tang</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">International Conference on Computer Vision (ICCV)</title>
				<imprint>
			<date type="published" when="2015">2015</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b44">
	<analytic>
		<title level="a" type="main">Fairface: Face attribute dataset for balanced race, gender, and age for bias measurement and mitigation</title>
		<author>
			<persName><forename type="first">K</forename><surname>Karkkainen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Joo</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Winter Conference on Applications of Computer Vision (WACV)</title>
				<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b45">
	<monogr>
		<title level="m" type="main">Learning multiple layers of features from tiny images</title>
		<author>
			<persName><forename type="first">A</forename><surname>Krizhevsky</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2009">2009</date>
		</imprint>
	</monogr>
	<note type="report_type">Technical Report</note>
</biblStruct>

<biblStruct xml:id="b46">
	<monogr>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">P</forename><surname>Kingma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Ba</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1412.6980</idno>
		<title level="m">Adam: A method for stochastic optimization</title>
				<imprint>
			<date type="published" when="2014">2014</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b47">
	<analytic>
		<title level="a" type="main">Image quality assessment: from error visibility to structural similarity</title>
		<author>
			<persName><forename type="first">Z</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Bovik</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Sheikh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Simoncelli</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Transactions on Image Processing</title>
		<imprint>
			<date type="published" when="2004">2004</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b48">
	<analytic>
		<title level="a" type="main">Multiscale structural similarity for image quality assessment</title>
		<author>
			<persName><forename type="first">Z</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><forename type="middle">P</forename><surname>Simoncelli</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">C</forename><surname>Bovik</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">The Thrity-Seventh Asilomar Conference on Signals, Systems &amp; Computers</title>
				<imprint>
			<publisher>Ieee</publisher>
			<date type="published" when="2003">2003. 2003</date>
			<biblScope unit="volume">2</biblScope>
			<biblScope unit="page" from="1398" to="1402" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b49">
	<analytic>
		<title level="a" type="main">Image quality metrics: Psnr vs. ssim</title>
		<author>
			<persName><forename type="first">A</forename><surname>Horé</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Ziou</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">International Conference on Pattern Recognition</title>
				<imprint>
			<date type="published" when="2010">2010</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b50">
	<analytic>
		<title level="a" type="main">The unreasonable effectiveness of deep features as a perceptual metric</title>
		<author>
			<persName><forename type="first">R</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Isola</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">A</forename><surname>Efros</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Shechtman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Wang</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the IEEE conference on computer vision and pattern recognition</title>
				<meeting>the IEEE conference on computer vision and pattern recognition</meeting>
		<imprint>
			<date type="published" when="2018">2018</date>
			<biblScope unit="page" from="586" to="595" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b51">
	<analytic>
		<title level="a" type="main">Analyzing and improving the image quality of stylegan</title>
		<author>
			<persName><forename type="first">T</forename><surname>Karras</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Laine</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Aittala</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Hellsten</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Lehtinen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Aila</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)</title>
				<meeting>the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)</meeting>
		<imprint>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b52">
	<analytic>
		<title level="a" type="main">Searching for mobilenetv3</title>
		<author>
			<persName><forename type="first">A</forename><surname>Howard</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Sandler</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Chu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L.-C</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Tan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Zhu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Pang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Vasudevan</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the IEEE/CVF International Conference on Computer Vision</title>
				<meeting>the IEEE/CVF International Conference on Computer Vision</meeting>
		<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
