<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Representation Learning via Frequency Filtering Encoder</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Tae-hoon Kim</string-name>
          <email>jeewook.kim@deepingsource.io</email>
          <email>pete.kim@deepingsource.io</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Deeping Source Inc.</institution>
          ,
          <addr-line>508, Eonju-ro, Gangnam-gu, Seoul</addr-line>
          ,
          <country>Republic of Korea</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Jonghu Jeong</institution>
          ,
          <addr-line>Minyong Cho, Philipp Benz, Jinwoo Hwang, Jeewook Kim, Seungkwan Lee</addr-line>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>The IJCAI-ECAI-22 Workshop on Artificial Intelligence Safety</institution>
        </aff>
      </contrib-group>
      <abstract>
        <p>Deep learning models are increasingly deployed in real-world applications. These models are often deployed on the server-side and receive user data in an information-rich representation to solve a specific task, such as image classification. Since images can contain sensitive information, which users might not be willing to share, privacy protection becomes increasingly important. Adversarial Representation Learning (ARL) is a common approach to train an encoder that runs on the client-side and obfuscates an image. It is assumed, that the obfuscated image can safely be transmitted and used for the task on the server without privacy concerns. However, in this work, we find that training a reconstruction attacker can successfully recover the original image of existing ARL methods. To this end, we introduce a novel ARL method enhanced through low-pass ifltering, limiting the available information amount to be encoded in the frequency domain. Our experimental results reveal that our approach withstands reconstruction attacks while outperforming previous state-of-the-art methods regarding the privacy-utility trade-of. We further conduct a user study to qualitatively assess our defense of the reconstruction attack. privacy-preserving machine learning, adversarial representation learning, image frequency filtering</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <sec id="sec-1-1">
        <title>Service providers, such as Amazon Rekognition and Mi</title>
        <p>crosoft Cognitive Services, frequently deploy deep
learning models in real-world applications in recent years.
The models run on the providers’ server can receive and
process user information in an information-rich
representation to solve a specific task. For example, the users
send their face images from their smartphone (client) to
the server and receive the processed results, such as face
identification. However, the raw images can also contain
veal or share, violating the users’ privacy. An adversary
could take over and abuse the images of the users. In
one possible attack scenario, adversaries can train a new
attacker model (e.g. neural network) that retrieves
private attributes, such as gender, emotional state, and race.</p>
      </sec>
      <sec id="sec-1-2">
        <title>Even the service provider could have malicious intent</title>
        <p>without the users’ knowledge. Hence, an obfuscation
method should be used to protect the users’ privacy.</p>
      </sec>
      <sec id="sec-1-3">
        <title>For privacy protection with deep learning models,</title>
        <p>
          ing [
          <xref ref-type="bibr" rid="ref1 ref2">1, 2</xref>
          ], split learning [
          <xref ref-type="bibr" rid="ref3 ref4">3, 4</xref>
          ], diferential privacy [
          <xref ref-type="bibr" rid="ref5 ref6 ref7">5, 6, 7</xref>
          ],
and homomorphic encryption [
          <xref ref-type="bibr" rid="ref10 ref8 ref9">8, 9, 10</xref>
          ] to instance
hiding mechanisms [
          <xref ref-type="bibr" rid="ref11 ref12 ref13 ref14">11, 12, 13, 14</xref>
          ], GAN-based obfuscation
techniques [
          <xref ref-type="bibr" rid="ref15 ref16">15, 16</xref>
          ] and adversarial representation
learning [
          <xref ref-type="bibr" rid="ref17">17</xref>
          ]. Among these works, however, adversarial
representation learning (ARL) is the one suitable for the service
additional information which users do not consent to re- Figure 1: An overview of our proposed method. The
proseveral prior works exist ranging from federated learn- can not abuse the obfuscated image for a privacy breach attack
provider to serve users with an obfuscation method. For
example, federated learning and instance hiding focus on
model training with privacy-safe data, not on inference
with obfuscated data [
          <xref ref-type="bibr" rid="ref1 ref11">1, 11</xref>
          ]. Furthermore, several
existing methods sufer under privacy leakage [
          <xref ref-type="bibr" rid="ref18 ref19 ref20">18, 19, 20</xref>
          ], and
the degree of computational complexity is too large to
be deployed in practice [
          <xref ref-type="bibr" rid="ref10 ref8 ref9">8, 9, 10</xref>
          ]. With ARL, the service
provider can train an obfuscator model and deploy it to
make data obfuscation possible on the user side [
          <xref ref-type="bibr" rid="ref21">21, 22</xref>
          ].
        </p>
        <p>
          Most previous ARL methods solve the problem of
privacy-safe transmission by optimizing 1) utility task
loss and 2) proxy adversary task loss [
          <xref ref-type="bibr" rid="ref21">23, 21, 24, 22</xref>
          ]. They
also introduce specific loss-design formulations, model
architecture design, and training schemes. The methods
are evaluated quantitatively with performance on both
utility and adversary tasks. Note that there usually exists
a trade-of between privacy and utility. We use a
reconstruction attack, to test the quality of the obfuscation. In
a reconstruction attack, a new model is trained that takes
the obfuscated representation as an input and outputs
the original image. As demonstrated in Figure 2, the
original data of existing ARL methods can successfully be
recovered from the obfuscated representation. This re- Figure 2: Results of the reconstruction attack with
varisult suggests that the private information is still encoded ous methods on CelebA. For a successful defense, the
reconin the obfuscated representations. structed image should not reveal 1) the identity of the original
        </p>
        <p>We present a novel ARL method that leverages fre- image and 2) the privacy attribute (in this case, gender ). Our
quency filtering, leveraging an extreme low-pass fre- method successfully defends the reconstruction attack while
quency filter (Figure 1). The representation filtering on all other approaches fail. Detailed results are further discussed
the frequency domain efectively limits the amount of in- in Section 5.
formation to be encoded. Our experimental results show
that our approach outperforms previous state-of-the-art
methods regarding the privacy-utility trade-of. We also images are used for the inference which means that there
present that our proposed method withstands the recon- are still potential threats for data breaches when inferring
struction attack better than existing ARL methods, which the target.
are evaluated through visual metrics and a user study.</p>
      </sec>
    </sec>
    <sec id="sec-2">
      <title>2. Related Work</title>
      <p>
        Data-privacy in Computer Vision For privacy-safe
data transmission, several approaches have been
proposed to tackle the problem of raw image sharing.
Federated learning [
        <xref ref-type="bibr" rid="ref1 ref2">1, 2</xref>
        ] and split learning [
        <xref ref-type="bibr" rid="ref3 ref4">3, 4</xref>
        ] aim to
train a machine learning model without directly sharing
raw images through sharing gradients or a processed
representation. These methods usually focus on the
model training, and not on inference with obfuscated
data. Homomorphic encryption [
        <xref ref-type="bibr" rid="ref10 ref8 ref9">8, 9, 10</xref>
        ] attempts to
train models on encrypted data, such that the data can
be shared in encrypted form and be processed without
decryption. Currently, this method sufers from a
considerably high computational cost. Instance hiding
mechanisms [
        <xref ref-type="bibr" rid="ref11 ref12 ref13 ref14">11, 12, 13, 14</xref>
        ] introduce random pixel mixing
and clipping algorithm to perturb images. The perturbed
images are used only for the training, and the original
Adversarial Representation Learning (ARL)
Another line of work focuses on the training framework
of ARL to address the utility-privacy trade-of of (a)
mitigation of privacy disclosure while (b) maintaining
task utility. ARL methods have found their
application in practical scenarios, such as information
censoring [25], learning fair representations [26, 27], the
mitigation of information leakage [
        <xref ref-type="bibr" rid="ref21">23, 21, 24</xref>
        ], collaborative
inference [28, 29, 22], and GAN-based obfuscation
techniques [
        <xref ref-type="bibr" rid="ref15 ref16">15, 16</xref>
        ]. Commonly, the ARL framework consists
of three entities: 1) an obfuscator, which transforms input
data to a private representation that retains utility, 2) a
task model, performing the utility task on the data
representation, 3) a proxy adversary, attempting to extract
sensitive attributes. Recent approaches [30, 31, 32, 24]
represent each component as deep neural networks (DNNs).
      </p>
      <p>MaxEnt [23] formulate the ARL problem as an
adversarial non-zero-sum game and minimizes the amount of
non-utility information, which they quantify through
entropy. Adversarial representation learning with
nonlinear functions through kernel representation with
theoretical guarantees are introduced in [33]. While most
of the previous methods represent the obfuscated output
leverages domain-preserving transformations, i.e.
images to images. Above mentioned ARL methods mainly
focused on designing special loss functions or model
architectures. To the best of our knowledge, our method
is the first ARL method that focuses on the efective
encoding of privacy-safe representation in the frequency
domain.</p>
      <p>
        There are three common attacks on privacy in ma- tor, resulting in () =  ̂ . Note that in prior works, the
chine learning. The first is the membership inference
attack [34], which attempts to infer whether a data sam- a feature map difering in shape from the original input
intermediate representation  ̂ was often represented as
as the intermediate feature of a DNN, Bertran et al. [
        <xref ref-type="bibr" rid="ref21">21</xref>
        ]  and  represent width and height, respectively, along
model, not the transmitted data. The second is the inver- input image. This setting allows us to leverage existing
      </p>
    </sec>
    <sec id="sec-3">
      <title>3. Problem Formulation</title>
      <sec id="sec-3-1">
        <title>We consider an image dataset  ∼  ∈</title>
        <p>R × ×3
, where
with a number of various attributes  ∼</p>
        <p>
          . Some of the
attributes are private attributes   ∼   and some are
utility attributes   ∼   , such that  = 
 ∪  
. Given a
utility task model   , we search for an intermediate
representation  ̂, from which   can infer the utility attributes,
but not the privacy attributes. This transformation can
also be represented through a DNN  , termed
obfuscaple is used for the machine learning model training. This
attack is more related to the attack on the server-side
sion attack [35] which attempts to infer raw data from
processed representation. This is the same attack
scenario as the aforementioned reconstruction attack. The
images. However, similar to [
          <xref ref-type="bibr" rid="ref21">21</xref>
          ], we represent the
obfuscated representation in the same shape as the original
image transformation techniques, such as transforming
them into a 2D Fourier representation. Additionally, this
form of intermediate representation allows us to analyze
last is the information leakage attack [23], for which ad- the representations visually.
versaries attempt to infer privacy-related information
from obfuscated representation. In this work the in- Threat Model
        </p>
      </sec>
      <sec id="sec-3-2">
        <title>Given the above problem formulation, version attack and the information leakage attack are considered as they are potential threats to transmitted privacy-sensitive images.</title>
        <sec id="sec-3-2-1">
          <title>Frequency Perspective in Computer Vision</title>
          <p>Prior
works have explored the behavior of DNNs from a
frequency perspective. Overall, there is solid evidence that
both high-frequency features and low-frequency features
can be helpful for classification [ 36, 37]. It has been
demonstrated that DNNs have an increased bias toward
texture compared to the object’s shape [38]. On the other
hand, DNNs trained only on low-pass filtered images
also generalize well and are capable of achieving high
accuracies [36]. Yin et al. [36] shows that adversarial
training and Gaussian data augmentation shift DNNs
towards utilizing low-frequency information in the
input. Wang et al. [37] points out that convolutional neural
an attacker can attempt to retrieve information about the
private attributes from the intermediate representation.</p>
          <p>This can be realized either by directly inferring private
information from the intermediate representation
(information leakage attack) or through the reconstruction of
the original input images from the intermediate
representations (reconstruction attack). In the information leakage
attack scenario an attacker is able to obtain data pairs
consisting of the corresponding intermediate
representation and their respective private attributes {, ̂  }. In
this scenario an attacker can attempt to train a model
 , which leaks the private information from the
repre
sentations   ()̂ =</p>
          <p>. In the reconstruction attack, given
image pairs of the original image and the intermediate
representation {, }̂</p>
          <p>the attacker attempts to obtain a
model   which retrieves the original image  from the
intermediate representation   ()̂ =  . In this work, we
since they are proven to be powerful for image processing
networks (CNNs) mainly exploit high-frequency compo- represent both attacker models 
 and   through DNNs,
nents. Similarly, Abello et al. [39] find that mid or
highlevel frequencies are disproportionately critical for CNNs. tasks.</p>
        </sec>
      </sec>
      <sec id="sec-3-3">
        <title>Ilyas et al. [40] also show similar findings that humanimperceptible features with high-frequency properties are suficient for the model to exhibit high generalization capability.</title>
      </sec>
      <sec id="sec-3-4">
        <title>In this work, we leverage previous insights that infor</title>
        <p>mation can be encoded in diferent frequency ranges of
images. We propose encoding information in the
lowfrequency band of images to securely transfer them
between diferent parties.
4.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Methodology</title>
      <p>C ×</p>
      <sec id="sec-4-1">
        <title>Fourier Transformation</title>
        <p>Fourier transform is a
common tool to perform frequency analysis [41]. We consider
the 2D discrete Fourier transformation ℱ ∶ R ×</p>
        <p>
          →
and the inverse Fourier transformation as ℱ −1.
After applying ℱ on an image, low frequencies are located
in the center of a Fourier image, while high frequencies
are located toward the boundaries. For low-pass
filtering, we set all frequency components outside of a central
circle with radius  in the frequency domain to zero and retically, the utility (higher the better) is upper bounded
apply ℱ −1 afterward. We normalize the radius to be in by 100%. In practice, however, we consider the upper
the range of [
          <xref ref-type="bibr" rid="ref1">0, 1</xref>
          ] by considering the center of the image bound as the utility performance of a ResNet18 [43]
as 0 and the corner as 1. We indicate low-pass filtering model trained on the original images. For privacy (lower
as  . the better), we consider the lower bound as the random
guess for the privacy attribute.
        </p>
        <p>Frequency Obfuscation We depict our proposed We also perform a reconstruction attack on the
obfusmethodology in Figure 1. Given an input image, the objec- cated images to recover corresponding original images.
tive is to obfuscate the image to achieve the best privacy- We evaluate the reconstruction attacks quantitative and
utility trade-of. Our obfuscator module consists of an qualitatively by calculating similarity scores between the
encoder architecture followed by frequency-filtering. We original and reconstructed images and conducting a user
choose the commonly used U-Net [42] architecture as study on the reconstructed images.
our encoder and pass the original image through it.
Formally, we express this as () , where we indicated the 5. Experiments
encoder with  . The subsequent frequency filtering is
realized via a low-pass filter  (()) . This procedure 5.1. Setup
completes the generation of the intermediate
representation through the obfuscator  =̂ () =  (()) . Dur- Datasets We conduct experiments on CelebA [44],
ing obfuscator training, we leverage a task model and a FairFace [45], and CIFAR10 [46]. Following the utility and
proxy adversary. The objective of the task model is to privacy task setting from DISCO [22], we set “Smiling”
predict the utility attribute from the intermediate rep- as the utility attribute and “Male” as the privacy attribute
resentation. The respective task loss can be calculated for CelebA, “Gender” as the utility attribute, and “Race”
with   = [ℒ  (  (()),   )], where ℒ indicates the task as the privacy attribute for FairFace. For CIFAR10, the
loss function, which is the cross-entropy function in utility task is defined as classifying living objects ( e.g.
our setup. The objective of proxy adversary model is “bird”, “cat”, etc.) or non-living objects (e.g. “airplane”,
to leak the privacy attribute from the intermediate rep- “automobile”, etc.) and the privacy task as classifying the
resentation. The proxy adversary loss can be calculated separate 10 classes.
as   = [ℒ  (  (()),   )], where ℒ indicates the
privacy loss function, which is also represented as the cross- Implementation details The encoder is a lightweight
entropy function. The obfuscator loss is represented as variant of U-Net [42], with 4× fewer intermediate feature
  =   −   . channels than the original version. We use an extreme</p>
        <p>
          Similar to the scenario introduced in DISCO [22] a low pass filter with radius,  = 0.01 for CelebA and
Fairpractical application scenario of our proposed approach Face, and  = 0.05 for CIFAR10. We apply a center-circled
is when the obfuscator module is present on a trusted iflter, which can adjust the level of obfuscation by
changclient device, which sends the intermediate feature repre- ing its radius (bandwidth). Section 6.2 discusses the efect
sentations to a server. Since an adversary can intercept of the radius. We normalize the radius by the length from
the communication between client and server, or the the filter’s center to the corner to make the value in the
server can also be malicious, we consider the server-side range [
          <xref ref-type="bibr" rid="ref1">0, 1</xref>
          ]. For both the utility and privacy task models,
an untrusted entity. we use ResNet-18 [43], and use the same dataset for
training both models. We use Adam [47] optimizer for all 3
Evaluation Protocol In the following, we outline our models with learning rate 10−4 for U-Net and 10−3 for the
evaluation protocol. We follow the general ARL eval- ResNet-18 models. We evaluate the top-1 accuracy for
uation protocol [22, 23]. Given an image classification both utility and privacy tasks. We used the lightweight
dataset, we specify certain classes as the utility and pri- U-Net as the reconstructor for the reconstruction attack.
vacy tasks, respectively. Based on the chosen tasks, fol- The reconstructor adversary is trained with the MSE
lowing our proposed method we obtain an obfuscator loss between the original and the reconstructed images.
and a utility task model. Note that this includes training The reconstructed images are evaluated using MSE,  1,
proxy adversaries. After training, we evaluate the mod- SSIM [48], MS-SSIM [49], PSNR [50], and LPIPS [51].
els on the utility task and report the accuracy as utility. MSE,  1, and PSNR compare the images pixel-wise while
Then we freeze the weights of the obfuscator and train an SSIM and MS-SSIM compare structural similarity (e.g.,
adversary model to predict the privacy attributes and re- brightness, contrast) between the images. LPIPS uses a
port the accuracy as privacy. To assess the privacy-utility pre-trained neural network’s feature map for comparison.
trade-of, we measure their diference ( Δ). These metrics are commonly used for comparing the
simAdditionally, we report the performance bounds. Theo- ilarity between images [22, 24, 52] and we consider them
Perf. Bounds
Noise
LP
U-Net
DISCO
Ours
        </p>
        <sec id="sec-4-1-1">
          <title>Compared Methods We compare our method with</title>
          <p>
            various baselines. As a simple baseline obfuscator, we
add Gaussian noise sampled from  (0,  2) to the input
image while obeying the image range of pixels in the
range [
            <xref ref-type="bibr" rid="ref1">0, 1</xref>
            ]. We indicate this method with Noise. We Table 2
use  2 = 4 for CelebA and FairFace and  2 = 0.64 for Similarity scores between the original image and the
reconCIFAR10, which obfuscate the images suficiently. To structed ones on CelebA. The upper/lower arrow suggests that
investigate the sole efect of the low-pass filtering, we each value is higher/lower the better, respectively. Our
apapply only the low-pass filter to the raw images. We name proach shows the best dissimilarity among all the metrics.
this baseline as LP. Complementary, we also compare
the U-Net without the low-pass filtering module as an
obfuscator. We call it U-Net. This setup is similar to Our method is a combination of LP and U-Net, and learns
DeepObfuscator [24] which uses an encoder, task model, to encode a representation into the restricted bandwidth,
and a proxy adversary. However, since DeepObfuscator which is limited by the frequency filtering module. This
has not open-sourced their code, we used our U-Net limited bandwidth helps the encoder to learn how to
exencoder as a method to compare. Finally, we compare our tract utility information efectively and remove privacy
method to the state-of-the-art ARL method DISCO [22], attributes to fully leverage the limited bandwidth. While
which selectively removes features via channel pruning the same data is used to train both utility and adversary
in the latent space. models, which is a generous and unrealistic condition
for the attackers to have, we found the adversary model
5.2. Results performed poorly. DISCO shows the lowest privacy
accuracy among all the datasets. However, the utility accuracy
Table 1 shows a comparison between the privacy and util- is lower than our method, so the utility-privacy gap is
ity accuracy of each obfuscation method. Our method smaller than ours.
resulted in the highest gap between utility and privacy In terms of the visual quality, our obfuscated
repreaccuracy on all datasets. For the methods without en- sentations appear as simple globs of color, making them
coder (i.e. Noise and LP ), the accuracy for both utility and unrecognizable to human observers (Figure 1). The
obprivacy decreases compared to training with the origi- fuscated representations from other methods also appear
nal image since these methods obfuscate images without obfuscated to the human eye. However, applying our
any prior knowledge of the tasks. These methods cannot best efort reconstruction attack, it is possible to
reconselectively restrict information for high utility and low struct the original image or infer the privacy attribute (i.e.
privacy leakage. U-Net showed high utility accuracy but gender) from reconstructed images. (Figure 2). The
refailed to defend against the privacy attack, although it constructed images from our method successfully defend
is trained with a proxy adversary. We conjecture that identity reconstruction and privacy attribute leakage,
simply taking the guidance of the proxy model loss is not with the reconstructed images all being relatively similar
enough for the encoder to learn to restrict information. to each other. The quantitative results of the
reconstruction attack in Table 2 further confirm this since all scores
achieve the best results in terms of dissimilarity for our
approach. We note that an adversary model trained with
the reconstructed images to infer the privacy attributes
performs worse than directly training the model with
the obfuscated images since the reconstructed images are
processed from the obfuscated images.
5.3. User Study
HP (r=0.80)
HP (r=0.85)
HP (r=0.90)
HP (r=0.95)
HP (r=0.99)
LP (r=0.01)
26.19
26.28
28.94
24.96
19.03
from the methods Noise, U-Net, and DISCO. More than
90% of answers were correct for the three methods. LP
showed a relatively low correct ratio (56.9%) and a high
“cannot judge” ratio (6.19%). Our method showed the
best for both, the lowest correct ratio of 45.83% and the
highest “cannot judge” ratio of 7.02%. We consider the
50% ratio for each “correct” and “wrong” answer as a
random guess since the labels for the test datasets are
balanced. Additionally, we note that “cannot judge” can be
considered as a random guess since without this option,
the users would have done a random choice. The results
indicate that our approach successfully protects against
reconstruction attacks in terms of human vision. The
results also align with the quantitative results (Table 2).
          </p>
          <p>In terms of obfuscation, our method shows the best
results, followed by LP. It reconfirms the usefulness of our
architecture design, the combination of the encoder and
the frequency filtering module.</p>
        </sec>
        <sec id="sec-4-1-2">
          <title>We present a user study to show our method’s robustness</title>
          <p>against the reconstruction attack on CelebA. Since the 6. Ablation Study
privacy task for the dataset is gender classification, the
reconstructed image’s gender should not be correctly clas- 6.1. High-pass filter
sified by a human observer if the obfuscation is successful.</p>
          <p>To conduct the experiment, we randomly sampled 30 im- Previously, we presented the efect of the low-pass
freages (15 for male and 15 for female), for which ResNet18 quency filtering module on ARL. The module
appropriclassifies the gender correctly. By doing so, we balanced ately limits the amount of encoded information in the
each class and addressed the ambiguity of the labels to obfuscated image. It retains the information at a
lowprevent unfair results. Then, we obfuscated the images frequency range. Using a high-pass filter, we can
leverusing each of the techniques and reconstructed them age the same intuition, by limiting the information to be
with their respective attacker models from Section 5.1. encoded in the high-frequency bandwidth. However, in
Examples of reconstructed images are shown in Figure 2. the following, we will present results indicating that the
We presented 180 reconstructed images to a group of low-pass filter is the superior method to use.
people and asked them to identify whether the person We conduct the same experiment from Section 5.2 on
in the reconstructed image is male, female, or cannot be FairFace with a high-pass filtering module for 5 radii
judged. We provided the last option to let the users skip (0.80, 0.85, 0.90, 0.95, 0.99). Contrary to the low-pass
the examples that are hard to judge. The test subjects ifltering, the filter removes frequencies inside the filter
were randomly selected and consist of 30 people who live radius, which leads to a radius of 0.99 as the most extreme
in Seoul, South Korea, and are in their 20s and 30s. high-pass filter. We call this method HP.</p>
          <p>As shown in Figure 3, people correctly identify the gen- The respective results are presented in Table 3. As the
der for the original images and the reconstructed ones filtering gets more extreme, the utility accuracy decreases
a privacy attack easily. Note that the utility accuracy
did not decrease even with the harshest filter. We
speculate that the extremely low-pass filtered representation
is enough for these specific utility tasks. Figure 4 and
Table 3 confirm that the radius is a crucial factor of privacy
and utility accuracy. Thus the radius is a hyperparameter
that should be tuned based on the privacy-utility gap.</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>7. Conclusion</title>
      <p>This work proposes a novel ARL method based on
frequency filtering, which is robust to privacy leakage
attacks while maintaining task utility. Our experiments
together with the privacy accuracy. The table also shows suggest that a combination of neural-net encoder and
lowthat our approach with a low-pass filter from Table 1 pass filter improves ARL training for the quantitative and
outperforms all results from the high-pass filter regarding qualitative metrics. The method outperforms other
comthe privacy-utility gap. The best privacy-utility gap with pared methods for the quantitative measure of
privacythe high-pass filter is 63.16% with a radius of 0.95, which utility trade-of and reconstruction attack (Section 5).
is 2.88%p lower than for the approach with low-pass Our user study suggests that the proposed method
efecifltering. It has been demonstrated that DNNs can learn tively defends against reconstruction attacks (Section 5.3).
from low-pass filtered images more eficiently than high- The ablation experiments justified the use of a low-pass
pass filtered ones [ 36]. Especially with the extreme high- filter and also showed that the filter radius adjusts the
pass (r=0.99), the model did not learn for both, the utility privacy-utility trade-of (Section 6).
and privacy tasks. For future work we consider the optimization of the</p>
      <p>Furthermore, from a practical point of view, we need to client-side model to reduce the computation burden by
usreduce the size of the obfuscated image to reduce the cost ing a lightweight architecture such as MobileNetV3 [53].
of transmission or storage. The most commonly used Furthermore, an adaptive selection of the
frequencyJPEG compression algorithm leverages the filtering of ifltering hyperparameter might increase the utility
accuhigh frequency. If we use a high-pass filter ARL method, racy and decrease the privacy accuracy.
encoded information in the high-frequency range would
be lost. To this end, encoding information into the
lowfrequency range is more suitable than the opposite to References
utilize the conventional compression algorithms further.
6.2. The efect of filter radius
One of the key points of our proposed method is the
frequency filtering module. The module has only one
parameter to consider, the filter’s radius. To gain insight
into choosing the parameter, we conducted experiments
with various radii. The same experiment from Section 5
on FairFace is done with 5 radii (0.01, 0.015, 0.02, 0.025,
0.03). The radius of 0.01 is the most extreme low-pass
iflter.</p>
      <p>Figure 4 (left) shows a trend of consistent utility
accuracy and increasing privacy accuracy. The utility
accuracies are around 89% with a small variance. The privacy
accuracies show an increasing tendency from 23.64% to
30.45% as the radius increases. It leads the privacy-utility
gap to decrease (Figure 4, right).</p>
      <p>The increased privacy accuracy aligns with our
intuition of limiting information in the obfuscated
representation. The wider radius allows the representation to have
more information, leading the adversary to exploit it for
[34] R. Shokri, M. Stronati, C. Song, V. Shmatikov, Mem- quality assessment: from error visibility to
strucbership inference attacks against machine learning tural similarity, Transactions on Image Processing
models, in: Symposium on security and privacy (2004).</p>
      <p>(SP), 2017. [49] Z. Wang, E. P. Simoncelli, A. C. Bovik, Multiscale
[35] M. Fredrikson, S. Jha, T. Ristenpart, Model inversion structural similarity for image quality assessment,
attacks that exploit confidence information and ba- in: The Thrity-Seventh Asilomar Conference on
sic countermeasures, in: ACM SIGSAC conference Signals, Systems &amp; Computers, 2003, volume 2, Ieee,
on computer and communications security, 2015. 2003, pp. 1398–1402.
[36] D. Yin, R. G. Lopes, J. Shlens, E. D. Cubuk, J. Gilmer, [50] A. Horé, D. Ziou, Image quality metrics: Psnr
A fourier perspective on model robustness in com- vs. ssim, in: International Conference on Pattern
puter vision, in: Advances in neural information Recognition, 2010.</p>
      <p>processing systems (NeurIPS), 2019. [51] R. Zhang, P. Isola, A. A. Efros, E. Shechtman,
[37] H. Wang, X. Wu, Z. Huang, E. P. Xing, High- O. Wang, The unreasonable efectiveness of deep
frequency component helps explain the generaliza- features as a perceptual metric, in: Proceedings of
tion of convolutional neural networks, in: Confer- the IEEE conference on computer vision and
patence on Computer Vision and Pattern Recognition tern recognition, 2018, pp. 586–595.
(CVPR), 2020. [52] T. Karras, S. Laine, M. Aittala, J. Hellsten, J.
Lehti[38] R. Geirhos, P. Rubisch, C. Michaelis, M. Bethge, F. A. nen, T. Aila, Analyzing and improving the
imWichmann, W. Brendel, Imagenet-trained cnns are age quality of stylegan, in: Proceedings of the
biased towards texture; increasing shape bias im- IEEE/CVF Conference on Computer Vision and
Patproves accuracy and robustness., in: International tern Recognition (CVPR), 2020.</p>
      <p>Conference on Learning Representations (ICLR), [53] A. Howard, M. Sandler, G. Chu, L.-C. Chen, B. Chen,
2019. M. Tan, W. Wang, Y. Zhu, R. Pang, V. Vasudevan,
[39] A. A. Abello, R. Hirata, Z. Wang, Dissecting the et al., Searching for mobilenetv3, in: Proceedings
high-frequency bias in convolutional neural net- of the IEEE/CVF International Conference on
Comworks, in: Proceedings of the IEEE/CVF Confer- puter Vision, 2019.
ence on Computer Vision and Pattern Recognition,
2021, pp. 863–871.
[40] A. Ilyas, S. Santurkar, D. Tsipras, L. Engstrom,</p>
      <p>B. Tran, A. Madry, Adversarial examples are not
bugs, they are features, Advances in neural
information processing systems (NeurIPS) (2019).
[41] J. S. Lim, Two-dimensional signal and image
pro</p>
      <p>cessing, Englewood Clifs (1990).
[42] O. Ronneberger, P. Fischer, T. Brox, U-net:
Convolutional networks for biomedical image segmentation,
in: International Conference on Medical image
computing and computer-assisted intervention, 2015.
[43] K. He, X. Zhang, S. Ren, J. Sun, Deep residual
learning for image recognition, in: Conference on
computer vision and pattern recognition (CVPR), 2016.
[44] Z. Liu, P. Luo, X. Wang, X. Tang, Deep learning face
attributes in the wild, in: International Conference
on Computer Vision (ICCV), 2015.
[45] K. Karkkainen, J. Joo, Fairface: Face attribute
dataset for balanced race, gender, and age for bias
measurement and mitigation, in: Winter
Conference on Applications of Computer Vision (WACV),
2021.
[46] A. Krizhevsky, Learning multiple layers of features</p>
      <p>from tiny images, Technical Report, 2009.
[47] D. P. Kingma, J. Ba, Adam: A method for
stochastic optimization, arXiv preprint arXiv:1412.6980
(2014).
[48] Z. Wang, A. Bovik, H. Sheikh, E. Simoncelli, Image</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>J.</given-names>
            <surname>Konečny</surname>
          </string-name>
          ̀ , H. B.
          <string-name>
            <surname>McMahan</surname>
            ,
            <given-names>F. X.</given-names>
          </string-name>
          <string-name>
            <surname>Yu</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Richtárik</surname>
            ,
            <given-names>A. T.</given-names>
          </string-name>
          <string-name>
            <surname>Suresh</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Bacon</surname>
          </string-name>
          ,
          <article-title>Federated learning: Strategies for improving communication eficiency</article-title>
          ,
          <source>arXiv preprint arXiv:1610.05492</source>
          (
          <year>2016</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>P.</given-names>
            <surname>Kairouz</surname>
          </string-name>
          , H. B.
          <string-name>
            <surname>McMahan</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Avent</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Bellet</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Bennis</surname>
            ,
            <given-names>A. N.</given-names>
          </string-name>
          <string-name>
            <surname>Bhagoji</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          <string-name>
            <surname>Bonawitz</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          <string-name>
            <surname>Charles</surname>
            , G. Cormode,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Cummings</surname>
          </string-name>
          , et al.,
          <article-title>Advances and open problems in federated learning</article-title>
          ,
          <source>Foundations and Trends® in Machine Learning</source>
          (
          <year>2021</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>O.</given-names>
            <surname>Gupta</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Raskar</surname>
          </string-name>
          ,
          <article-title>Distributed learning of deep neural network over multiple agents</article-title>
          ,
          <source>Journal of Network and Computer Applications</source>
          (
          <year>2018</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>P.</given-names>
            <surname>Vepakomma</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Gupta</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Swedish</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Raskar</surname>
          </string-name>
          ,
          <article-title>Split learning for health: Distributed deep learning without sharing raw patient data</article-title>
          , arXiv preprint arXiv:
          <year>1812</year>
          .
          <volume>00564</volume>
          (
          <year>2018</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>C.</given-names>
            <surname>Dwork</surname>
          </string-name>
          ,
          <article-title>Diferential privacy: A survey of results</article-title>
          ,
          <source>in: International conference on theory and applications of models of computation</source>
          ,
          <year>2008</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Ji</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z. C.</given-names>
            <surname>Lipton</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Elkan</surname>
          </string-name>
          ,
          <article-title>Diferential privacy and machine learning: a survey and review</article-title>
          ,
          <source>arXiv preprint arXiv:1412.7584</source>
          (
          <year>2014</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>M.</given-names>
            <surname>Abadi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Chu</surname>
          </string-name>
          , I. Goodfellow, H. B.
          <string-name>
            <surname>McMahan</surname>
            , drigues, G. Reeves, G. Sapiro, Adversarially learned
            <given-names>I.</given-names>
          </string-name>
          <string-name>
            <surname>Mironov</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          <string-name>
            <surname>Talwar</surname>
            ,
            <given-names>L. Zhang,</given-names>
          </string-name>
          <article-title>Deep learning with representations for information obfuscation and indiferential privacy, in: ACM SIGSAC conference ference</article-title>
          ,
          <source>in: International Conference on Machine on computer and communications security</source>
          ,
          <year>2016</year>
          .
          <source>Learning (ICML)</source>
          ,
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>E.</given-names>
            <surname>Hesamifard</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Takabi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Ghasemi</surname>
          </string-name>
          , Cryptodl: [22]
          <string-name>
            <given-names>A.</given-names>
            <surname>Singh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Chopra</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Garza</surname>
          </string-name>
          ,
          <string-name>
            <surname>E. Zhang,</surname>
          </string-name>
          <article-title>Deep neural networks over encrypted data, arXiv</article-title>
          <string-name>
            <given-names>P.</given-names>
            <surname>Vepakomma</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Sharma</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Raskar</surname>
          </string-name>
          , Disco: Dypreprint arXiv:
          <volume>1711</volume>
          .05189 (
          <year>2017</year>
          ).
          <article-title>namic and invariant sensitive channel obfuscation</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>C.</given-names>
            <surname>Juvekar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Vaikuntanathan</surname>
          </string-name>
          ,
          <string-name>
            <surname>A.</surname>
          </string-name>
          <article-title>Chandrakasan, for deep neural networks</article-title>
          , in: Conference on Com{
          <article-title>GAZELLE}: A low latency framework for secure puter Vision and Pattern Recognition (CVPR</article-title>
          ),
          <year>2021</year>
          .
          <article-title>neural network inference</article-title>
          ,
          <source>in: USENIX Security</source>
          [23]
          <string-name>
            <given-names>P. C.</given-names>
            <surname>Roy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V. N.</given-names>
            <surname>Boddeti</surname>
          </string-name>
          ,
          <source>Mitigating information Symposium</source>
          ,
          <year>2018</year>
          .
          <article-title>leakage in image representations: A maximum en-</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>K.</given-names>
            <surname>Nandakumar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Ratha</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Pankanti</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Halevi</surname>
          </string-name>
          , To- tropy approach,
          <source>in: Proceedings of the IEEE/CVF wards deep neural network training on encrypted Conference on Computer Vision</source>
          and Pattern Recogdata, in: Conference on Computer Vision and Pat- nition,
          <year>2019</year>
          , pp.
          <fpage>2586</fpage>
          -
          <lpage>2594</lpage>
          . tern Recognition
          <string-name>
            <surname>Workshops (CVPR-W)</surname>
          </string-name>
          ,
          <year>2019</year>
          . [24]
          <string-name>
            <given-names>A.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Guo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F. D.</given-names>
            <surname>Salim</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Chen</surname>
          </string-name>
          , Deep-
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Fu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Mi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <article-title>Mixup based obfuscator: Obfuscating intermediate representaprivacy preserving mixed collaboration learning, tions with privacy-preserving adversarial learning in: International Conference on Service-Oriented on smartphones</article-title>
          ,
          <source>in: International Conference System Engineering (SOSE)</source>
          ,
          <year>2019</year>
          .
          <article-title>on Internet-of-Things Design and Implementation,</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Huang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Song</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Arora</surname>
          </string-name>
          , Instahide:
          <year>2021</year>
          .
          <article-title>Instance-hiding schemes for private distributed</article-title>
          [25]
          <string-name>
            <given-names>H.</given-names>
            <surname>Edwards</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Storkey</surname>
          </string-name>
          ,
          <article-title>Censoring representations learning</article-title>
          ,
          <source>in: International Conference on Machine with an adversary</source>
          ,
          <source>in: International Conference on Learning (ICML)</source>
          ,
          <year>2020</year>
          .
          <source>Learning Representations (ICLR)</source>
          ,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>M.</given-names>
            <surname>Shin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Hwang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Kim</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Park</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Bennis</surname>
          </string-name>
          , S.- [26]
          <string-name>
            <given-names>C.</given-names>
            <surname>Louizos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Swersky</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Welling</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Zemel</surname>
          </string-name>
          , L. Kim,
          <article-title>Xor mixup: Privacy-preserving data aug- The variational fair autoencoder (</article-title>
          <year>2016</year>
          ).
          <article-title>mentation for one-shot federated learning</article-title>
          , arXiv [27]
          <string-name>
            <given-names>D.</given-names>
            <surname>Madras</surname>
          </string-name>
          , E. Creager,
          <string-name>
            <given-names>T.</given-names>
            <surname>Pitassi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Zemel</surname>
          </string-name>
          , Learning preprint arXiv:
          <year>2006</year>
          .
          <volume>05148</volume>
          (
          <year>2020</year>
          ).
          <article-title>adversarially fair and transferable representations,</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>E.</given-names>
            <surname>Borgnia</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Geiping</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Cherepanova</surname>
          </string-name>
          , L. Fowl, in: International Conference on Machine Learning A.
          <string-name>
            <surname>Gupta</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Ghiasi</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Huang</surname>
            ,
            <given-names>M. Goldblum,</given-names>
          </string-name>
          (ICML),
          <year>2018</year>
          . T. Goldstein, Dp-instahide: Provably defusing [28]
          <string-name>
            <given-names>P.</given-names>
            <surname>Vepakomma</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Singh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Gupta</surname>
          </string-name>
          ,
          <string-name>
            <surname>R.</surname>
          </string-name>
          <article-title>Raskar, poisoning and backdoor attacks with diferen- Nopeek: Information leakage reduction to share tially private data augmentations, arXiv preprint activations in distributed deep learning</article-title>
          ,
          <source>in: 2020 InarXiv:2103</source>
          .
          <year>02079</year>
          (
          <year>2021</year>
          ). ternational Conference on Data Mining Workshops
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <surname>T</surname>
          </string-name>
          .
          <article-title>-h.</article-title>
          <string-name>
            <surname>Kim</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Kang</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          <string-name>
            <surname>Pulli</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Choi</surname>
          </string-name>
          ,
          <source>Training (ICDMW)</source>
          ,
          <year>2020</year>
          .
          <article-title>with the invisibles: Obfuscating images</article-title>
          to share [29]
          <string-name>
            <given-names>S. A.</given-names>
            <surname>Osia</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. S.</given-names>
            <surname>Shamsabadi</surname>
          </string-name>
          ,
          <string-name>
            <surname>S.</surname>
          </string-name>
          <article-title>Sajadmanesh, safely for learning visual recognition models, arXiv</article-title>
          <string-name>
            <given-names>A.</given-names>
            <surname>Taheri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Katevas</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H. R.</given-names>
            <surname>Rabiee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N. D.</given-names>
            <surname>Lane</surname>
          </string-name>
          , preprint arXiv:
          <year>1901</year>
          .
          <volume>00098</volume>
          (
          <year>2019</year>
          ). H.
          <string-name>
            <surname>Haddadi</surname>
          </string-name>
          ,
          <article-title>A hybrid deep learning architecture for</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>C.</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Ren</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Qin</surname>
          </string-name>
          ,
          <string-name>
            <surname>K.</surname>
          </string-name>
          <article-title>Ren, privacy-preserving mobile analytics</article-title>
          ,
          <source>IEEE Internet Ganobfuscator: Mitigating information leakage un- of Things Journal</source>
          (
          <year>2020</year>
          ).
          <article-title>der gan via diferential privacy</article-title>
          , Transactions on [30]
          <string-name>
            <given-names>F.</given-names>
            <surname>Pittaluga</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Koppal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Chakrabarti</surname>
          </string-name>
          ,
          <source>Learning Information Forensics and Security</source>
          (
          <year>2019</year>
          ).
          <article-title>privacy preserving encodings through adversarial</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>J.</given-names>
            <surname>Donahue</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Simonyan</surname>
          </string-name>
          ,
          <article-title>Large scale adversar- training, in: Winter Conference on Applications of ial representation learning</article-title>
          ,
          <source>Advances in Neural Computer Vision (WACV)</source>
          ,
          <year>2019</year>
          .
          <source>Information Processing Systems</source>
          <volume>32</volume>
          (
          <year>2019</year>
          ). [31]
          <string-name>
            <given-names>S.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Du</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Shrivastava</surname>
          </string-name>
          , L. Zhong, Privacy
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>L.</given-names>
            <surname>Lyu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Yu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <article-title>Threats to federated learning: adversarial network: representation learning for A survey</article-title>
          , arXiv preprint arXiv:
          <year>2003</year>
          .
          <volume>02133</volume>
          (
          <year>2020</year>
          ).
          <article-title>mobile data privacy</article-title>
          , ACM on Interactive, Mobile,
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>D.</given-names>
            <surname>Pasquini</surname>
          </string-name>
          , G. Ateniese,
          <string-name>
            <given-names>M.</given-names>
            <surname>Bernaschi</surname>
          </string-name>
          , Unleash- Wearable and Ubiquitous
          <string-name>
            <surname>Technologies</surname>
          </string-name>
          (
          <year>2019</year>
          ).
          <article-title>ing the tiger: Inference attacks on split learning</article-title>
          , [32]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Jin</surname>
          </string-name>
          , Towards privacyin:
          <source>ACM SIGSAC Conference on Computer and preserving visual recognition via adversarial trainCommunications Security</source>
          ,
          <year>2021</year>
          .
          <article-title>ing: A pilot study</article-title>
          , in: European Conference on
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>O.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Sun</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Gao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , J. Xie, Computer Vision (ECCV),
          <year>2018</year>
          . V.
          <string-name>
            <surname>Smith</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Wang</surname>
          </string-name>
          ,
          <article-title>Label leakage</article-title>
          and protec- [33]
          <string-name>
            <given-names>B.</given-names>
            <surname>Sadeghi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Yu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Boddeti</surname>
          </string-name>
          ,
          <article-title>On the global optima tion in two-party split learning, arXiv preprint of kernelized adversarial representation learning</article-title>
          ,
          <source>arXiv:2102.08504</source>
          (
          <year>2021</year>
          ). in: International Conference on Computer Vision
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>M.</given-names>
            <surname>Bertran</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Martinez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Papadaki</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Qiu</surname>
          </string-name>
          , M. Ro
          <string-name>
            <surname>-</surname>
          </string-name>
          (ICCV),
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>