=Paper=
{{Paper
|id=Vol-2670/MediaEval_19_paper_57
|storemode=property
|title=Image
Enhancement and Adversarial Attack Pipeline for Scene Privacy Protection
|pdfUrl=https://ceur-ws.org/Vol-2670/MediaEval_19_paper_57.pdf
|volume=Vol-2670
|authors=Muhammad Bilal Sakha
|dblpUrl=https://dblp.org/rec/conf/mediaeval/Sakha19
}}
==Image
Enhancement and Adversarial Attack Pipeline for Scene Privacy Protection==
Image Enhancement and Adversarial Attack Pipeline for Scene Privacy Protection Muhammad Bilal Sakha1 1 Habib University, Pakistan mbilal.sakha@gmail.com ABSTRACT 2 APPROACHES In this paper, we propose approaches to prevent automatic inference 2.1 Fusion based approaches of scene class by classifiers and also enhance (or maintain) the vi- CartoonGAN style transfer and Iterative least-likely class sual appeal of images. The task is part of the Pixel Privacy challenge adversarial attack: In the first approach, we use an image style of the MediaEval 2019 workshop. The fusion based approaches we transfer method based on Generative Adversarial Networks (GANs) propose apply adversarial perturbations on the images enhanced [4] called CartoonGAN [1], which enhances the image by applying by image enhancement algorithms instead of the original images. cartoon style effects. On these enhanced set of images, we then They combine the benefits of image style transfer/contrast enhance- apply a white-box targeted adversarial attack called the Iterative ment and the white-box adversarial attack methods and have not least-likely class method [7], which is a variant of the Fast Gradient been previously used in the literature for fooling the classifier and Sign Method (FGSM) proposed by [5]. The Iterative least-likely enhancing the images at the same time. We also propose to use class method tries to make an adversarial image by adding noise simple Euclidean transformations which include image translation to the clean image, so that it will be classified as the class with the and rotation and show their efficacy in fooling the classifier. We lowest confidence score for clean image. For choosing optimal ϵ test the proposed approaches on a subset of the Places365-standard (limit on the perturbation size), instead of doing binary search on dataset and get promising results. each example because of the computational expense, we choose the value of ϵ to be 8/255 on the basis of experimental results on a 1 INTRODUCTION subset of validation set images. When enhancing the images using CartoonGAN, Hayao style is chosen because it results in the largest Social media users unintentionally expose private information increase of mean aesthetic score among different CartoonGAN when sharing photos online [12], such as locations a user visited styles on the validation images. etc., which can be automatically inferred by state of the art methods CartoonGAN style transfer and PGD: In a slightly modified [2]. The focus of Pixel Privacy task of MediaEval 2019 workshop version, we now apply Projected Gradient Descent (PGD) [11] ad- [10] is to protect user uploaded multimedia data online. The task versarial attack after enhancing the images with CartoonGAN style objective is to use image transformation algorithms for blocking transfer. Here, we apply an untargeted adversarial attack, unlike in the automatic inference of scene class by convolutional neural net- the previous method where the target class is the least-likely class work (ConvNet) based ResNet50 classifier [6] trained on Places365- of clean image. For the PGD adversarial attack, we chose the value standard dataset [15]. The proposed methods should also either of ϵ to be 2/255 and the stepsize is chosen as 1/ϵ on the basis of increase (or maintain) the visual appeal of an image. Additional empirical results on a subset of validation images. details of the task can be found in [10]. Image contrast enhancement & Iterative least-likely class: We propose to combine image style-transfer and image enhance- In this approach, we first enhance the contrast of the images using ment with adversarial image perturbations to increase the visual the method proposed by [14] and then perturb the enhanced images appeal of the images, in addition to blocking the automatic infer- using the Iterative least-likely class adversarial method [7]. The rea- ence of scene class information by the classifier. We also apply son for applying image processing to enhance the images initially white-box (where the attacker has access to the model’s parame- is because the adversarial perturbation methods, reduce the visual ters) adversarial perturbations alone to compare the performance appeal of the images, so enhancing the visual appeal of the images to the fusion based approaches. Finally, we use simple euclidean before applying adversarial perturbations will not only result in bet- operations like image translation and rotation to show how they ter performance on image quality metrics, but may also incentivize are also able to fool the classifier. The proposed approaches are users to use this method over adversarial perturbations alone. In evaluated on the basis of reduction in the top-1 classifier accuracy the image contrast enhancement approach by [14], the input image and Neural Image Assessment (NIMA) [13] score is used to evalu- is fused with the synthetic image, which is obtained by finding the ate the image quality of the transformed images. The motivation best exposure ratio to well-expose the under-exposed regions in behind proposing fusion based approaches is to incentivize the the original image. Both the images are then fused according to social media users to use such methods for not only protecting the the weight matrix, which is designed using illumination estimation privacy-sensitive information in the photos, but also to enhance techniques and the output is the contrast enhanced image. On these their photos as an added bonus. enhanced set of images, we then apply the Iterative least-likely Copyright 2019 for this paper by its authors. Use class method, with the same parameters values as mentioned in the permitted under Creative Commons License Attribution first approach. 4.0 International (CC BY 4.0). MediaEval’19, 27-29 October 2019, Sophia Antipolis, France MediaEval’19, 27-29 October 2019, Sophia Antipolis, France M. Sakha Figure 1: Original sample image from the Places356-standard dataset and the images transformed using different approaches with their corresponding top-5 classifier predictions. 2.2 White-box Private-FGSM adversarial attack Table 1: Accuracy and NIMA score of different approaches In order to compare the adversarial image perturbations with pre- vious fusion based approaches, we use a more powerful variant of Run/Method Accuracy NIMA score FGSM method, called Private-Fast Gradient Sign Method (P-FGSM) 1. CartoonGAN + least-likely class 0.167% 4.37 recently proposed by [8]. The values of ϵ and σ used for this method 2. CartoonGAN + PGD 14% 4.77 are set to 8/255 and 0.99 respectively. 3. Contrast Enh. + least-likely class 0% 4.47 4. Private-FGSM 0% 4.49 2.3 Euclidean transformations 5. Euclidean transformations 1 6.667% 4.42 Inspired from the center crop and random crop operations in [9] Original test images 100% 4.64 to fool the classifier, we choose to explore other simple geometric operations on images, which are often overlooked in favor of adver- 1 Euclidean transformations evaluated on a smaller subset of sarial attacks to fool the classifier. We consider two basic euclidean test dataset consisting of 60 images. transformations i.e. image translation and rotation. To choose the optimal translation and rotation value to fool the classifier, we use the robust optimization method proposed by [3], instead of the the top-1 classifier accuracy to 0%, at the cost of added noise in the computationally expensive grid-search. For majority of the images, submitted images, which is reflected in the reduced NIMA score of we constrain translation to be within 20% of image size in each 4.49. Private-FGSM attack and previously used Iterative least-likely spatial direction and rotation up to 20°, and fill the resulting empty class methods are bounded by l ∞ norm, which results in small image spaces with zero pixel value. noise evenly distributed in the image, as can be seen by zooming the transformed images in Figure 1. 3 RESULTS AND EVALUATION Euclidean transformations: The final run of euclidean trans- In the Pixel Privacy task of the MediaEval 2019 workshop, the formations which consists of translation and rotation operations participants are allowed to submit five runs for the task, which achieves 6.667% top-1 classifier accuracy with a reasonable NIMA are evaluated on the basis of top-1 classification accuracy (lower is score of 4.42. For each image, finding the optimal translation and better) and NIMA score [13] (higher is better), as shown in Table 1. rotation value to fool the classifier is computationally expensive Figure 1 shows the original image and the transformed images by due to number of random transformations, therefore we test this different approaches and the corresponding top-5 class prediction. approach on smaller subset of test dataset consisting of 60 images Fusion based approaches: The performance of CartoonGAN + called test_manual, provided by the task organizers. Iterative least-likely class adversarial method is good in terms of the top-1 accuracy, however it has the worst NIMA score of 4.37 4 CONCLUSION AND OUTLOOK among all runs. CartoonGAN + PGD adversarial method has the In this paper, different approaches have been proposed for the best NIMA score of 4.77 among all runs, but considerably higher Pixel Privacy task of MediaEval 2019 workshop. The fusion based classifier accuracy of 14%, which it is still less than 50%. approaches combining style transfer/image enhancement with ad- For Contrast Enhancement + Iterative least-likely class run, we versarial attacks are chosen to increase the image appeal score get the lowest 0% top-1 accuracy and 4.47 NIMA score. The images beforehand, as reducing the classifier accuracy through adversarial enhanced using contrast enhancement method [14] look visually perturbations decrease image appeal score, due to addition of noise. more appealing to the naked eye, however the NIMA score after In future, increasing image appeal by using the state of the art applying only contrast enhancement is still slightly less than that deep learning based image enhancement methods for image de- of the clean images which is unexpected. noising, color/contrast/exposure adjustment etc. and then applying Private-FGSM adversarial attack: Private-FGSM attack reduces adversarial perturbation in our opinion will yield better results. Pixel Privacy MediaEval’19, 27-29 October 2019, Sophia Antipolis, France REFERENCES [1] Yang Chen, Yu-Kun Lai, and Yong-Jin Liu. 2018. CartoonGAN: Gen- erative adversarial networks for photo cartoonization. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 9465–9474. [2] Jaeyoung Choi, Martha Larson, Xinchao Li, Kevin Li, Gerald Fried- land, and Alan Hanjalic. 2017. The geo-privacy bonus of popular photo enhancements. In Proceedings of the 2017 ACM on International Conference on Multimedia Retrieval. ACM, 84–92. [3] Logan Engstrom, Brandon Tran, Dimitris Tsipras, Ludwig Schmidt, and Aleksander Madry. 2019. Exploring the Landscape of Spatial Robustness. In International Conference on Machine Learning. 1802– 1811. [4] Ian Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde-Farley, Sherjil Ozair, Aaron Courville, and Yoshua Bengio. 2014. Generative adversarial nets. In Advances in neural information process- ing systems. 2672–2680. [5] Ian J Goodfellow, Jonathon Shlens, and Christian Szegedy. 2014. Explaining and harnessing adversarial examples. arXiv preprint arXiv:1412.6572 (2014). [6] Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. 2016. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition. 770–778. [7] Alexey Kurakin, Ian Goodfellow, and Samy Bengio. 2016. Adversarial examples in the physical world. arXiv preprint arXiv:1607.02533 (2016). [8] C. Y. Li, A. S. Shamsabadi, R. Sanchez-Matilla, R. Mazzon, and A. Cavallaro. 2019. Scene Privacy Protection. In Proc. IEEE Int. Conf. on Acoustics, Speech and Signal Processing. Brighton, UK. [9] Zhuoran Liu and Zhengyu Zhao. 2018. First Steps in Pixel Privacy: Exploring Deep Learning-based Image Enhancement against Large- Scale Image Inference.. In MediaEval. [10] Zhuoran Liu, Zhengyu Zhao, and Martha Larson. 2019. Pixel Privacy 2019: Protecting Sensitive Scene Information in Images. In Working Notes Proceedings of the MediaEval 2019 Workshop. [11] Aleksander Madry, Aleksandar Makelov, Ludwig Schmidt, Dimitris Tsipras, and Adrian Vladu. 2017. Towards deep learning models resis- tant to adversarial attacks. arXiv preprint arXiv:1706.06083 (2017). [12] Tribhuvanesh Orekondy, Bernt Schiele, and Mario Fritz. 2017. Towards a visual privacy advisor: Understanding and predicting privacy risks in images. In Proceedings of the IEEE International Conference on Computer Vision. 3686–3695. [13] Hossein Talebi and Peyman Milanfar. 2018. NIMA: Neural image assessment. IEEE Transactions on Image Processing 27, 8 (2018), 3998– 4011. [14] Zhenqiang Ying, Ge Li, Yurui Ren, Ronggang Wang, and Wenmin Wang. 2017. A new image contrast enhancement algorithm using exposure fusion framework. In International Conference on Computer Analysis of Images and Patterns. Springer, 36–46. [15] Bolei Zhou, Agata Lapedriza, Aditya Khosla, Aude Oliva, and Antonio Torralba. 2017. Places: A 10 million Image Database for Scene Recog- nition. IEEE Transactions on Pattern Analysis and Machine Intelligence (2017).