=Paper=
{{Paper
|id=Vol-2670/MediaEval_19_paper_9
|storemode=property
|title=Pixel
Privacy 2019: Protecting Sensitive Scene Information in Images
|pdfUrl=https://ceur-ws.org/Vol-2670/MediaEval_19_paper_9.pdf
|volume=Vol-2670
|authors=Zhuoran Liu,Zhengyu Zhao,Martha Larson
|dblpUrl=https://dblp.org/rec/conf/mediaeval/Liu0L19
}}
==Pixel
Privacy 2019: Protecting Sensitive Scene Information in Images==
Pixel Privacy 2019: Protecting Sensitive Scene Information in Images Zhuoran Liu, Zhengyu Zhao, Martha Larson Radboud University, Netherlands {z.liu,z.zhao,m.larson}@cs.ru.nl ABSTRACT Pixel Privacy task focuses on the protection of user-uploaded multi- media data online. Specifically, it benchmarks image transformation algorithms that protect privacy-sensitive images against automatic inference. The image transformations should block automatic classi- fiers that infer sensitive scene categories and increase (or maintain) image visual appeal at the same time. The task in 2019 is to develop image transformations under the condition that all information of the attack model is available for transformation development. Under this white-box setting, the decreased accuracy of the attack Figure 1: Examples of validation images in MediaEval Pixel model and the visual appeal of the protected images are considered Privacy task 2019. Images are randomly selected from cate- for protection evaluation. gory bedroom (top row) and closet (bottom row). 1 INTRODUCTION The Pixel Privacy task was introduced as a brave new task in the MediaEval Multimedia Evaluation Benchmark in 2018 [8]. The task The MediaEval Pixel Privacy task aims to promote the development focused on sensitive scene categories of social images, and required of algorithms that protect the privacy-sensitive information of user- participants to protect images against an automatic scene classifier. generated multimedia data online. To achieve this goal, participants Examples from the 2019 validation set are shown in Figure 1. In are encouraged to develop image transformation algorithms that 2019, we again focus on the protection of sensitive scene categories increase (or maintain) the visual appeal of images, while at the and use the same basic task formulation and source data. The task same time protecting privacy-sensitive information in the images. has been refined in several ways in order to allow us to gain more Ideally, Ideally, users should find that the transformed images are insight from the results. In 2019, we retain the whitebox setting, interchangeable with the original image, for whatever purpose the meaning that the attacking classifier is known and all informa- original image was intended. The transformed images should be tion of the attack model is available for protection development. able to mislead automatic scene classifiers. Also, we retain the untargeted setting, meaning that there is no The task is motivated by the potential risk of the privacy-sensitive particular target class into which the image must be misclassified. information implicit in user-generated data, which is accumulated Instead, any misclassification counts as protection. The important by large social networks. Accumulated social media data can be change for this year is that the test set only contains images that misappropriated for commercial purposes that are not transparent the attacking classifier classifies correctly. For the purposes of eval- to users [9]. Although algorithms [11, 13] have been developed uation, we find the images that the attacking classifier misclassifies to improve the the situation of privacy protection in multimedia to be less interesting because they can be considered to already be online, users themselves still do not have many choices to control protected. Also, this year, we pay closer attention to the pipeline. the information implicit in their own multimedia data. In addition, Specifically, images are downsized before being fed into the attack given the large amount of accumulated data, potential privacy risks classifier. Participants are required to protect the images in this could be aggravated by massive data breaches [9]. Privacy-sensitive downsized format, to control for the impact of the downsizing on information can be processed by automatic algorithms, allowing the protection. malicious actors to select potential victims as the target for specific To achieve the goal of privacy protection and visual appeal im- crimes, a practice known as cybercasing [4]. For example, based on provement, researchers participating in the task may consider re- user-uploaded images, the trajectory of an individual can be cal- lated work on different multimedia technologies. Image enhance- culated by geo-location prediction algorithms based on computer ment and style transfer [6] techniques can be exploited to increase vision algorithms. This information can be exploited by a criminal visual appeal and protect privacy. Early work showed the basic to plan a burglary by only accessing the visual contents of these ability of standard Instagram filters to block the inference of loca- social photos. Combining mined information from different sources tion information [3]. Last year, one participant paper [2] pursued a is also likely to aggravate online crimes, e.g., telecommunication color harmony based enhancement approach, which focused on im- fraud [1] or blackmail. proving visual appeal and another investigated style transfer [10]. Copyright 2019 for this paper by its authors. Use Image aesthetics assessment [14] and image quality assessment permitted under Creative Commons License Attribution methods [5] could be helpful to control the visual quality of trans- 4.0 International (CC BY 4.0). formed images. Knowledge of adversarial examples in machine MediaEval’19, 27-29 October 2019, Sophia Antipolis, France MediaEval’19, 27-29 October 2019, Sophia Antipolis, France M. Larson et al. learning can also be applied for privacy protection purpose [10, 11]. For some creative image transformation ideas, it may not be However, in 2018, participants did not fully exploit the whitebox feasible to develop fully automatic transformation algorithms. To information, which we hope they will do in 2019. leverage participants’ creativity and explore unexpected new ways in improving the visual appeal, we also provide a special test set 2 TASK DEFINITION AND DATA (MEPP19test_manual). It is a subset of test set and contains one image for each category. Manual image transformations can be ap- As stated above, the Pixel Privacy task 2019 focuses on the pro- plied on this special test set, and these images can also be submitted tection of privacy-sensitive scene category information of social for evaluation. images. A scene category can be understood to be the identity of the setting in which a photo was taken. Participants are asked to develop protection approaches on validation set to decrease the 3 EVALUATION attack accuracy while increasing image appeal. Afterwards, these Participants submit the transformed test set for evaluation and each developed approaches can be applied on test set images, and the team can maximally submit five runs. Submitted images will be protected images are submitted for evaluation. evaluated with respect to protection and appeal. The performance of The task provides 60 privacy-sensitive categories chosen from transformation approaches with respect to protection is evaluated the original 365 scene categories from Places365-Standard dataset [15], by measuring the drop of prediction accuracy of the attack model. which were original introduced in 2018. The task data set is a subset Once the prediction accuracy has reached a certain level of protec- of this dataset. The Places365-Standard dataset contains 1,803,460 tion, performance of transformations with respect to appeal will training images and 365 scene categories. The number of images be carried out with an automatic aesthetics assessment algorithm. per category varies from 3,068 to 5,000. The attack algorithm is To this end, the automatic algorithm NIMA [14] trained on the trained to detect all 365 categories. The attack classifier in the task AVA [12] dataset will be used for visual appeal evaluation, as was is a PyTorch ResNet501 [7] classifier trained on the training set of also done in 2018. This evaluation method aligns with practical the Places365-Standard dataset, as was also used in 2018. needs from users for multimedia protective technologies. A validation set (MEPP18val) is provided to allow participants In order to gain further insight in the appeal of images, we will to develop their image transformation algorithms. Figure 1 shows perform further manual assessment on cases in which the NIMA examples of image from the validation set. We also provide a test scores for different protection algorithms diverge dramatically. We set (MEPP19test) to evaluate the performance of the transforma- will select the images that have the highest variance of NIMA tion algorithms. MEPP18val contains 3000 images (50 from each scores across runs submitted by all participating teams, and have of the 60 classes), while MEPP19test contains 600 images (around these images inspected by a small panel of computer vision experts. 10 from each of the 60 classes). Note that if the original images The experts will choose the best and the worst examples from the without modification can already block the attack model, no protec- pool of all protected versions. These examples will be qualitatively tion transformation is needed. Further, images which the original analyzed in order to gain further insight into the relative strengths version is incorrectly classified by the classifier, may be correctly and weaknesses of the different protection algorithms. classified after transformation. To be able to measure protection performance without interference from these effects, MEPP19test is a subset of last year’s test set (MEPP18test), and contains only 4 DISCUSSION AND OUTLOOK the images that were correctly classified by the attack classifier. One question remains is that whether changing the label of the Pixel Privacy task 2019 is a simplified version of social image image from the proper one to an arbitrary one is enough to help privacy protection, and in particular, uses an untargeted white-box users hide their privacy-sensitive information? From Figure 1, we protection setting. Here, we provide some more details about what can imagine that if the label of an images is changed from bedroom this means. The white-box setting is that all information of the to closet, the criminal may still be able to mine the information that attack model is available for image transformation development, this image is took from home. In this case, protection by changing which means the exact neural network architecture, pre-trained the ground truth label to an arbitrary one is not enough. Another weights and related preprocessing details are available to partici- question is that in practical cases model information is not available, pants. Untargeted setting defines no target categories for the pro- which means that the white-box setup may not be valid for image tected images. In other words, if the predicted label is different from protection in real life. the ground truth then the protection is successful. Pixel Privacy task is a highly simplified task that defines how to Preprocessing the transformed images may have strong influ- protect users’ multimedia data online in a user-controlled manner. ences on the protection performance evaluation. For this reason, in In practice, the social multimedia data may have different types, the task setting, no resizing and cropping are be applied in the pro- e.g., text, video and speech data, and the threat models can be cessing step. Normalization is the only preprocessing step carried complicated too. The goal of the task is to provide a foundation upon out during evaluation. Small images (256*256) of Places365-Standard which solutions addressing progressively more realistic versions of dataset are used as standard input, and they can be downloaded the problem may be developed in the future. directly from the official website of places data set2 . ACKNOWLEDGMENTS 1 http://places2.csail.mit.edu/models_places365/resnet50_places365.pth.tar This work is part of the Open Mind research program, financed by 2 http://places2.csail.mit.edu/download.html the Netherlands Organization for Scientific Research (NWO). Pixel Privacy MediaEval’19, 27-29 October 2019, Sophia Antipolis, France REFERENCES [1] 2017. Scammers still up to their tricks despite local efforts to stop them, China Daily, 21 July. (2017). http://www.chinadaily.com.cn/opinion/ 2017-07/21/content_30195232.htm, Online; accessed 8-Aug-2019. [2] Simon Brugman, Maciej Wysokinski, and Martha Larson. 2018. Media- Eval 2018 Pixel Privacy Task: Views on image enhancement. In Work- ing Notes Proceedings of the MediaEval 2018 Workshop. [3] Jaeyoung Choi, Martha Larson, Xinchao Li, Kevin Li, Gerald Friedland, and Alan Hanjalic. 2017. The Geo-Privacy Bonus of Popular Photo En- hancements. In ACM International Conference on Multimedia Retrieval (ICMR). ACM, 84–92. [4] Gerald Friedland and Robin Sommer. 2010. Cybercasing the Joint: On the Privacy Implications of Geo-tagging. In Proceedings of the 5th USENIX Conference on Hot Topics in Security (HotSec). [5] Leonardo Galteri, Lorenzo Seidenari, Marco Bertini, and Alberto Del Bimbo. 2017. Deep generative adversarial compression artifact removal. In The IEEE Conference on Computer Vision and Pattern Recog- nition (CVPR). 4826–4835. [6] Leon A. Gatys, Alexander S. Ecker, and Matthias Bethge. 2016. Image Style Transfer Using Convolutional Neural Networks. In The IEEE Conference on Computer Vision and Pattern Recognition (CVPR). 2414– 2423. [7] Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. 2016. Deep residual learning for image recognition. In IEEE Conference on Com- puter Vision and Pattern Recognition (CVPR). 770–778. [8] Martha Larson, Zhuoran Liu, Simon Brugman, and Zhengyu Zhao. 2018. Pixel Privacy: Increasing Image Appeal while Blocking Au- tomatic Inference of Sensitive Scene Information. In Working Notes Proceedings of the MediaEval 2018 Workshop. [9] Sam Levin. 2017. Facebook Told Advertisers It Can Identify Teens Feeling ’Insecure’ and ’Worthless’, The Guardian, 1 May. (2017). https://www.theguardian.com/technology/2017/may/01/ facebook-advertising-data-insecure-teens, Online; accessed 12-Jul- 2019. [10] Zhuoran Liu and Zhengyu Zhao. 2018. First Steps in Pixel Privacy: Exploring Deep Learning-based Image Enhancement against Large- scale Image Inference. In Working Notes Proceedings of the MediaEval 2018 Workshop. [11] Zhuoran Liu, Zhengyu Zhao, and Martha Larson. 2019. Who’s Afraid of Adversarial Queries? The Impact of Image Modifications on Content-based Image Retrieval. In ACM International Conference on Multimedia Retrieval (ICMR). ACM, 306–314. [12] Naila Murray, Luca Marchesotti, and Florent Perronnin. 2012. AVA: A large-scale database for aesthetic visual analysis. In Computer Vision and Pattern Recognition (CVPR). IEEE, 2408–2415. [13] Eleftherios Spyromitros-Xioufis, Symeon Papadopoulos, Adrian Popescu, and Yiannis Kompatsiaris. 2016. Personalized Privacy-aware Image Classification. In ACM International Conference on Multimedia Retrieval (ICMR). ACM, 71–78. [14] Hossein Talebi and Peyman Milanfar. 2018. Nima: Neural image assessment. IEEE Transactions on Image Processing 27, 8 (2018), 3998– 4011. [15] Bolei Zhou, Agata Lapedriza, Aditya Khosla, Aude Oliva, and Antonio Torralba. 2017. Places: A 10 million image database for scene recogni- tion. IEEE transactions on pattern analysis and machine intelligence 40, 6 (2017), 1452–1464.