=Paper=
{{Paper
|id=Vol-2283/MediaEval_18_paper_56
|storemode=property
|title=Exploring Three Views on Image Enhancement for Pixel Privacy
|pdfUrl=https://ceur-ws.org/Vol-2283/MediaEval_18_paper_56.pdf
|volume=Vol-2283
|authors=Simon Brugman,Maciej Wysokiński,Martha Larson
|dblpUrl=https://dblp.org/rec/conf/mediaeval/BrugmanWL18
}}
==Exploring Three Views on Image Enhancement for Pixel Privacy==
Exploring three views on image enhancement for Pixel Privacy Simon Brugman Maciej Wysokiński Martha Larson Radboud University, the Netherlands Universidad Complutense, Spain Radboud University, the Netherlands simon.brugman@cs.ru.nl maciwyso@ucm.es m.larson@cs.ru.nl ABSTRACT The aim of the MediaEval 2018 Pixel Privacy task is to increase image appeal while blocking automatic inference of sensitive scene information. We investigate three different views from which we could consider enhancement: the view of the image aesthetics field, the view of automatic large-scale aesthetics inference models, and the view of social media users who reflect on their own photo- (a) Original image (b) Enhanced image graphic practices. Systematic image editing can do better than one-size-fits-all-filters with helping casual social media users find Figure 1: The original image (left) is classified by ResNet50 the desired photo look. Machine learning aesthetics assessment as hotel/outdoor, the enhanced image as fire_escape. falls short when inferring individual preferences. A qualitative user study gives insight into the diversity and complexity of preferences. amount of activity, such as Instagram (more than 1B monthly ac- tive users [15]) and Flickr (the iPhone is the most used camera [7]). 1 INTRODUCTION These apps allow users to edit images internally, for example, ap- plying filters, supporting extremely fast sharing of edited images. The MediaEval Pixel Privacy task aims at protecting users from Currently, the state of the art in mobile apps for cameras is large-scale inference of sensitive information while increasing im- predefined filters, which can change in hue, saturation or lightness age appeal. As we develop Pixel Privacy technologies, we want to or add visual effects like blur or noise. Filtered photos, especially understand how to apply and assess image enhancement. In this with increased colour temperature, exposure, and contrast, are more paper, we consider three views on image enhancement. likely to be viewed (+21%) and commented on (+45%) than unfiltered • The view from the field of image aesthetics: Here we explore photos [2]. These filters have the disadvantage of depriving users of what aspects of overall colour harmony we can systematise with- editing control. Predefined filters are the same each time the filter out full understanding of the content of the image. is used and may limit the ability of users to achieve the desired • The view of the field of machine learning on automatic in- photo look. Here, we aim to discover contributions from the field ference of aesthetics: We would like to better understand the of image aesthetics that would allow us to improve the flexibility potential of this technology for aesthetics evaluation of image of photo filters to increase image aesthetics and add user appeal. enhancement in the Pixel Privacy task. The users of image sharing networks can be divided into people • The view of social media users: We survey a small group of with aesthetic knowledge and casual photographers. The former participants who have the habit of consciously reflecting on their group tends towards smooth changes, supported by manual editing, own photographic practices. This qualitative user study aims at the latter usually prefers to achieve more dramatic change [2]. Our discovering strong and weak points of our image enhancements. goal is to discover dramatic changes consistent with image content, In the following sections, we discuss each view in turn. Note that but not requiring full image understanding. Early explorations have in this work we assume an interconnection between enhancement directed our attention to colour grading and cropping for image and appeal. Consistently with [11, 14] we consider that improving enhancement. Figure 1 shows a colour transformation whose goal aesthetics also improves appeal. is to increase the appeal of the original image. The example was chosen because it is one of the promising cases where the classifier 2 SYSTEMATIC IMAGE EDITING used in the Pixel Privacy task [9] is misdirected by a transformation. We consider the field of image aesthetics in order to discover aspects What aspects of overall colour harmony can we systematise with- of photos that can be changed systematically, leading to an aes- out full image understanding? As an initial attempt, we convert thetic improvement or an increase of appeal without full knowledge the input image to HSV colour space [19], obtaining pixel values of what is being depicted in the photo. Such aspects would lend expressed in terms of the three-dimensional nature of human colour themselves well to automatization. Our interest in automatization perception [20]: (1) hue, which refers to pure colour, (2) saturation is related to the observation that automatic filters are currently in from white light to pure colour and (3) value, which refers to illumi- widespread use and assume that transformations must be fast to nation values. Assigning the hue values to the specific ranges in the match the speed of what is currently offered by apps. RGB colour wheel [13] (primary, secondary and tertiary colours) The number of amateur photographers is growing as smartphone it is possible to identify dominant values in order to carry out an usage increases [16]. Popular camera mobile apps attract a large overall harmony shift sensitive to tones, tints, and shades. In this experiment, we manipulate only hue values shifting them Copyright held by the owner/author(s). MediaEval’18, 29-31 October 2018, Sophia Antipolis, France to different ranges in the RGB colour wheel according to the near- est detected harmony: monochromatic, analogous, complementary, MediaEval’18, 29-31 October 2018, Sophia Antipolis, France Simon Brugman, Maciej Wysokiński, and Martha Larson The Pixel Privacy task could benefit from automatic assessment that treats all users equally in terms of the prediction error of their appeal judgements. 4 PERCEPTION OF IMAGE ENHANCEMENT The user study is aimed at gathering qualitative insight into aspects Figure 2: A histogram of the per image mean and standard of image enhancements important for user preference. The study deviation as calculated on ground truth and as predicted by compares three approaches: (1) systematically increasing overall NIMA, figure from [18]. colour harmony and improving composition (cf. Section 2), (2) double complementary, split complementary, triadic complemen- enhancing the images intuitively, carried out by an artist, who tary [6] (pages 22-28). This methodology is similar to the geometri- restricted the enhancements to the same sort of manipulations cal formulation of classical colour harmony by Moon-Spencer [12]. that were applied systematically in (1), and (3) the style transfer We also applied a forced crop that considers the rule of thirds. approach described in [10]. Each approach is used to generate an Note that visual perception involves both a form corresponding enhanced image from ten original images from the manual test set to structure and colours as a feature of reflected light [3] (page 20). of the 2018 Pixel Privacy task, resulting in 30 image pairs. The order Our experiment disregards form. The juxtaposition of the original of the pairs is randomised. and harmony-shifted images can intensify the sensation of artificial For each pair the original and enhanced image are randomly colours. However, colours are a response to light, and convincing assigned to be Image A and Image B. Study participants look at looking colours are not absolute but can vary. The perceptual dif- both images and then answer the question “Which image would ference is reduced by the colour constancy phenomenon [1] (page you prefer to share?” using a 5-point scale running between A and 6-9). Note that the desired colour harmony can differ for each user B. Additionally they give qualitative feedback by giving a short and also may be more or less suitable for a given original image. elaboration on their preference. The interface allows the user to toggle between image A and image B. Toggling makes the interface 3 MACHINE LEARNING AESTHETICS more closely resemble the user interfaces in existing applications Technology for large-scale aesthetic inference is widely available (e.g., Instagram) and is also intended to eliminate unwanted direct and can be used by different multimedia applications for which comparisons of the two images. We had access to a group of peo- user appeal is important, such as search engines (e.g., [11, 14]) ple with conscious knowledge about images (e.g., photography or and automated photo album management systems (e.g., [4]). We computer vision expertise), and, for this preliminary, we selected consider state-of-the-art advances in automatic image aesthetic the study participants (ten in total) from this group. The rationale assessment for evaluating appeal of image enhancement. In the is that this group would be better able to identify which of their re- task, our interest is focused on the user’s personal perspective when actions is related to image transformations (as opposed to content) sharing a picture on social media, as this is likely to lead to adoption and express their reactions in words. of the privacy-preserving image enhancements. On average, study participants preferred the original image over A recent survey on image aesthetics assessment discusses visual the enhanced image. For systematic enhancement (1), enhanced features (hand-crafted and deep features), data set characteristics images were preferred in 2 of the 10 cases, compared to 3/10 for in- and evaluation metrics [5]. Neural-network-based machine learn- tuitive enhancement (2). We identified several high-level categories ing models are able to assess image aesthetics more accurately than capturing generalisations in the reasons given by study partici- traditional approaches. They do not require explicit incorporation pants for their image preferences: colours (harmony, cold/warm), of expert knowledge of photography. There are efforts to improve composition (ratio, perspective, focus, information, framing), no on the state of the art. In [8, 11], the user ratings are extended difference, authenticity (water is not purple). Notable was that for with rater IDs, enabling user-specific models. NIMA [18] is a neu- the systematic enhancement, composition change has an effect on ral architecture for image assessment that predicts a distribution the perceived authenticity and image quality (with respect to focus). of ratings from one to ten. It improves handling of ground-truth ambiguity by optimizing the Earth Mover’s Distance on ordered 5 OUTLOOK user score distributions. In addition to the mean user rating, the In this paper, we have explored three different views on image distribution of NIMA can capture agreement of user ratings. The enhancement. Aspects from the field of image aesthetics can be loss used by NIMA can also be used for tuning image enhancement systematised for specific image enhancement, making the change methods [17] and as a metric for perceptual distance [21]. more dependent on the content of the image, while not requiring From the nature of how NIMA is learned, it tends to avoid uncer- full understanding of the image. For the Pixel Privacy task, machine tain predictions when relevant information is missing. The image learning aesthetic assessment does not treat users equally in terms alone does not provide all information of the user state that could of the error of prediction of their appeal judgements, which is a influence the rating (e.g., memories from the moment the user took potential limitation. A user study with experts gave valuable insight the photo, current mood). Figure 2 illustrates that there is a mis- into the diversity of preferences for hue and composition. match between the distributions of per-image means and standard deviations when comparing the ground truth to the predictions of ACKNOWLEDGMENTS NIMA. NIMA predicts the overall reception of an image by users This work is part of the Open Mind research programme, financed and does not attempt to predict reception of images by single users. by the Netherlands Organisation for Scientific Research (NWO). Pixel Privacy Task MediaEval’18, 29-31 October 2018, Sophia Antipolis, France REFERENCES 598–610. [1] George A Agoston. 2013. Color Theory and its Application in Art and [20] Stephen Westland, Kevin Laycock, Vien Cheung, Phil Henry, and Design. Vol. 19. Berlin, Heidelberg. Forough Mahyar. 2007. Colour Harmony. Journal of the International [2] Saeideh Bakhshi, David A Shamma, Lyndon Kennedy, and Eric Gilbert. Colour Association (JAIC) 1 (2007), 1–15. 2015. Why We Filter Our Photos and How It Impacts Engagement. [21] Richard Zhang, Phillip Isola, Alexei A. Efros, Eli Shechtman, and Oliver In Proceedings of the 9th International Conference on Web and Social Wang. 2018. The Unreasonable Effectiveness of Deep Features as a Media (ICWSM). AAAI, 12–21. Perceptual Metric. In The IEEE Conference on Computer Vision and [3] José María Cuasante, Cuevas María, and Blanca Fernández Quesada. Pattern Recognition (CVPR). 2005. Introducción al Color. Akal, D.L., Tres Cantos (Madrid). [4] Jingyu Cui, Fang Wen, Rong Xiao, Yuandong Tian, and Xiaoou Tang. 2007. EasyAlbum: An Interactive Photo Annotation System based on Face Clustering and Re-ranking. In Proceedings of the SIGCHI confer- ence on Human factors in computing systems. ACM, 367–376. [5] Yubin Deng, Chen Change Loy, and Xiaoou Tang. 2017. Image aesthetic assessment: An experimental survey. IEEE Signal Processing Magazine 34, 4 (2017), 80–106. [6] Edith Anderson Feisner and Ronald Reed. 2013. Color Studies. Fairchild Books, New York. [7] Flickr. 2018. Camera Finder. https://www.flickr.com/cameras/. (2018). Accessed: 2018-11-14. [8] Shu Kong, Xiaohui Shen, Zhe Lin, Radomir Mech, and Charless Fowlkes. 2016. Photo Aesthetics Ranking Network with Attributes and Content Adaptation. In Proceedings of the 14th European Conference on Computer Vision (ECCV). Springer, 662–679. [9] Martha Larson, Zhuoran Liu, Simon Brugman, and Zhengyu Zhao. 2018. Pixel Privacy: Increasing Image Appeal while Blocking Au- tomatic Inference of Sensitive Scene Information. In Working Notes Proceedings of the MediaEval 2018 Workshop. [10] Zhuoran Liu and Zhengyu Zhao. 2018. First Steps in Pixel Privacy: Exploring Deep Learning-based Image Enhancement against Large- scale Image Inference. In Working Notes Proceedings of the MediaEval 2018 Workshop. [11] Ning Ma, Alexey Volkov, Aleksandr Livshits, Pawel Pietrusinski, Houdong Hu, and Mark Bolin. 2018. An Universal Image Attrac- tiveness Ranking Framework. arXiv preprint arXiv:1805.00309 (2018). [12] Parry Moon and Domina Eberle Spencer. 1944. Geometric formulation of classical color harmony. Journal of the Optical Society of America (JOSA) 34, 1 (1944), 46–59. [13] José María Parramón. 1998. Teoría y Práctica del Color. Parramón ediciones, Barcelona. [14] Yale Song, Miriam Redi, Jordi Vallmitjana, and Alejandro Jaimes. 2016. To click or not to click: Automatic selection of beautiful thumbnails from videos. In Proceedings of the 25th ACM International on Conference on Information and Knowledge Management (CIKM). ACM, 659–668. [15] Statista. 2018. Number of monthly active Instagram users from Janu- ary 2013 to June 2018 (in millions). (2018). https://www.statista.com/ statistics/253577/number-of-monthly-active-instagram-users/ Ac- cessed: 2018-11-14. [16] Statista. 2018. Number of smartphone users worldwide from 2014 to 2020 (in billions). https://www.statista.com/statistics/330695/ number-of-smartphone-users-worldwide. (2018). Accessed: 2018- 11-14. [17] Hossein Talebi and Peyman Milanfar. 2018. Learned perceptual image enhancement. In 2018 IEEE International Conference Computational Photography (ICCP). IEEE, 1–13. [18] Hossein Talebi and Peyman Milanfar. 2018. NIMA: Neural image assessment. IEEE Transactions on Image Processing 27, 8 (2018), 3998– 4011. [19] A Vadivel, Shamik Sural, and Arun K Majumdar. 2005. Human color perception in the HSV space and its application in histogram genera- tion for image retrieval. In Color Imaging X: Processing, Hardcopy, and Applications, Vol. 5667. International Society for Optics and Photonics,