<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>The Computational Memorability of Iconic Images</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Lisa Saleh</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Nanne van Noord</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Multimedia Analytics Lab, University of Amsterdam</institution>
        </aff>
      </contrib-group>
      <fpage>55</fpage>
      <lpage>71</lpage>
      <abstract>
        <p>The perception of historic events is frequently shaped by speci昀椀c images that have been ascribed an iconic status. These images are widely reproduced and recognised and can therefore be considered memorable. A question that arises given such images is whether the memorability of iconic images is intrinsic or whether it is shaped. In this work we analyse the memorability of iconic images by means of computational techniques that are speci昀椀cally designed to measure the intrinsic memorability of images. To judge whether iconic images are inherently more memorable we establish two baselines based on datasets of diverse imagery and of newspaper imagery. Our 昀椀ndings show that iconic images are not more memorable than modern day newspaper imagery or when compared to a diverse set of everyday images. In fact, by and large many of the iconic images analysed score on the low end of the memorability spectrum. Additionally, we explore the variation in memorability of reproductions of iconic images and 昀椀nd that certain images have been edited resulting in higher memorability scores, but that the images by and large are reproduced with memorability close to the original.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Memorability</kwd>
        <kwd>Iconicity</kwd>
        <kwd>Computer Vision</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>the course of history. Due to this in昀氀uence iconic images have been widely researched with
the aim to understand what makes them unique.</p>
      <p>
        Di昀erent aspects of iconic images have been researched. One o昀琀en noted criteria for
iconicity is the importance of the captured historic event. Due to the democratisation of photography,
images of recent historic events are abundant. Thus an image that merely displays a signi昀椀cant
historic event is not su昀케cient to reach iconic status. For an image to become iconic it must be
widely spread in the media1[
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. Therefore it is useful to know what images the media tends
to publish and which characteristics they carry. For example, one study on media portrayal
of photographs of hurricane Katrina uncovered that photographs with some visual themes are
more o昀琀en published than others [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]. Besides visual themes, there are plenty of other aspects
that can in昀氀uence if an image can be iconic. To recall the de昀椀nition of an iconic photograph,
the photograph must be widely recalled and recognized. An image that is widely recalled and
recognized needs either to be really memorable or to be displayed very o昀琀en. A more
memorable image is more likely to be recalled. Thus high memorability of an image may contribute
to its iconic status. As there has been no research on the intrinsic memorability of iconic
images, this will be the focus of this paper. Speci昀椀cally, we investigate whether there are general
patterns concerning memorability of iconic images that are salient even when considering the
unique circumstances by which these images have become iconic.
      </p>
      <p>
        Memorability of images is a complex characteristic to predict; Using only the surface features
of the image for predicting memorability is not comprehensive. Automating the prediction
of memorability with Convolutional Neural Networks (CNN) has been a successful solution
to this problem 1[
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. It is proven that features such as object size and brightness do a昀ect
image memorability 1[0]. However, a CNN like MemNet outperforms these surface features
in predicting memorability1[
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. When considering image memorability what a method like
MemNet measures is the intrinsic memorability of an image, which can be interpreted as the
innate probability of an image to be remembered. It does not guarantee that an image will be
remembered, and it is certainly not the only factor, but all other factors being equal an image
with a higher intrinsic memorability is more likely to be remembered.
      </p>
      <p>Besides the advantages of automatizing the process by using a CNN, the usage of a
prediction CNN has another advantage: As iconic photographs are images that are widely spread it
is likely that a great share of people has already seen said image. Therefore it would be
di昀케cult to test the memorability of these images with a memory task. It would be a challenge to
guarantee that these people have never seen these images before. Thus the challenge of
testing the memorability of iconic photographs context-independent can be solved by using the
above-mentioned MemNet. Additionally, the iconicity of some images is determined by the
fame of the subject. MemNet is not trained to recognize celebrities or other symbols. In this
research, the focus is on what image aspects the memorability in昀氀uenced, without analysis of
the context of the images. This makes the usage of MemNet a great 昀椀t. In this paper, iconic
photographs are tested on their memorability by the use of the CNN MemNet to explore to
what extent the memorability of iconic photographs can be measured.</p>
      <p>To investigate this we formulate the following two research questions:
RQ 1: How does the memorability of iconic images compare to a baseline?
RQ 2: To what extent does the memorability of iconic images di昀er across variations?</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related work</title>
      <sec id="sec-2-1">
        <title>2.1. Iconic Images</title>
        <p>
          There has been little to no computational research on iconic images, but in the following we
will give a brief overview of other research directions. As iconic images are not a bound set,
di昀erent de昀椀nitions and what criteria they should meet have been proposed [
          <xref ref-type="bibr" rid="ref11 ref17 ref2 ref7">7, 17, 2, 11</xref>
          ]. There
are di昀erences in the criteria brought up, but the following six proposed by Perlmut2te2r] a[re
used most o昀琀en: ”(1) signi昀椀cance of the reported event; (2) capacity to represent the event
as a whole; (3) celebrity of the image promoted by the media; (4) prominence of display of
the image; (5) frequent repetition of the image across media outlets; (6) ability to generate a
primordial theme in society such as good versus evil.” These criteria are not met to the same
extent for every iconic image; it is not a blueprint for iconic images. For example, both the
image of the Falling Man of the Twin Towers and the hijacked plane crashing into the Twin
Towers represent the September 11 attacks. The image of the Falling Man has a less capacity
to represent the event as a whole than the image of the airplane, as the attack itself is displayed
more comprehensively in the latter image. But even though the Falling Man image does not
meet the second criteria to the full extent, it is still an image to be considered iconic.
        </p>
        <p>
          Additionally, other studies have added criteria to be more comprehensive. For example, in
some studies more stress has been placed on the symbolic meaning that people have of the
image [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ]. An image can become a symbol with a meaning beyond the historical image it is
attached to. The Guerrillero Heroico image of Che Guevara is used as a symbol for revolution
beyond in Cuba [
          <xref ref-type="bibr" rid="ref20">20</xref>
          ]. Even though this is not signi昀椀cant for every iconic image, it is still a
unique aspect for some. Furthermore, the criteria from Perlmut2te2r] p[ut the focus on the
image and less on the reception of the image. One aspect of the reception that is unmentioned
in these six criteria is that an iconic image also should be widely recalled and part of the
collective memory of a certain with a group’s identity7][. This is another criterion o昀琀en added
to complement the criteria from Perlmutter. That an image must be widely recalled is
important to add because this can be a way of researching what image have an iconic status with
qualitative research as done by for example Hoeven12[].
        </p>
        <p>
          But as other factors of the reception, like the symbolic meaning or being a part of the
collective memory, of an iconic image are not fully encoded in the image itself and are subjective,
it is yet challenging to study computational19[]. It is yet di昀케cult to study the semantics of
images computationally, as this o昀琀en is can not be conducted from the visual data or metadata.
Therefore the study of the perception of iconic images is o昀琀en limited to interviews or
largescale surveys [
          <xref ref-type="bibr" rid="ref19">19</xref>
          ]. Hence the study of iconic images in computer vision is yet to be immersed,
and potentially the computational study of memorability can open this door.
        </p>
      </sec>
      <sec id="sec-2-2">
        <title>2.2. Memorability</title>
        <p>
          Due to technological advancements capturing and sharing have become more convenient.
Mass media and its consumption lead people to be exposed to countless images every day.
People are exposed to many images daily and some of those are better remembered than
others. Which image people remember and which they forget is related to the memorability of
the image. Di昀erent academic 昀椀elds have an interest in image memorability. For example,
research in cognitive sciences answered how certain activities in the brain correlate to the
memorability of images. Additionally, image memorability has been a subject of research in
the computer vision 昀椀eld in recent years 1[
          <xref ref-type="bibr" rid="ref10 ref14 ref4 ref6 ref9">6, 10, 9, 14, 4</xref>
          ]. Taking the next step from
computing how memorable images are with MemNet, to analyzing what image attributes leads to
memorable image and what connection memorability has with other qualities like the emotion
it portrays. In line with researching what attributes in昀氀uence memorability, GANalyze was
created [
          <xref ref-type="bibr" rid="ref10">10</xref>
          ]. With GANalyze images are tweaked to increase memorability. Some attributes
emerged to o昀琀en be increased when optimizing the memorability, such as redness, brightness
and object size [
          <xref ref-type="bibr" rid="ref10">10</xref>
          ]. Thus images that have higher redness, brightness or a bigger object size
are more likely to be memorable. Even though these attributes might be simple and intuitive
ways to predict memorability, the prediction of memorability is complex and can not solely be
explained by these attributes 1[
          <xref ref-type="bibr" rid="ref6">6</xref>
          ].
        </p>
        <p>
          When diving further into the correlation between memorability and some image aspects,
collected across di昀erent studies, the following observations were made. In some studies, it
appeared that a number of object categories, like people, animals and vehicles were relatively
more memorable 2[
          <xref ref-type="bibr" rid="ref1">1</xref>
          ]. Images with humans are especially memorable when the face is
visible and has eye contact with the camera 1[
          <xref ref-type="bibr" rid="ref3">3</xref>
          ]. In scenic images, indoor scenes appeared to be
relatively more memorable13[]. Additionally, it appeared that spatial layout is highly
correlated with memorability; Images where the object is bigger and centred in the image are more
likely to be memorable2[
          <xref ref-type="bibr" rid="ref1">1</xref>
          ]. This is in line with research which showed that images with a
high aesthetic level are also more likely to be memorab1l5e];[the spatial layout of an image is
strongly connected to its aesthetics. Moreover, it also appeared that images in which humans
portray negative emotions like disgust and fear, tend to be recalled bett1e6r].[ Finally, only
the most memorable images have a correlation with populari1t6y][. Even though mapping for
these above attributes and many more, there is still about 25% of the variance in memorability
that is unaccounted for 2[
          <xref ref-type="bibr" rid="ref3">3</xref>
          ]. Thus looking for the attributes in images that are correlated with
memorability will give some intuitive sense of memorability, but will not be comprehensive.
        </p>
      </sec>
      <sec id="sec-2-3">
        <title>2.3. MemNet &amp; LaMem</title>
        <p>In this paper, memorability will be measured computationally through a score given by
MemNet. MemNet is a CNN trained on an annotated dataset called LaMem. LaMem is a large-scale
dataset constructed of multiple already existing datasets that were annotated in a perception
study with human subjects who performed a memorability task. The datasets used are highly
diverse, which leads to LaMem being a diverse dataset totalling 60.000 image1s6][. It contains
images of humans, animals, landscapes, and even art, which also includes abstract images. This
diversity of image types is displayed in Figur1e.</p>
        <p>
          For the LaMem annotation task a stream of images was shown to participants and a昀琀er a
varying number of distractors the participants had to indicate if they had seen the image before.
Thus when an image was remembered o昀琀en the image would be scored as more memorable.
A昀琀er performing this task on a group of participants a memorability score for each image
could be calculated. This score is the annotation for each image in the LaMem dataset. This
makes LaMem thereby a su昀케cient foundation for a CNN to compute a memorability score
(0 − 1 scalar) for di昀erent types of images [
          <xref ref-type="bibr" rid="ref16">16</xref>
          ]. LaMem was shown to have an inter-annotator
rank correlation o0f.68, with MemNet achieving a remarkably high rank correlation 0o.6f4
[16, p. 6]. MemNet has since been further validated in subsequent studies and is, therefore, a
reliable measure of intrinsic image memorabili2ty3,[
          <xref ref-type="bibr" rid="ref14 ref21 ref4">21, 14, 4</xref>
          ].
        </p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Memorability of Iconic Images</title>
      <p>The following step-by-step procedure, shown in Figu2rerepresents our proposed method that
can be used for measuring a memorability across a set of images.</p>
      <p>The procedure we propose involves a comparative process between two datasets (A and B).
Here, dataset A represents the collection to be studied, whereas dataset B functions as a baseline.
The general idea of this approach is to not only measure the memorability scores but also
judge whether the scores di昀er from a meaningful baseline. Each dataset is pre-processed by
transforming the images to a suitable input resolution for the CNN (typically a low resolution
like256 × 256) and normalizing them. Once the dataset are pre-processed the images will be
passed through the pre-trained CNN to obtain a per-image memorability score.</p>
      <p>The method for comparing between the memorability scores of datasets A and B depends
on the sizes of the respective datasets. In the case of small datasets, the analysis will be mostly
done by qualitatively looking at the images to draw conclusions based on the speci昀椀c images.
When dealing with large-scale datasets, the comparison can be based on a statistical analysis.
Whilst our baseline datasets are large-scale, the dataset of iconic images used is too small scale
to reliably perform statistical analysis.</p>
      <p>Further analysis beyond the comparative analysis will consist of sampling images which are
either random or statically interesting, according to their attribute score. Sampling random
images will give an indicative perspective on how the data is constructed. This method of
analysis is based on Distant Viewing3[], a method for studying large visual corpora, where the
main take way is to perform the computational analysis whilst also viewing the corpora. The
latter is import because the intuitive semantics that may get lost in a computational analysis
will thereby also be studied. As the semantics are important for iconic images, this method
suits this paper well. Conclusions drawn from these analyses can then be linked to prior work.</p>
      <sec id="sec-3-1">
        <title>3.1. Comparing the Memorability of Iconic Images</title>
        <p>
          For our dataset of iconic images we use a selection of 26 images that were initially in a
largescale survey on iconic images 1[
          <xref ref-type="bibr" rid="ref2">2</xref>
          ]. Smits and Ros [
          <xref ref-type="bibr" rid="ref25">25</xref>
          ] further used this set and collected
variations of the images in the form of online circulations. These 26 images are a part of a
global visual memory: ’a limited set of images that people all over the world have seen and
remembered’ [
          <xref ref-type="bibr" rid="ref12">12</xref>
          ]. Each image and its scores are depicted in Figur3e. These images are in
order from highest to lowest memorability score. When comparing the scores with the images
we make the following observations. Firstly, it is notable is that the three images with the
highest memorability score are similar. The images of Che Guevara, Sharbat Gula, and the
Migrant Mother are all portraits where the face of the subject 昀椀lls much of the image. Their
expression is visible to the viewer as all three are looking in the (rough) direction of the camera.
That these portraits have a high memorability score is in line with the prior work; Images with
bigger object size, a face that has eye contact with the camera and humans as their main object
are more likely to have a higher memorability score. Which makes it in the lines of expectation
that images like the Holocaust survivors, Raising a 昀氀ag over the Reichstag and Tank Man have
a lower memorability score; These images have a small object size and either has no human
faces or a high amount of di昀erent faces. On the whole we see a diverse range of memorability
scores across the dataset.
        </p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Iconic Images compared to a Baseline</title>
        <p>
          To place the memorability scores in context, we compare to two di昀erent baselines. The 昀椀rst
baseline used is the LaMem dataset; This dataset consists of images with di昀erent subjects
and themes, and functions as a diverse baseline representing a ‘general image collectio1n6]’.[
This is an annotated dataset for memorability, but the annotations were not used; Images of
the LaMem dataset were inserted in MemNet to create a memorability score generated with
the same circumstances as the dataset it is compared with. The second baseline used was the
GoodNews dataset, which consists of all kinds of images used in the New York Times from 1818
until 2019 [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ]. These images vary from news photography to sports photography, to images
from the cooking appendix and so on. The range of types of images is displayed in Figu4r. e
These images are mainly images that were made by professional photographers and selected
by the editors of the paper. Thus these images o昀琀en have some journalistic value and meet
the aesthetic standards of the paper. Therefore the GoodNews dataset is used as a baseline to
represent images depicted in media. As iconic images are widely published in the media, the
GoodNews dataset will help us determine whether iconic images are particularly memorable
Sharbat Gula
        </p>
        <p>0.9120
Survivor
of Hutu
death
camp
0.8290</p>
        <p>Abu
Ghraib
prisoner
0.8098
Mao Zedong</p>
        <p>0.7767
Times Square Kiss</p>
        <p>0.7610
Assassination of
Inejiro Anasuma
0.7248
Burning</p>
        <p>Monk
0.7476
Spanish
Soldier
0.6982
Raising a
Flag over</p>
        <p>the
Reichstag
0.6796</p>
        <p>Che Guevera</p>
        <p>0.8894
The Falling</p>
        <p>Man
0.8190
Hinden</p>
        <p>burg
Disaster
0.7940
Mohandas</p>
        <p>Ghandi
0.7725
Napalm Girl</p>
        <p>0.7588
Vultre and
the girl
0.7462
Raising the</p>
        <p>flag on
Iwo Jima
0.7173
Coup in</p>
        <p>Chile
0.6981
Holocaust
survivors
0.6449</p>
        <p>Migrant Mother</p>
        <p>0.8767
Man on the Moon</p>
        <p>0.8168
Alan Kurdi</p>
        <p>0.7897
Hijacked Airplane</p>
        <p>0.7697
Viet Cong
Execution</p>
        <p>0.7533
Situation</p>
        <p>Room
0.7447
Kent State
Shootings</p>
        <p>0.7151
Tank Man
0.6939
when compared to other media images.</p>
        <p>The distribution of the memorability scores of the LaMem and GoodNews dataset is
visualised in Figure5. Additionally, the scores of the iconic images are plotted with a scatter plot
at the bottom of the Figure. From the distributions we can observe that the GoodNews images
are generally more memorable than the LaMem images. Additionally, the LaMem dataset has
a wider distribution than the GoodNews dataset. There are fewer images in the GoodNews
dataset with a very low memorability score, which might re昀氀ect that all images in GoodNews
have been approved by an editor of the paper. It is very likely an editor would select a
picture with at least some traits that correlate with memorability. For example, images that are
selected by an editor o昀琀en need to have either a clear main object, thus big object size or some
aesthetic value. Taking into account scatter plot of the iconic images we observe that iconic
images are generally (slightly) less memorably than the images depicted in the media.</p>
        <p>To further clarify how the iconic images compare to both datasets we plot the distribution
of the iconic images according to the quartiles of both LaMem and GoodNews in Figu6r.eWe
can observe that most iconic images fall into the lower quartiles for both datasets, scoring
slightly higher when compared to LaMem. As the LaMem dataset is representative of all kinds
of images, it can be concluded that these iconic images are slightly less memorable images.
When comparing the iconic images to the GoodNews dataset, it stands out that most of the
iconic images have a memorability score that overlaps with the 昀椀rst quantile of the GoodNews
dataset. The distribution of the iconic images in the quartiles of the GoodNews dataset is
signi昀椀cantly shi昀琀ed to the less memorable side. Thus iconic images are less memorable than
images depicted in the media.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Memorability across Variations</title>
      <p>Until now we have analysed the most canonical versions of the 26 iconic images. However, it
is known that edits have been made in prominently published versions of iconic images. To
explore whether these edits have been made (implicitly or explicitly) to boost the memorability
of iconic images we further analyse the circulations collected by Smits and R2o5s].[For each
of the 26 iconic images they used the Google Cloud Vision API to retrieve online circulations,
thereby creating a dataset of 900k images. Per image the number of variations di昀ers from the
Che Guevara portrait having over 100k variations retrieved to the image of Mao Zedong and
the founding of the PRC, having less than 3k variations. Figur7e shows how many images
were retrieved for each of the images.</p>
      <p>Due to how the variations were collected, in iterations based on previous retrieval results, the
di昀erences to the original get progressively larger the deeper we get into the list of variations.
The variations of the iconic image vary in di昀erent crops, di昀erent shades of colors and so
on. There are also variations where other images and text are added to the iconic image, this
can di昀er from a logo of a broadcaster to a book cover where the original image also appears.
Furthermore, internet memes, collages and other types of photo-shopped images appear in the
dataset. Additionally, the API is not infallible, there are some images in the dataset where the
original image does not appear at all. Roughly speaking we observe that images which have
been circulated less there are also fewer edits. To control for some of this variation, and to limit
the scale of the comparison, we only use the 昀椀rst 10k variations for each iconic image.</p>
      <p>The memorability scores for the variations of each iconic image are visualised in Fig7u. re
Images to the right on the x-axis should generally have less resemblance with the original
image. When viewing the distributions of scores for the variations some patterns can be found.
We recognise four groups: (1) distributions with relatively little spread across the variations,
(2) distributions that fan outward towards the end of the plot, (3) distributions that have a large
spread from start to 昀椀nish, and (4) distributions where we can recognise clear clusters.</p>
      <p>Examples of the 昀椀rst group, with little spread, at the Migrant Mother, the Situation Room, and
the assassination of Anasuma. Within this group we can recognise two subcategories, as they
are either images that have less than 10k variations or they are images where the canonical
form is most dominant. Because for the latter subcategory we have only looked at the 昀椀rst
10k images retrieved, it is still very likely that there are also many variations that di昀er more
strongly circulating on the internet. But these would only be retrieved further in the dataset
than the 昀椀rst 10k images selected. However, it is still noticeable that the 昀椀rst 10k variations
have more resemblance with the original iconic image than for other widely circulated iconic
images.</p>
      <p>For the second group examples have most of their spread at the end of the graph, such as
the Burning Monk, the Falling man and the Tank Man images. In these images, it appears that
there is de昀椀nitely a big share of images that have a big resemblance with the original picture
but variations that di昀er more already appear within the 昀椀rst 10k variations.</p>
      <p>In the third group there is a lot of spread from beginning to 昀椀nish, these are images like
Gandhi, Raising the 昀氀ag over Iwo Jima and the Image of the Spanish Soldier. These images
appear in a lot of di昀erent variations. This can be explained by it being popular photos for
web pages, book covers and other types of editing where the context of the image gets altered
heavily.</p>
      <p>Lastly is the group where we can recognise di昀erent clusters. This group includes images
like Napalm Girl, the Hijacked Plane and the portrait of Sharbat Gula. All these images have an
alternative popular variation, which leads to a cluster forming in the 昀椀gure. In the following
section, the Napalm Girl image will be highlighted to give an example of what these variations
look like. Two variations of the Sharbat Gula image and the Hijacked plane are displayed
in Figure8 with their memorability score in the caption of each image. The variations that
are highlighted in this 昀椀gure are average images from the main clusters. In the Sharbat Gula
variations graph, the bigger cluster on the top of the graph is the original image (Fig8uar)eand
the cluster that lies under most of the images are depicted in Figu8rbe. In the Hijacked Plane,
there is a similar pattern, the graph of the variations has two distinct clusters, one at the top
which consists of images like the original image (Figur8ec and the other cluster which has a
lower memorability score than the 昀椀rst cluster. This cluster consists of images like in Fig8udr,e
which has a wider crop than the original image such that the building on the le昀琀 still remains
in the image.
(a) The original image with a memorability score
of 0.9120
(b) The variation of the original with a recent
image placed alongside has a memorability score
of 0.8511
(c) The published image with a memorability (d) A variation of the published image with a
score of 0.7697 wider crop and a memorability score of 0.7095</p>
      <sec id="sec-4-1">
        <title>4.1. The Memorability of the Variations of the Napalm Girl Image</title>
        <p>A number of di昀erent crops of the Napalm girl image circulate online. The scatter plot which
displays the memorability scores and their variations is depicted in Fig9u.rTehe outliers on the
upper side of this plot, images with a score above 0.85, are either images heavily photoshopped
or images that got in this dataset but in which the Napalm Girl image does not appear. One
of the aspects that stands out from this graph is the big cluster of images which are under the
trend line and appear from about 4.000 on the x-axis. When sampling these images it appears
that these are mainly the Napalm Girl image with a tighter crop than the original image. This
is displayed in Figure10. Where the original image is displayed in Figu1r0ea and the tighter
crop in Figure10b. The part of the image that is only visible in the wider crop consists of a big
part of the sky and a photographer on the right. The tight crop is more focused on the children,
which are the main subject of the picture, and thereby take up a larger portion of the image.
The tighter crop having a higher memorability score 昀椀ts the notion that images with a bigger
object size are more memorable and hints that memorability might be an (implicit) criteria for
edits done by photo-editors.</p>
        <p>Another frequent variation was the image in Figur1e0c. This image has a signi昀椀cantly
higher score than the original picture and was the only variation with a score in this range
that still fully depicted the original image. The original image was edited with a red-hued 昀椀lter
over the image. That this red-hued image has such a higher memorability score is expected as
the redness of an image has been demonstrated to positively a昀ect memorability.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Discussion</title>
      <p>The 昀椀ndings in this paper are based on the result generated by MemNet. Even though MemNet
is a validated method of predicting memorability, the actual memorability and the computed
(a) The original full photograph with a memora- (b) Widely published tighter crop with a
memobility of 0.6767 rability of 0.7652</p>
      <p>(c) A variation with a red 昀椀lter and a memorability of 0.8200
memorability score could still di昀er. In conducted research, it became apparent that image
memorability and some qualities were correlated, like aesthetics and certain emotions. But
even though they are correlated, MemNet is not trained on these qualities; Thus, for example,
images that do not have the more common features that positively in昀氀uence the memorability,
like a big object size, but are very memorable because they portray a strong negative emotion
that also in昀氀uences memorability, possibly do not get a high memorability score from MemNet.
This highlights a mismatch between intrinsic memorability and actually being remembered. A
clear example of this mismatch is the Napalm Girl image, this image has a memorability score
below the average of LaMem. But Napalm Girl displays a scene of terror, that evokes strong
negative emotions like sadness and anger. Those strong emotions make this image more
memorable and highlight a limitation of (computational) intrinsic memorability methods. Moreover,
frequent exposure of less memorable images might also lead to increased remembrance, which
might also play a role for this image.</p>
      <p>A possible limitation of this work is that the dataset was selected by Dutch researchers.
Despite being selected with the aim to represent international iconic images, the images are
predominantly known in the Western World. This could in昀氀uence the results, but this might
also interact with MemNet. As MemNet is trained on LaMem which consists mainly of images
from the Western World, this bias is is matched - whilst it should not in昀氀uence the analysis itself
it does limit to what extent we can generalize about the results. Additionally, the dataset of the
iconic images is mostly from the 20th century when color photography was not as common.
Most of the 26 images are in black and white. When experimenting with the di昀erences in
memorability score for the same image in greyscale as in color, it a昀ected the memorability
score; While the LaMem dataset averaged a memorability score 0o.7f645 this score dropped
to 0.7456 when all images were converted to greyscale. From this we can observe that colour
plays a role, but not to the extent that is changes our conclusions. Even with taking these
points into account, the observations we made align with existing theories on memorability.</p>
    </sec>
    <sec id="sec-6">
      <title>6. Conclusion</title>
      <p>In answering the research question: How does the memorability of iconic images compare to
a baseline? It appeared that of the iconic images, the portraits where the face was a big part
of the image had a higher memorability score. This con昀椀rmed the research on memorability
where a correlation between higher memorability and big object size, and a face as an image
object was established. Additionally, the iconic images with small object sizes and small faces
had a lower memorability score. The memorability scores for these iconic images align with
theories on memorability. Furthermore, it appeared that images that are depicted in the media
are generally more memorable than all other images. Few images are depicted in media that
have a relatively low memorability score since all images depicted in media are selected by
editors. Iconic images are generally slightly less memorable than other images and are on the
lower side of the memorability of images in the media.</p>
      <p>To answer the second research question: To what extent does the memorability iconic images
di昀er across variations? When comparing the spread of the memorability of the variations of
the iconic images, there were certain patterns to be found. Some had clear clusters, which were
other popular variations of the original image. Looking at examples of di昀erent variations and
memorability scores they show that altering an image can in昀氀uence the memorability score. In
these examples, it appeared that variations of the images with a red hue or tighter crop were
more memorable. This is in line with previous work.</p>
      <p>On the whole we can conclude that computational measures of memorability do not fully
capture the memorability of iconic images, as many iconic images are remembered much better
than what their memorability score would imply. While the reasons for this may be manifold
we expect that frequent exposure and strong emotional content play an important role.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>R.</given-names>
            <surname>Adler-Nissen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K. E.</given-names>
            <surname>Andersen</surname>
          </string-name>
          , and
          <string-name>
            <given-names>L.</given-names>
            <surname>Hansen</surname>
          </string-name>
          . “Images, emotions, and
          <article-title>international politics: the death of Alan Kurdi”</article-title>
          .
          <source>InR:eview of International Studies 46.1</source>
          (
          <issue>2020</issue>
          ), pp.
          <fpage>75</fpage>
          -
          <lpage>95</lpage>
          . doi:
          <volume>10</volume>
          .1017/s0260210519000317.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>K.</given-names>
            <surname>Andén-Papadopoulos</surname>
          </string-name>
          .
          <article-title>“The Abu Ghraib torture photographs: News frames, visual culture, and the power of images”</article-title>
          .
          <source>InJ:ournalism 9</source>
          .1 (
          <issue>2008</issue>
          ), pp.
          <fpage>5</fpage>
          -
          <lpage>30</lpage>
          . doi:
          <volume>10</volume>
          .1177/1464 884907084337.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>T.</given-names>
            <surname>Arnold</surname>
          </string-name>
          and
          <string-name>
            <given-names>L.</given-names>
            <surname>Tilton</surname>
          </string-name>
          . “
          <article-title>Distant viewing: analyzing large visual corpora”D.Iingi:tal Scholarship in the Humanities 34</article-title>
          .Supplement_
          <volume>1</volume>
          (
          <issue>2019</issue>
          ), pp.
          <fpage>i3</fpage>
          -
          <lpage>i16</lpage>
          . doi:
          <volume>10</volume>
          .1093/llc/fqz01 3.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Baveye</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Cohendet</surname>
          </string-name>
          ,
          <string-name>
            <surname>M.</surname>
          </string-name>
          <article-title>Perreira Da Silva, and</article-title>
          <string-name>
            <given-names>P. Le</given-names>
            <surname>Callet</surname>
          </string-name>
          . “
          <article-title>Deep Learning for Image Memorability Prediction: The Emotional Bias”</article-title>
          .
          <source>IPnr:oceedings of the 24th ACM International Conference on Multimedia. Mm '16</source>
          . Amsterdam, The Netherlands: Association for Computing Machinery,
          <year>2016</year>
          , pp.
          <fpage>491</fpage>
          -
          <lpage>495</lpage>
          . doi:
          <volume>10</volume>
          .1145/2964284.2967269. url: https://d oi.
          <source>org/10</source>
          .1145/2964284.2967269.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>A. F.</given-names>
            <surname>Biten</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Gómez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Rusiñol</surname>
          </string-name>
          , and
          <string-name>
            <given-names>D.</given-names>
            <surname>Karatzas</surname>
          </string-name>
          . “Good News,
          <article-title>Everyone! Context driven entity-aware captioning for news images”</article-title>
          . InC:oRR abs/
          <year>1904</year>
          .01475 (
          <year>2019</year>
          ). arXiv:
          <year>1904</year>
          .01475.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <surname>P. Bruegel</surname>
          </string-name>
          <article-title>the ElderT</article-title>
          .
          <source>riumph of Death</source>
          .
          <volume>1563</volume>
          . url: https://www.museodelprado.es/en/th e-collection/
          <article-title>art-work/the-triumph-of-death/</article-title>
          <source>d3d82b0b-9bf2-4082-ab04-66ed53196</source>
          .ccc
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>A. A.</given-names>
            <surname>Cohen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Boudana</surname>
          </string-name>
          , and
          <string-name>
            <given-names>P.</given-names>
            <surname>Frosh</surname>
          </string-name>
          . “
          <article-title>You Must Remember This: Iconic News Photographs and Collective Memory”</article-title>
          .
          <source>InJ:ournal of Communication 68.3</source>
          (
          <issue>2018</issue>
          ), pp.
          <fpage>453</fpage>
          -
          <lpage>479</lpage>
          . doi:
          <volume>10</volume>
          .1093/joc/jqy017.
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>N. S.</given-names>
            <surname>Dahmen</surname>
          </string-name>
          and
          <string-name>
            <given-names>A.</given-names>
            <surname>Miller</surname>
          </string-name>
          . “
          <article-title>Rede昀椀ning iconicity: A 昀椀ve-year study of visual themes of Hurricane Katrina”</article-title>
          .
          <source>InV:isual Communication Quarterly 19.1</source>
          (
          <issue>2012</issue>
          ), pp.
          <fpage>4</fpage>
          -
          <lpage>19</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>R.</given-names>
            <surname>Dubey</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Peterson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Khosla</surname>
          </string-name>
          ,
          <string-name>
            <surname>M.-H. Yang</surname>
            , and
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Ghanem</surname>
          </string-name>
          . “
          <source>What Makes an Object Memorable?” In:2015 IEEE International Conference on Computer Vision</source>
          (ICCV).
          <year>2015</year>
          , pp.
          <fpage>1089</fpage>
          -
          <lpage>1097</lpage>
          . doi:
          <volume>10</volume>
          .1109/iccv.
          <year>2015</year>
          .
          <volume>130</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>L.</given-names>
            <surname>Goetschalckx</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Andonian</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Oliva</surname>
          </string-name>
          , and
          <string-name>
            <given-names>P.</given-names>
            <surname>Isola</surname>
          </string-name>
          . “
          <article-title>GANalyze: Toward Visual De昀椀nitions of Cognitive Image Properties”</article-title>
          .
          <source>InP: roceedings of the IEEE/CVF International Conference on Computer Vision</source>
          (ICCV).
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>R.</given-names>
            <surname>Hariman</surname>
          </string-name>
          and
          <string-name>
            <given-names>J. L.</given-names>
            <surname>Lucaites</surname>
          </string-name>
          .
          <article-title>No caption needed: Iconic photographs, public culture, and liberal democracy</article-title>
          . University of Chicago Press,
          <year>2007</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <surname>R. van der Hoeven.</surname>
          </string-name>
          “
          <article-title>The Global Visual Memory: A Study of the Recognition and Interpretation of Iconic and Historical Photographs”</article-title>
          .
          <source>PhD thesis</source>
          .
          <source>Universiteit Utrecht</source>
          ,
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>P.</given-names>
            <surname>Isola</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Parikh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Torralba</surname>
          </string-name>
          ,
          <article-title>and</article-title>
          <string-name>
            <given-names>A.</given-names>
            <surname>Oliva</surname>
          </string-name>
          . “
          <article-title>Understanding the Intrinsic Memorability of Images”</article-title>
          .
          <source>In: Advances in Neural Information Processing Systems</source>
          . Ed. by J.
          <string-name>
            <surname>Shawe-Taylor</surname>
            , R. Zemel,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Bartlett</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Pereira</surname>
            , and
            <given-names>K.</given-names>
          </string-name>
          <string-name>
            <surname>Weinberger</surname>
          </string-name>
          . Vol.
          <volume>24</volume>
          . Curran Associates, Inc.,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>P.</given-names>
            <surname>Jing</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Su</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Nie</surname>
          </string-name>
          , and
          <string-name>
            <given-names>H.</given-names>
            <surname>Gu</surname>
          </string-name>
          . “
          <article-title>Predicting Image Memorability Through Adaptive Transfer Learning From External Sources”</article-title>
          .
          <source>IInE:EE Transactions on Multimedia 19.5</source>
          (
          <issue>2017</issue>
          ), pp.
          <fpage>1050</fpage>
          -
          <lpage>1062</lpage>
          . doi:
          <volume>10</volume>
          .1109/tmm.
          <year>2016</year>
          .
          <volume>2644866</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>P.</given-names>
            <surname>Jing</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Su</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Nie</surname>
          </string-name>
          , and
          <string-name>
            <given-names>H.</given-names>
            <surname>Gu</surname>
          </string-name>
          . “
          <article-title>Predicting Image Memorability Through Adaptive Transfer Learning From External Sources”</article-title>
          .
          <source>IInE:EE Transactions on Multimedia 19.5</source>
          (
          <issue>2017</issue>
          ), pp.
          <fpage>1050</fpage>
          -
          <lpage>1062</lpage>
          . doi:
          <volume>10</volume>
          .1109/tmm.
          <year>2016</year>
          .
          <volume>2644866</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>A.</given-names>
            <surname>Khosla</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. S.</given-names>
            <surname>Raju</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Torralba</surname>
          </string-name>
          ,
          <article-title>and</article-title>
          <string-name>
            <given-names>A.</given-names>
            <surname>Oliva</surname>
          </string-name>
          . “
          <article-title>Understanding and Predicting Image Memorability at a Large Scale”</article-title>
          .
          <source>In2:015 IEEE International Conference on Computer Vision</source>
          (ICCV).
          <year>2015</year>
          , pp.
          <fpage>2390</fpage>
          -
          <lpage>2398</lpage>
          . doi:
          <volume>10</volume>
          .1109/iccv.
          <year>2015</year>
          .
          <volume>275</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>M.</given-names>
            <surname>Mortensen</surname>
          </string-name>
          . “Constructing, con昀椀rming, and
          <article-title>contesting icons: the Alan Kurdi imagery appropriated by #humanitywashedashore, Ai Weiwei, and Charlie Hebdo”</article-title>
          .
          <source>IMne:dia, Culture &amp; Society 39.8</source>
          (
          <issue>2017</issue>
          ), pp.
          <fpage>1142</fpage>
          -
          <lpage>1161</lpage>
          . doi:
          <volume>10</volume>
          .1177/0163443717725572.
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>M.</given-names>
            <surname>Moss</surname>
          </string-name>
          .
          <article-title>Toward the visualization of history: the past as image</article-title>
          .
          <source>Lexington Books</source>
          ,
          <year>2008</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <surname>N. van Noord. “</surname>
          </string-name>
          <article-title>A survey of computational methods for iconic image analysis”.DInig:ital Scholarship in the Humanities (</article-title>
          <year>2022</year>
          ). doi:
          <volume>10</volume>
          .1093/llc/fqac003.
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>M.</given-names>
            <surname>Parker</surname>
          </string-name>
          . “
          <source>The Retrospective Iconicity of 'Guerrillero Heroico'”S.aInlf:ord Postgraduate Annual Research Conference</source>
          .
          <year>2009</year>
          , p.
          <fpage>292</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>S.</given-names>
            <surname>Perera</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Tal</surname>
          </string-name>
          , and L.
          <string-name>
            <surname>Zelnik-Manor</surname>
          </string-name>
          .
          <article-title>“Is Image Memorability Prediction Solved?</article-title>
          ” In: 2019 IEEE/CVF Conference on
          <article-title>Computer Vision and Pattern Recognition Workshops (CVPRW)</article-title>
          . Los Alamitos, CA, USA: IEEE Computer Society,
          <year>2019</year>
          , pp.
          <fpage>800</fpage>
          -
          <lpage>808</lpage>
          .
          <year>doi1</year>
          :
          <fpage>0</fpage>
          .1109/cvprw.
          <year>2019</year>
          .
          <volume>00108</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>D. D.</given-names>
            <surname>Perlmutter</surname>
          </string-name>
          .
          <article-title>Photojournalism and foreign policy : icons of outrage in international crises</article-title>
          .
          <source>Praeger series in political communication. Westport, CO [etc: Praeger</source>
          ,
          <year>1998</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <given-names>N. C.</given-names>
            <surname>Rust</surname>
          </string-name>
          and
          <string-name>
            <given-names>V.</given-names>
            <surname>Mehrpour</surname>
          </string-name>
          . “
          <article-title>Understanding Image Memorability”</article-title>
          .
          <source>ITnr:ends in Cognitive Sciences 24.7</source>
          (
          <issue>2020</issue>
          ), pp.
          <fpage>557</fpage>
          -
          <lpage>568</lpage>
          . doi: https://doi.org/10.1016/j.tics.
          <year>2020</year>
          .
          <volume>04</volume>
          .001.
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [24]
          <article-title>Sarcophagus with the Battle of Marathon. 490 Bc</article-title>
          . url: https://www.livius.org/pictures/it aly/brescia-brixia/marathon-reli.ef/
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          [25]
          <string-name>
            <given-names>T.</given-names>
            <surname>Smits</surname>
          </string-name>
          and
          <string-name>
            <given-names>R.</given-names>
            <surname>Ros</surname>
          </string-name>
          . “
          <article-title>Quantifying Iconicity in 940K Online Circulations of 26 Iconic Photographs”</article-title>
          .
          <source>In:Computational Humanities Research</source>
          .
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>