=Paper= {{Paper |id=Vol-3834/paper28 |storemode=property |title=Automated Image Color Mapping for a Historic Photographic Collection |pdfUrl=https://ceur-ws.org/Vol-3834/paper28.pdf |volume=Vol-3834 |authors=Taylor Arnold,Lauren Tilton |dblpUrl=https://dblp.org/rec/conf/chr/ArnoldT24 }} ==Automated Image Color Mapping for a Historic Photographic Collection== https://ceur-ws.org/Vol-3834/paper28.pdf
                                Automated Image Color Mapping for a Historic
                                Photographic Collection
                                Taylor Arnold1,† , Lauren Tilton2,∗,†
                                1
                                    Data Science & Linguistics, University of Richmond, U.S.A
                                2
                                    Rhetoric & Communication Studies, University of Richmond, U.S.A


                                              Abstract
                                              In the 1970s, the United States Environmental Protection Agency sponsored Documerica, a large-scale
                                              photography initiative to document environmental subjects nation-wide. While over 15,000 digitized
                                              public-domain photographs from the collection are available online, most of the images were scanned
                                              from damaged copies of the original prints. We present and evaluate a modified histogram matching
                                              technique based on the underlying chemistry of the prints for correcting the damaged images by using
                                              training data collected from a small set of undamaged prints. The entire set of color-adjusted Documer-
                                              ica images is made available in an open repository.

                                              Keywords
                                              computer vision, color analysis, histogram matching, documentary photography




                                1. Introduction
                                Many of the most important environmental laws and federal agencies in the United States came
                                into existence during the large-scale political and social environmental movement that formed
                                during the 1960s and 1970s [8]. Significant political advances that continue into the present
                                day include the Clean Air Act (1963), the Clean Water Act (1972), and the Resource Conserva-
                                tion and Recovery Act (1976). The United States Environmental Protection Agency (EPA) was
                                founded in 1970, to manage, advocate, and set standards for these new legal frameworks [21].
                                The EPA continues today with a staff of over 16,000 and a budget of over $12 million [3].
                                   Documerica was an EPA-funded project running from 1972 to 1977 that aimed to “photo-
                                graphically document subjects of environmental concern in America during the 1970s” [1]. The
                                photographs captured a wide range of topics. Mixed through images of water pollution, chem-
                                ical spills, and factory smoke plumes are pastoral landscapes from national parks around the
                                country [7]. Urban images of junkyards and trash are juxtaposed with images of cleaner tech-
                                nologies, mass transit, and Americans of all ages at play [22]. Over 15,000 photographic prints
                                from the collection have been digitized and made available through the National Archives of
                                the United States. As a product of the federal government, these images are in the public do-

                                CHR 2024: Computational Humanities Research Conference, December 4 – 6, 2024, Aarhus University, Denmark
                                ∗
                                  Corresponding author.
                                †
                                  These authors contributed equally.
                                £ tarnold2@richmond.edu (T. Arnold); ltilton@richmond.edu (L. Tilton)
                                ç https://statsmaths.github.io (T. Arnold); https://laurentilton.com/ (L. Tilton)
                                ȉ 0000-0003-0576-0669 (T. Arnold); 0000-0003-4629-8888 (L. Tilton)
                                            © 2024 Copyright for this paper by its authors. Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0).




                                                                                                               37
CEUR
                  ceur-ws.org
Workshop      ISSN 1613-0073
Proceedings
main and serve as a rich potential source of documentary evidence of the United States in the
1970s and the early years of the modern environmental movement.
   Unfortunately, a damaged set of Documerica prints was used for digitization. The scanned
photographic prints have an intense red shift, causing everything from the sky, to lakes, to
trees to have a red/orange hue in place of their normal expected colors. The sharpness of
the images has not been significantly affected indicating that the damage was due to a slow
degradation over time through a combination of heat and light. The color shift is sufÏciently
strong to reduce the aesthetic and rhetorical power of the images that are available online [12].
Interestingly, the National Archives holds two additional sets of Documerica prints. These two
other sets reveal the aesthetic qualities of the original photographs and open the door to the
possibility of correcting the damaged colors in digital prints.
   In this article, we present and evaluate an algorithm to automatically correct the color of the
damaged digitized Documerica images using a small set of the undamaged prints as training
data. Our goal is to restore the aesthetic qualities of the images rather than the impossible task
of perfectly matching the colors in the reference images [4, 17]. Through several quantitative
and qualitative analyses, we show that our transformed images more closely represent the
expected colors of the photographed scenes. Our technique is sufÏciently tractable that it could
be applied to other color-shifted collections that do not have a reference set to train against.


2. Data
The photographers employed by the Documerica project took photographs using the Ko-
dachrome color reversal film produced by Eastman Kodak [7]. Unlike many other technologies
for color photography, Kodachrome film used a subtractive technology that recorded light by
measuring the amount of cyan, magenta, and yellow dye needed to print a reconstruction of
the image [19]. A special setup was required to create prints from the images. Special labs were
able to take film and turn it into color photographic slides through three differently colored
developers [6].
   The National Archives of the United States has three copies of prints from Documerica: (1)
archival copies held in cold storage and available on special request, (2) a non-circulating set
held in their print room, and (3) a damaged set of prints that were used for the digitization
process.1 The authors visited the National Archives and selected a set of 23 slides from across
the collection to manually re-scan from both the cold-storage and non-circulating print set.
We selected images that contained a variety of different colors. After determining that the cold
storage images were indistinguishable from the print room copies, we decided to only re-scan
the latter. We then manually cropped each of these images to match the digitized prints. One
selected image had already been replaced with a corrected scan online. This left a training set
of 22 images for which we had manual reference scans from the print room and the damaged,
digitized copies on National Archives website.

1
    It is not entirely certain how the latter set was damaged or why it was selected for digitization. The third set was
    previously used as a circulation copy, and was likely damaged due to its continued circulation. Personal commu-
    nications with the current archivists suggest that because the digitization was done off-site, someone selected the
    circulating copy as to send off before realizing how much it had become damaged.




                                                           38
Figure 1: Set of selected photographs from the Documerica collection. The first column contains dig-
itized damaged prints available online; the second column shows manually scanned from the undam-
aged prints available at the U.S. National Archives; the final column shows the automatically corrected
images using our method.




                                                 39
3. Overview of Color Mapping Algorithms
Color adjustment of digital photographs is a well-studied topic and an essential element of
the workflow of modern photographers. For example, a common technique for professional
photographers is first to photograph a color reference card containing a variety of boxes with
known hues. Then, commercial software can use this reference image to learn how to adjust
any other images taken with the same equipment in similar lighting conditions [23, 24]. Adobe
Photoshop provides automated algorithms for matching color across images, which is useful
for applications such as putting together images from a photoshoot done under various lighting
conditions [9]. Color adjustment is also an essential stylistic element for photographers [16].
Commercial photo editing software such as Adobe Lightroom provides a variety of methods
for automatically applying a suite of filters to a set of images—a technique commonly used, for
example, by wedding photographers to create a distinct style—or for manually adjusting the
color of an individual image.
   In addition to commercial software for color adjustment, there has also been considerable
academic research on the specific task of matching the color between two images, a process
known as color mapping or image color transfer. In one of the earliest studies of color map-
ping, Reinhard et al. introduced a technique consisting of a simple standardization of color
intensities in a specific color space [20]. They showed that this technique worked well as
both a mild correction to standardize images of the sun setting over the horizon and more
drastic applications of style transfer from an oil painting to a modern photograph. In a more
recent survey, Faridul et al. provide a nomenclature of available techniques for color trans-
fer: geometry-based, statistical, and user-assisted [10]. Statistical techniques extend the ideas
of standardizing the mean and variance of color intensities to more involved transformations
known as histogram matching [18]. Much of the sophistication of novel methods has focused
on making differential transformations of various types over various parts of the image [13].
One motivation for localized changes is adjusting lighting conditions as a pre-processing step
for other algorithms.


4. CMY Median Histogram Matching (MHM)
Our approach to image color mapping with the Documerica collection is closely related to
the well-known histogram matching technique. In this technique we attempt to match the
distribution of colors in one image with the distribution of colors in a reference image through
a monotonic transformation of the pixel intensities [18, 14]. Two special considerations in our
application require some minor changes to the standard histogram matching algorithm.
   First of all, through our understanding of the materiality of the prints, it is likely that the
color shift that damaged the prints likely affected the images along the dimensions of the three
developer colors. The contributing elements of heat, light, and/or humidity would affect each
of these dyes differently according to their chemical composition. Our qualitative analysis of
the consistent red shift in the digitized images further indicates the need for a color correction
applied individually to each of the CMY color channels. So, in contrast to the standard strategy
for histogram matching that uses color spaces adapted to the sensitivity of the human eye, we




                                               40
will apply our directly to the RGB/CMY pixel intensities [25].2
      The second difference from a standard histogram matching algorithm is that we want to
avoid overfitting to the distribution of a single image. In fact, one of the goals of our transfor-
mation is to restore the diversity of lighting and colors that are in the source materials. Instead
of perfectly matching the distribution for a single image, we want to create a transformation
by averaging the histogram transformations across the entire training set of 22 pairs of images.
      The goal of our adapted technique, which we will refer to as median histogram matching
(MHM), is to learn three monotonic functions 𝑓𝐶̂ , 𝑓𝑀̂ , and 𝑓𝑌̂ that each map an input color
intensity (from 0 to 1) into an output color intensity on the same scale. Applying these to the
cyan, magenta, and yellow components of a damaged image should result in a transformed
image that more closely represents the undamaged form of the print. In addition to being
motivated by the chemistry of the prints, the transformation mirrors tools available in popular
photo editing software such as the Photos application of macOS and Adobe Photoshop.
      Differences in the technology available for digitizing the undamaged prints—including the
lighting, orientation, overscan size, and print-specific artifacts—made it infeasible to line up
our reference images pixel-by-pixel. Following related work on color correction, we focused
on lining up the general distribution of each color channel intensity. For each image 𝑖 and color
channel 𝑗, we computed the percentiles for the color intensities of both the damaged image and
the scanned undamaged print. Then, we constructed a function 𝑓𝑗𝑖̂ by matching the two sets of
percentiles to one another. For example, if the 20th percentile of the damaged image had a cyan
intensity of 0.6 and the undamaged percentile was 0.82, 𝑓𝑗𝑖̂ (0.6) would be set to 0.82. We then set
𝑓𝑗𝑖̂ (0) to 0 and 𝑓𝑗𝑖̂ (1) to 1 and filled in the intermediate values through linear interpolation. Then,
the final predictions for each color channel 𝑗 (𝑓𝑗̂ ) are given by taking the median across all of
the images in our training collection. These transformations are guaranteed to be monotonic
and to be well-defined for all input values.


5. Results
5.1. Transformations
Figure 2 shows the learned MTM transformations derived from our training set and the method
outlined in the previous section. The grey lines show the estimates 𝑓 𝑖̂ for each image 𝑖, with
the darker colored lines showing the final learned transformations based on the median across
the entire training set. These transformations were also applied in the final column of Figure 1.
   The cyan color channel transformation is the most different from an identity mapping. The
transformation suggests that the amount of cyan intensity should be increased, which aligns
with our qualitative analysis of the collection has having a general red tint (the intensity of
cyan can be computed by taking the inverse of the red pixel intensity). The transformation
indicates that the cyan dye as degraded at a rate faster than the other two dyes. The magenta
and yellow curves, on the other hand, are much closer to the identity transformation. While the


2
    The representation of a pixel in CMY space can be computed by simply subtracting the RGB representation from
    1. So, matching the CMY distribution is equivalent to matching the RGB representation.




                                                        41
                                       cyan                                   magenta                                   yellow

                  1.00




                  0.75
PRINT ROOM SCAN




                  0.50




                  0.25




                  0.00

                         0.00   0.25   0.50   0.75   1.00   0.00   0.25         0.50        0.75   1.00   0.00   0.25   0.50     0.75   1.00
                                                                          DIGITIZED IMAGE




Figure 2: Learned MHM transformations of the CMY color channels. Grey lines show the estimates
of individual images; the colored lines are the medians across the entire training set.


algorithm does indicate some changes for individual images, these are relatively minor when
we take the median across the entire training set.
   We see that there are considerable differences across training images. These differences are
less pronounced than they first appear because over half of all the pixel intensities in the dig-
itized images have intensities over 0.8. Looking at the upper part of each curve, we see that
these are much more stable than the rest of the transformation. Additionally, the transforma-
tions on the bottom of the curves correspond to parts of the image that have a small amount
of the given color dye. Often these are very bright parts of the image that are close to white,
and the differences indicated on the chart have a small visual effect, as we see in the following
subsection.
   The differences across training images highlight two points of caution. First, the adjustments
are nowhere near perfect. The final exact colors can still be quite different than the original
prints, the negatives themselves, or the light reflected by the objects being photographed. Sec-
ondly, in order to find a good transformation for another collection. Taking the median across
several hand-selected transformations will likely work well, whereas applying a filter trained
on only one image to a larger collection can quickly lead to overfitting.

5.2. Leave-One-Out Cross-Validation
We can use quantitative measurements to see how well the learned median transformations
work on the training data itself. In order to do this, we use a leave-one-out cross-validation
technique in which we compute the median transformations with the 𝑖th image removed—
which we denote by 𝑓𝑗−𝑖 ̂ (𝑥)—and then compare this to the transformation given by using only
the 𝑖th image itself [11]. To compare these two functions, we use a distance metric 𝑑 given by




                                                                            42
Table 1
Average errors (sum of the 𝐿2 norms across color channels, multiplied by 100) for predicting the color
channel transformation on the training set of images. Standard errors are given after the means. Uni-
form errors are computed equally across the color intensities. Weighted errors use the empirical distri-
bution for each image.
                                                      Uniform Error        Weighted Error
                     Identity Transformation           3.375 ±0.400         3.021 ±0.301
                    Leave-One-Out Estimator            1.154 ±0.227         0.990 ±0.217


Table 2
Average distance between a set of manually transformed images by a Wikipedia editor and the trans-
formed images through our median estimator. Distances are the average Euclidean distances per pixel.
The UV and AB color spaces use the same coordinates as the CIELUV and CIELAB color spaces, but
compute distances without the perceptual lightness component.
            Color Space                    CIELUV               UV          CIELAB          AB
            Identity Transformation      25.78 ±2.04         23.62 ±2.08   1239 ±115   1097 ±126
            Median Estimator             16.77 ±1.20         12.95 ±1.59    602 ±49     471 ±56


the sum of the 𝐿2 norms from each of the three color channels:
                                                         1                    2
                            𝑑 (𝑓 𝑖̂ ; 𝑓 −𝑖
                                        ̂ )=     ∑          𝑖̂       ̂
                                                                     −𝑖
                                                       ∫ |𝑓𝑗 (𝑥) − 𝑓𝑗 (𝑥)| 𝑑𝑥
                                               𝑗∈{c,m,y} 0

To get have a sense of the scale of this metric, we will compare it with using the identity
transformation to estimate 𝑓 𝑖̂ . This metric treats all input color intensities equally. We also
compute a weighted version of the metric weighted by the density of the colors in the image
itself to provide a more accurate measurement of how much the two transformations differ on
a specific image.
   The results of the leave-one-out cross-validation are shown in Table 1. The cross-validated
median estimator has an error rate about three times smaller than the error from using the
input images. The errors are slightly smaller when weighted by the empirical distribution of
the intensities, but the relative relationships remain similar. Only 1 of the 22 training images
has a cross-validation error larger for the median estimator than the identity transformation,
with every other image being improved by a factor of at least 50%.

5.3. Comparison to Manual Edits
One of the motivations for the work in this paper was our perceived ability to make the dig-
itized, damaged Documerica prints look much closer to what we expected through manual
color adjustments in commercial photo editing tools. One editor on the English version of
Wikipedia, posting under the username Hohum, had a similar experience with the Documer-
ica photographs. In January 2023, they uploaded manually corrected versions of 14 Documerica
photographs to Wikipedia. No description of the method used was given. It seems unlikely that




                                                       43
the user had access to the undamaged prints; they have uploaded thousands of other corrected
images under a large variety of categories on Wikimedia Commons. While we should not treat
the edits by Hohum as a gold-standard transformation that we should aim to replicate per-
fectly, comparing our method to the manually edited photographs offers further quantitative
evidence for the efÏcacy of our automated technique.
   As we did in comparing the transformations in the leave-one-out cross-validation, we will
look at the average distances between the reference image (here, the corrected version uploaded
to Wikimedia Commons) to both the uncorrected image and our automatically corrected im-
age. Here, though, we look at the average Euclidean distance at the level of individual pixels.
Because we are interested in measuring the visual perception of the transformation, we will
first convert the images into two different color spaces designed to represent human visual per-
ceptions: CIELAB and the CIELUV [15]. We report the results with and without considering
the lightness dimension of the colors.
   Table 2 compares our transformation to the independent, manual transformations performed
by Wikipedia user Hohum. Across all four selected color spaces, the median transformation
method is roughly twice as close to the manual adjustment than the original image. For all the
photos, the distance from the median transformation is closer to the manual transformation
in the CIELAB and AB color spaces. Only one image, which consists almost entirely of a
field of monochromatic corn shot from above, has a worse CIELUV distance in the median
transformation.

5.4. Qualitative comparison
By looking at a subset of the transformations applied outside of our training set, we can quali-
tatively describe how well the algorithm produces more realistic colors in the images. Creating
more realistic images that are more aesthetically interesting and engaging is the ultimate goal
of our work. Figure 3 shows 18 Documerica images and result of applying our transformation.
The red shift in the original images is very noticeable in the first and third columns. The trans-
formed images lose this red tone and become more color-appropriate. The image of the two
people fishing and the group of people on a bus both show no elements that still have any red
hues. On the other hand, images such as the sunset and the rocky cliff with a bridge still retain
their expected red characteristics. In other words, we have not overcorrected the images by
making them all too blue. Instead, the updated collection shows a diverse set of colors that
help to re-establish the true scale and scope of the Documerica project itself.


6. Conclusion and Future Work
The Documerica project produced a large set of historically important documentary pho-
tographs [5]. With the increased present-day attention to environmental issues, the collection
has the potential to have a visible role in helping to understand the longer history of environ-
mentalism in the United States. However, the unrealistic red shift of the digitized collection has
until now reduced its aesthetic and rhetorical appeal and thus its general reach. Through the
adjustment of this shift through the algorithmic color transfer outlined in this paper and the




                                               44
Figure 3: Set of selected photographs from the Documerica collection, with the digitized damaged
prints available online to the left of the automatically corrected images using our method. Images are
scaled to have the same aspect ratio to make them easier to display.




                                                 45
publication of the corrected images, we hope to help rectify this situation.3 As a next step—or,
more accurately, the original motivation for this work—we plan to build an interactive digital
public interface to help explore and understand the Documerica collection across its various
spatial, temporal, and visual components [2].


References
    [1] U. S. N. Archives. DOCUMERICA: The Environmental Protection Agency’s Program to Pho-
        tographically Document Subjects of Environmental Concern, 1972–1977. https://catalog.ar
        chives.gov/id/542493.
    [2] T. Arnold and L. Tilton. Distant Viewing: Computational Exploration of Digital Images.
        MIT Press, 2023.
    [3] A. J. Barnes, J. D. Graham, and D. M. Konisky. Fifty years at the US Environmental Protec-
        tion Agency: progress, retrenchment, and opportunities. Rowman & Littlefield Publishers,
        2021.
    [4] R. Barthes. “Rhétorique de l’image”. In: communications 4.1 (1964), pp. 40–51.
    [5] B. Berman and M. M. Cronin. “Project DOCUMERICA: A Cautionary Tale”. In: Journal-
        ism History 43.4 (2018), pp. 186–197.
    [6] T. W. Bober, D. Vacco, T. J. Dagon, and H. E. Fowler. “The Photographic Process”. In:
        Handbook of Industrial and Hazardous Wastes Treatment (2004), p. 297.
    [7] B. I. Bustard and D. S. Ferriero. Searching for the Seventies: The DOCUMERICA Photogra-
        phy Project. Foundation for the National Archives, 2013.
    [8] N. Carter. The politics of the environment: Ideas, activism, policy. Cambridge University
        Press, 2018.
    [9] M. Evening. The Adobe Photoshop Lightroom 5 Book: The Complete Guide for Photogra-
        phers. Pearson Education, 2013.
[10]      H. S. Faridul, T. Pouli, C. Chamaret, J. Stauder, E. Reinhard, D. Kuzovkin, and A. Trémeau.
          “Colour mapping: A review of recent methods, extensions and applications”. In: Com-
          puter Graphics Forum 35.1 (2016), pp. 59–88.
[11]      T. Hastie, R. Tibshirani, J. H. Friedman, and J. H. Friedman. The elements of statistical
          learning: data mining, inference, and prediction. Vol. 2. Springer, 2009.
[12]      R. Hirsch. Seizing the Light: A social & Aesthetic history of photography. Routledge, 2017.
[13]      Y. Hwang, J.-Y. Lee, I. So Kweon, and S. Joo Kim. “Color transfer using probabilistic mov-
          ing least squares”. In: Proceedings of the IEEE conference on computer vision and pattern
          recognition. 2014, pp. 3342–3349.
[14]      W. Jia, H. Zhang, X. He, and Q. Wu. “A comparison on histogram based image matching
          methods”. In: 2006 IEEE International Conference on Video and Signal Based Surveillance.
          Ieee. 2006, pp. 97–97.
3
    To download the color-adjusted collection, see: https://distantviewing.org/downloads.




                                                         46
[15]   M. Mahy, L. Van Eycken, and A. Oosterlinck. “Evaluation of uniform color spaces devel-
       oped after the adoption of CIELAB and CIELUV”. In: Color Research & Application 19.2
       (1994), pp. 105–121.
[16]   B. Matiash. The Visual Palette: Defining Your Photographic Style. Rocky Nook, Inc., 2015.
[17]   G. Mu. Traité du signe visuel: Pour une rhétorique de l’image. Seuil, 1992.
[18]   L. Neumann and A. Neumann. “Color style transfer techniques using hue, lightness and
       saturation histogram matching”. In: Computational Aesthetics in Graphics, Visualization
       and Imaging. 2005, pp. 111–122.
[19]   S. Pénichon. Twentieth Century Colour Photographs: The complete guide to processes, iden-
       tification & preservation. Thames & Hudson, 2013.
[20]   E. Reinhard, M. Adhikhmin, B. Gooch, and P. Shirley. “Color transfer between images”.
       In: IEEE Computer graphics and applications 21.5 (2001), pp. 34–41.
[21]   M. Rinde. “Richard Nixon and the rise of American environmentalism”. In: Distillations
       3.1 (2017), pp. 16–29.
[22]   B. L. Shubinski. “From FSA to EPA: Project documerica, the dustbowl legacy, and the
       quest to photograph 1970s America”. PhD thesis. University of Iowa, 2009.
[23]   A. Temkin. Color chart: Reinventing color, 1950 to today. The Museum of Modern Art,
       2008.
[24]   A. Varichon. Color Charts: A History. Princeton University Press, 2024.
[25]   X. Xiao and L. Ma. “Color transfer in correlated color space”. In: Proceedings of the
       2006 ACM international conference on Virtual reality continuum and its applications. 2006,
       pp. 305–309.




                                                47