<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>MediaEval 2013 Visual Privacy Task: Physics-Based Technique for Protecting Privacy in Surveillance Videos</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Mohamed Sedky</string-name>
          <email>m.h.sedky@staffs.ac.uk</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Claude C. Chibelushi</string-name>
          <email>c.c. chibelushi@staffs.ac.uk</email>
          <email>chibelushi@staffs.ac.uk</email>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Mansour Moniri</string-name>
          <email>m.moniri@staffs.ac.uk</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Staffordshire University</institution>
          ,
          <addr-line>Beaconside, Stafford, ST18 0AD</addr-line>
          ,
          <country country="UK">United Kingdom</country>
          ,
          <addr-line>+44(0)1785353260</addr-line>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Staffordshire University</institution>
          ,
          <addr-line>Beaconside, Stafford, ST18 0AD</addr-line>
          ,
          <country country="UK">United Kingdom</country>
          ,
          <addr-line>+44(0)1785353634</addr-line>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Staffordshire University</institution>
          ,
          <addr-line>Beaconside, Stafford, ST18 0AD</addr-line>
          ,
          <country country="UK">United Kingdom</country>
          ,
          <addr-line>+44(0)1785353802</addr-line>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2013</year>
      </pub-date>
      <fpage>18</fpage>
      <lpage>19</lpage>
      <abstract>
        <p>This paper describes a physics-based technique for protecting the privacy of people in videos as defined by the MediaEval 2013 Visual Privacy task. We propose a physics-based approach which estimates the full spectrum of the surface spectral reflectance from the video. Whereby the wavelength which corresponds to the global minimum of the spectral curve (an intrinsic feature of the material) at a pixel is calculated and converted to RGB values which are used to filter pixels that belong to a moving object. This effectively implements visual privacy protection by replacing foreground pixel colour by another which is related to intrinsic optical properties of the original pixel. Both objective and subjective evaluations are performed using both video analytics algorithm and user studies in order to evaluate the proposed technique.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>In this challenge, we propose a physics-based technique to protect
privacy in surveillance videos. Privacy protection approaches may
be classified depending on the image representation used as
physics-based or non-physics-based. Non-physics based
approaches use one of the known colour spaces, with no explicit
physics underpinning, as a cue to model the object. The word
physics refers to the extraction of intrinsic features about the
materials contained in the object based on an understanding of the
underlying physics which govern the image formation. This
process is achieved by applying physics-based image formation
models which attempt to estimate or eliminate the illumination
and/or the geometric parameters in order to extract information
about the surface reflectance.</p>
      <p>
        Conventional video cameras, analogous to a retina, sense reflected
light so that colour values are the integration of the product of
incident illumination power spectral distribution, object’s surface
Copyright is held by the author/owner(s).
spectral reflectance and camera sensors sensitivities. Humans have
the ability to separate the illumination power spectral distribution
from the surface spectral reflectance when judging object
appearance, such ability is called colour constancy ‎[
        <xref ref-type="bibr" rid="ref5">5</xref>
        ].
      </p>
    </sec>
    <sec id="sec-2">
      <title>2. SYSTEM DESCRIPTION</title>
      <p>Our proposed privacy-protection technique consists of two steps.
First, a change detection step segments moving objects (the
foreground). Second, a filter changes foreground pixels, so as to
achieve visual privacy, by converting the surface spectral
reflectance at these pixels into RGB values as perceived by
humans.
2.1</p>
    </sec>
    <sec id="sec-3">
      <title>Change Detection</title>
      <p>
        Spectral-360 ‎[
        <xref ref-type="bibr" rid="ref6">6</xref>
        ], a novel change detection algorithm has been
used. The algorithm uses colour constancy techniques to
computationally estimate a consistent physics-based colour
descriptor model of surface spectral reflectance from the video
and then to correlate the full-spectrum reflectance of the
background and foreground pixels to segment the foreground
from a static background. The rationale behind this approach is
that the segmentation between foreground and background objects
can be done through the matching between the surface spectral
reflectance over the visible wavelengths of a reference
background frame and each new frame. The challenge of this
approach arises from the new idea of processing the full-spectrum
of the surface spectral reflectance instead of the three samples
used by other colour spaces. The spectral representation uses a
linear model, which consists of a number of basis functions
(pretrained from a set of materials) and weights (calculated for the
object under investigation). A numerical estimation of the
physics-based model for image formation and the real-time
transformation from the video to the physical parameters is carried
out.
      </p>
    </sec>
    <sec id="sec-4">
      <title>2.2 Privacy Filter</title>
      <p>
        In order to build a computational physical model, the illumination
is estimated by segmenting areas in the image which represent
high specularities (highlights); McCamy’s formula ‎[
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] is then
applied and the correlated colour temperature is calculated. The
illumination spectral power distribution is then calculated using
Plank’s formula ‎[
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. Using the dichromatic model, and by
assuming diffuse-only reflection and the existence of a dominant
illuminant, the surface spectral reflectance is then recovered. The
wavelength that corresponds to the global minimum of the surface
spectral reflectance is then calculated and converted to RGB
values as perceived by humans.
      </p>
    </sec>
    <sec id="sec-5">
      <title>3. RESULTS</title>
      <p>
        The resulting filtered videos have been evaluated using both
objective and subjective procedures. The objective metrics
compare each pair of original and filtered video in terms of: face
detection accuracy, human body tracking accuracy, image quality
metrics ‎[
        <xref ref-type="bibr" rid="ref4">4</xref>
        ].
As a whole, our technique shows a high degree of privacy, and an
average level of intelligibility, but low level of visual
appropriateness. The visual appropriateness of our proposed filter
can be adjusted for future improvement.
      </p>
    </sec>
    <sec id="sec-6">
      <title>4. CONCLUSIONS</title>
      <p>This paper argues that image formation models offer interesting
new alternative physics-based cues for privacy protection
compared with other representations. Features such as surface
spectral reflectance have not been applied yet in the field of
privacy protection. The reason is the computational complexity of
such models, and hence possible unfeasibility of real-time
implementation. This challenge was tackled, firstly, by choosing
an appropriate reflection model, the dichromatic reflection model.
Secondly, by setting a feasible set of assumptions for such model
which best match the reduction of model complexity and does not
contradict with real-world operational conditions. In this paper we
have proposed a physics-based approach which segments the
moving objects and then estimates the full spectrum of surface
spectral reflectance from the video, where the wavelength that
corresponds to the global minimum is calculated and converted to
RGB values as perceived by humans. This effectively replaces
foreground pixel colours by others, as a simple privacy protection
mechanism. Our ongoing work is investigating enhanced pixel
replacement schemes, beyond the obfuscation mechanism
described here, so as to protect privacy while minimizing the
change of the semantic content of the image.</p>
      <p>Team: CIISS1</p>
      <sec id="sec-6-1">
        <title>Intelligibility</title>
      </sec>
      <sec id="sec-6-2">
        <title>Privacy</title>
      </sec>
      <sec id="sec-6-3">
        <title>Appropriateness</title>
      </sec>
      <sec id="sec-6-4">
        <title>Team: CIISS</title>
      </sec>
      <sec id="sec-6-5">
        <title>Intelligibility</title>
      </sec>
      <sec id="sec-6-6">
        <title>Privacy</title>
      </sec>
      <sec id="sec-6-7">
        <title>Appropriateness</title>
        <p>Our Score
0.453317
0.812541</p>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>Hoghai</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Shengyong</given-names>
            <surname>Chen</surname>
          </string-name>
          and
          <string-name>
            <given-names>N.</given-names>
            <surname>Kubota</surname>
          </string-name>
          .
          <article-title>Intelligent Video Systems and Analytics: A Survey</article-title>
          .
          <source>In IEEE Transactions on Industrial Informatics</source>
          , volume
          <volume>9</volume>
          , issue 3, pp.
          <fpage>1222</fpage>
          -
          <lpage>1233</lpage>
          ,
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>M.</given-names>
            <surname>Sedky</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C. C.</given-names>
            <surname>Chibelushi</surname>
          </string-name>
          and
          <string-name>
            <given-names>M.</given-names>
            <surname>Moniri</surname>
          </string-name>
          .
          <article-title>Smart Video Surveillance for Workplace Applications: Implications, Technologies and Requirements</article-title>
          .
          <source>In Proceedings of The 5th IASTED International Conference on Visualization, and Image Processing</source>
          , Spain,
          <year>2005</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>A.</given-names>
            <surname>Cripps</surname>
          </string-name>
          . Workplace Surveillance.
          <source>Intern at NSW Council for Civil Liberties</source>
          <year>2004</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>A.</given-names>
            <surname>Badii</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Einig</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Piatrik</surname>
          </string-name>
          .
          <article-title>Overview of the MediaEval 2013 Visual Privacy Task</article-title>
          .
          <source>MediaEval 2013 Workshop</source>
          , Barcelona, Spain,
          <source>October 18-19</source>
          ,
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>J.</given-names>
            <surname>Ho</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B. V.</given-names>
            <surname>Funt</surname>
          </string-name>
          , and
          <string-name>
            <given-names>M. S.</given-names>
            <surname>Drew</surname>
          </string-name>
          .
          <article-title>Separating a colour signal into illumination and surface reflectance components: Theory and applications</article-title>
          .
          <source>In: IEEE Transaction on Pattern Analysis and Machine Intelligence</source>
          , vol.
          <volume>12</volume>
          , pp.
          <fpage>966</fpage>
          -
          <lpage>977</lpage>
          ,
          <year>1990</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>M.</given-names>
            <surname>Sedky</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Moniri</surname>
          </string-name>
          and
          <string-name>
            <given-names>C. C.</given-names>
            <surname>Chibelushi</surname>
          </string-name>
          .
          <article-title>Object Segmentation Using Full-Spectrum Matching of Albedo Derived from Colour Images. PCT patent application international application no</article-title>
          .
          <source>PCT/GB2009/002829, EP2374109</source>
          ,
          <year>2009</year>
          .
          <source>US patent no. 2374109 12.10</source>
          .
          <string-name>
            <surname>2011</surname>
            <given-names>US</given-names>
          </string-name>
          ,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <surname>McCamy</surname>
            and
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Calvin</surname>
          </string-name>
          .
          <article-title>Correlated Colour Temperature as an Explicit Function of Chromaticity Coordinates</article-title>
          .
          <source>In: Journal of Colour Research &amp; Application</source>
          , vol.
          <volume>17</volume>
          , no.
          <issue>2</issue>
          , pp.
          <fpage>142</fpage>
          -
          <lpage>144</lpage>
          ,
          <year>1992</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>