<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>TUB @ MediaEval 2013 Visual Privacy Task: Reversible Scrambling with colour-preservative Characteristic</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Sebastian Schmiedeke</string-name>
          <email>schmiedeke@nue.tu-berlin.de</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Pascal Kelm</string-name>
          <email>kelm@nue.tu-berlin.de</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Thomas Sikora Communication Systems Group</string-name>
          <email>sikora@nue.tu-berlin.de</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Technische Universität Berlin</institution>
          ,
          <country country="DE">Germany</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2013</year>
      </pub-date>
      <fpage>18</fpage>
      <lpage>19</lpage>
      <abstract>
        <p>This paper describes our participation in the Visual Privacy Task of MediaEval 2013, which aims to obscure human occurrence in image sequences. As a result the recorded person should be unrecognisable. We use an approach which pseudo-randomly scrambles pixels within speci ed regions. This technique is reversible and preserves the colour characteristic of each region. So, colour-based approaches will still be able to automatically distinguish between di erently dressed individuals. The evaluations of our results show that the privacy aspect got a very high score in both objective and subjective metrics. Our approach has a lack of intelligibility since it was measured by applying the Histogram of Oriented Gradients which might be fail on scrambled areas since edges are not preserved.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. INTRODUCTION</title>
      <p>
        Video surveillance of public spaces is expanding.
Consequently, individuals are increasingly concerned about the
"invasiveness" of such ubiquitous surveillance and fear that
their privacy is at risk. The demands of stakeholders to
prevent criminal activities are often seen to be in con ict with
the privacy requirements of individuals. The main challenge
is to preserve the anonymity of the surveyed individuals and
also to ful l the stakeholders needs. The problem of privacy
protection in video surveillance is concerned in this year's
MediaEval Visual Privacy Task [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. A typical way to protect
privacy in images and videos is to apply techniques such as
blurring or masking. Since these techniques are irreversible,
scrambling is introduced in [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]: A transform-domain
scrambling technique, where pixels in the respective regions are
pseudo-randomly scrambled based on a secret key. Our
approach is quite similar, but applied on the pixel domain to
be independent in the encoder used.
      </p>
    </sec>
    <sec id="sec-2">
      <title>METHODOLOGY</title>
      <p>Our proposed privacy-protection approach consists of a
scrambling module that obfuscates regions of detected
persons. A human detector was not built since these regions
were provided by the organisers.</p>
      <p>These regions are then vertically partitioned into slices
having a width of one pixel. Our scrambling module
obfuscates these regions by shu ing these slices. So, an
obfuscated region di ers from its original version in a changed
sequence of these slices.</p>
      <p>For this scheme, we choose a column-wise swapping pixels
to recognise background structures and action. As seen in
Figure 1, the action of stealing can be coarsely recognised by
the extended arm visible through the horizontal black line
in the scrambled image.</p>
      <p>
        The shu e algorithm is based on a modi ed variant of
the Fisher-Yates method [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] which generates "random"
permutations. The original sequence c consisting of N slices
is changed by swapping each i-th slice with the j-th slice,
where j is de ned by a pseudo-random number generator.
The algorithm is implemented as depicted in Algorithm 1:
Algorithm 1 Modi ed Fisher-Yates shu e
1: procedure Shuffle(i; j)
2: i 0
3: for i &lt; N do
4: j random number 2 i
5: t c(i)
6: c(i) c(j)
7: c(j) t
8: end for
9: end procedure
j &lt; N
      </p>
      <p>So, the permutation of the slices of the detected regions
is determined by the order generated by a pseudo-random
sequence. The pseudo-random sequence is repeatable due to
the characteristics of the pseudo-random number generator.
This generator produces a random, but repeatable sequence
of integer numbers by specifying a certain, but xed seed.
This seed is generated from the hash value of a chosen
password. This value is xed for all regions in each frame and
video sequence. Since the pseudo-random sequence is
repeatable thru the given seed, the permutation of slices is
reversible. So, the scrambled image regions can be recovered
by knowing the password and the coordinates of each
region. Overlapping regions were united before scrambling to
ensure its reversibility. We choose for scrambling instead of
cryptography to be robust against image compression
artefacts and transmission errors. If these errors occur parts of
the image will be also distorted, but these errors are not
propagated.
reversible
scrambling
SCD
SCD</p>
    </sec>
    <sec id="sec-3">
      <title>EXPERIMENTS</title>
      <p>
        The video sequences of the VPT dataset [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] are
obscured by scrambling the areas of detected persons instead
of detected faces or parts within. We are sure that
individuals can be identi ed not only by their face but also their
clothes or accessories. So, the individuals are anonymised at
best and a colour-based cluster algorithm may also be able
to group regions depicting the same person.
      </p>
      <p>The evaluation of the obscured videos took place using
objective and subjective procedures. The objective metrics like
face detection, object tracking and person re-identi cation
were compared for each sequence pair.
Three metrics are used to evaluated the results:
Appropriateness, intelligibility, and privacy.</p>
      <p>Appropriateness stands for the in uence of the obscuring
lter on the human perception of the image distortion. The
objective score is measured using SSIM and PSNR, here a
score of ca. 0.41 is reached which is below the average. A
reason might be that quite large regions are scrambled which
a ect the pixel-based calculation of SSIM and PSNR. The
subjective score is based on aspects such as pleasantness,
distraction, and user acceptance. Here the score of ca. 0.15
is quite low resulting from distraction of the users.</p>
      <p>Intelligibility stands for the ability of tracking obscured
persons. The objective score is measured by applying the
Histogram of Oriented Gradients as a human detector. Since
our approach scrambles the slices of detected regions an
edge-based human detector is unlikely able to detect
person on these scrambled slices. So, a sub-standard value is
expected: 0.35. The subjective score is 0.18.</p>
      <p>
        The privacy metric concerns about the identi cation of
individuals through their faces or personal accessories. The
objective score is measured by a re-identi cation score from
a visual model [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] and a face detection using the Viola-Jones
algorithm from OpenCV. A score of ca. 0.80 is achieved
which is above the average due the scrambled person
regions. The subjective score is 0.97 even higher which is also
above the average. A high subjective score was excepted,
since it is very hard for the human eye to recognise a body
shape or even a speci c individual on a scrambled image.
      </p>
      <p>Table 1 provides the evaluation scores for these three
metrics.
4.</p>
    </sec>
    <sec id="sec-4">
      <title>CONCLUSION</title>
      <p>We propose a reversible approach for scrambling regions
within images or videos to obscure its content. This
approach ensures a maximum of privacy, but fails in the other
aspects, like appropriateness and intelligibility. In our mind,
coarse action within these scrambled regions are still
recognisable as depicted in Figure 1. The clue is that these regions
can be de-scrambled, if the coordinates of the regions and the
password which generated the seed for the pseudo-random
number generation are known.</p>
    </sec>
    <sec id="sec-5">
      <title>ACKNOWLEDGMENTS</title>
      <p>The research leading to these results has received funding
from the European Community's FP7 under grant
agreement number FP7-261743 (VideoSense). We would also like
to thank Martha Larson for organizing this event.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>A.</given-names>
            <surname>Badii</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Einig</surname>
          </string-name>
          , and
          <string-name>
            <given-names>T.</given-names>
            <surname>Piatrik</surname>
          </string-name>
          .
          <article-title>Overview of the MediaEval 2013 Visual Privacy Task</article-title>
          . In MediaEval 2013 Workshop, Barcelona, Spain, October
          <volume>18</volume>
          -19
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>F.</given-names>
            <surname>Dufaux</surname>
          </string-name>
          and
          <string-name>
            <given-names>T.</given-names>
            <surname>Ebrahimi</surname>
          </string-name>
          .
          <article-title>Video surveillance using jpeg 2000</article-title>
          .
          <source>In Optical Science and Technology, the SPIE 49th Annual Meeting</source>
          , pages
          <volume>268</volume>
          {
          <fpage>275</fpage>
          .
          <source>International Society for Optics and Photonics</source>
          ,
          <year>2004</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>R.</given-names>
            <surname>Durstenfeld</surname>
          </string-name>
          . Algorithm 235:
          <article-title>Random permutation</article-title>
          .
          <source>Commun. ACM</source>
          ,
          <volume>7</volume>
          (
          <issue>7</issue>
          ):
          <fpage>420</fpage>
          {,
          <year>July 1964</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>P.</given-names>
            <surname>Korshunov</surname>
          </string-name>
          and
          <string-name>
            <given-names>T.</given-names>
            <surname>Ebrahimi</surname>
          </string-name>
          .
          <article-title>PEViD: privacy evaluation video dataset</article-title>
          .
          <source>Applications of Digital Image Processing XXXVI</source>
          ,
          <fpage>25</fpage>
          -
          <lpage>29</lpage>
          August
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>