<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">MediaEval 2013 Visual Privacy Task: Using Adaptive Edge Detection for Privacy in Surveillance Videos</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Volker</forename><surname>Eiselein</surname></persName>
							<email>eiselein@nue.tu-berlin.de</email>
							<affiliation key="aff0">
								<orgName type="department">Communications Systems Group Technische</orgName>
								<orgName type="institution">Universität Berlin</orgName>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Tobias</forename><surname>Senst</surname></persName>
							<email>senst@nue.tu-berlin.de</email>
							<affiliation key="aff1">
								<orgName type="department">Communications Systems Group Technische</orgName>
								<orgName type="institution">Universität Berlin</orgName>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Ivo</forename><surname>Keller</surname></persName>
							<email>keller@nue.tu-berlin.de</email>
							<affiliation key="aff2">
								<orgName type="department">Communications Systems Group Technische</orgName>
								<orgName type="institution">Universität Berlin</orgName>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Thomas</forename><surname>Sikora</surname></persName>
							<email>sikora@nue.tu-berlin.de</email>
							<affiliation key="aff3">
								<orgName type="department">Communications Systems Group Technische</orgName>
								<orgName type="institution">Universität Berlin</orgName>
							</affiliation>
						</author>
						<title level="a" type="main">MediaEval 2013 Visual Privacy Task: Using Adaptive Edge Detection for Privacy in Surveillance Videos</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">52C44084868835D23B61A4F8F1570F09</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-19T17:59+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Privacy preservation</term>
					<term>video analysis</term>
					<term>obfuscation</term>
					<term>adaptive edge detection</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>In this paper we present a system for preserving the privacy of individuals in a video surveillance scenario. While a person's privacy should not be revealed to a viewer of the video without special needs, it is still important that the action in a scene as the semantic content of a video remain perceivable by a human observer.</p><p>The proposed system uses edge detection and adaptive thresholding in order to estimate the persons' silhouettes in a video scene and thus rendering most of their actions visible, while hiding sensitive personal information. In order to obtain a more complete contour around a person, an adaptive thresholding scheme using edge histograms is used as well as background subtraction which limits the edge extraction to foreground masks and thus avoids distraction of the viewer's eyes to background structures.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">INTRODUCTION</head><p>With the increasing usage of Video Surveillance systems and a continuously growing amount of CCTV data being recorded and looked at, the need to apply privacy protection techniques in this area rises as well. In order to increase acceptance of CCTV-cameras among people being observed, it is of special importance to ensure their personal rights be not violated, while the surveillance system is still able to work and security staff is able to identify critical events in a video stream. The MediaEval 2013 Visual Privacy Task <ref type="bibr" target="#b1">[1]</ref> addresses this issue and provides an evaluation based on the PEViD Dataset <ref type="bibr" target="#b5">[5]</ref>.</p><p>Copyright is held by the author/owner(s).</p><p>MediaEval 2013 Workshop, October 18-19, 2013, Barcelona, Spain For obfuscation of personal data, we propose to show only the contour of a person in the video. In case of a low number of persons in the scene, their movements and actions could be identified while it is still impossible to identify personal details such as the person's face, color of clothes or skin color. In this paper, we show a system which uses adaptive edge detection to determine the contour of a person and in the privacy-filtered result video replaces the interior of a person's silhouette by background information. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">SYSTEM DESCRIPTION</head><p>The proposed system uses a background model in order to exchange the information within a person's contour by recent background information (see Fig. <ref type="figure" target="#fig_0">1</ref>). The bounding boxes which are provided as manually-annotated ground truth for the purpose of MediaEval are used in order to identify where the privacy filter needs to be applied in the image. In these regions, edge detection is performed, and the result is blended into the background image. As the bounding boxes for objects could also be provided automatically (e.g. by human detection algorithms or trackers such as <ref type="bibr" target="#b3">[3]</ref>), our system does not necessarily depend on manual annotations.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1">Double Background model</head><p>Our algorithm uses two background models: a) a standard Gaussian-Mixture Model (GMM) similar to <ref type="bibr" target="#b4">[4]</ref> in order to obtain accurate foreground masks and b) a very simplified background model which essentially consists of only a single frame. The latter is used to maintain a very recent background information. This allows the viewer (or an analytics algorithm) to identify e.g. objects left in the scene or graffiti sprayed on the walls. As one of the aims of our system is to give the user a clear impression of what is going on in the scene, it is crucial to provide very recent and fast background information. We therefore just use a simple on-line learning method which adapts the background pixels for the regions without objects as follows:</p><formula xml:id="formula_0">BGi(x, y) = (1 − α) • BGi−1(x, y) + α • I(x, y)<label>(1)</label></formula><p>In order to adapt the background model quickly to the new frame, we use rather high values for α (e.g. α = 0.35) and update this model every 5th frame. This also has the advantage that the system can cope quickly with slight camera movements. In some of the PEViD videos, the camera shakes a bit and the image is shifted for some pixels. While first the background image in our system gets a bit blurry in these situations, after a few frames our system maintains the current background and is not disturbed by older values anymore. The GMM uses a much slower learning rate and is only used to mask out background parts within the given region of interest (ROI) -a task for which the previously described background model b) alone would be to simple.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2">Adaptive Edge Detection</head><p>We use Canny edge detection <ref type="bibr" target="#b2">[2]</ref> in order to extract the silhouettes of the persons in the image. After a noise reduction step, this algorithms computes the gradients in the image and performs thresholding with hysteresis. A common problem for this algorithm is the choice of the two thresholds T1, T2 used in the hysteresis step. While T1 sets the minimum edge level accepted for the starting points, T2 determines which edge level must be kept during the hysteresis. Usually, it is hard to set general values which give satisfying results for arbitrary video sequences.</p><p>In our system which has to work under varying conditions (indoor, outdoor, different lighting, changing weather...), we propose to adapt the threshold according to the gradient histogram in the ROI. For the given ROI, we compute the absolute values of the gradient information and set up a histogram over every ROI. Using the assumption, that most of a human's silhouette can be recovered by the highest 15 % of the gradient information, we can thus find T1 for the Canny algorithm. In order to close holes in the silhouette, T2 is set to T2 = 0.9 • T1. This automatic choice of T1 and T2 adapts well to most scenes in the data set and gives a good contour information of a walking person in most cases.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">EVALUATION RESULTS</head><p>The evaluation of the proposed filter is based on objective and subjective metrics described in <ref type="bibr" target="#b1">[1]</ref>. The results are shown in Table <ref type="table" target="#tab_0">1</ref>. The proposed filter score has been compared with the average score of all 9 participants of the MediaEval Visual Privacy Task challenge. The results are very promising in terms of the privacy behavior where the adaptive edge filter is competitive with other participants.</p><p>In regards to the intelligibility and appropriateness the results of the subjective and objective evaluations related to the average of all participants are contrary. While for these categories the score of the objective results are less than the average, the score of the subjective results is higher than the average score. As the proposed filter extracts and reprojects the edge information into the image and fills the interior of </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">CONCLUSIONS</head><p>In this paper we showed how adaptive edge detection can be used to preserve the privacy of people in CCTV videos. The subjective test demonstrates that the proposed filter outperforms most of the other privacy filters proposed in the MediaEval 2013 challenge. Thus, our method is of special interest for semi-automatic video surveillance systems which need to hide sensitive personal information while preserving the context of actions.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: original PEViD frame (left) and privacyfiltered result (right)</figDesc><graphic coords="1,316.82,382.33,117.04,65.83" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1 :</head><label>1</label><figDesc>Evaluation results for the proposed method compared to 9 other methods in the workshop.a person with background information, it is obvious that a standard person detection and tracking method which is usually based on extracting edges by HoG or wavelet filtering and is trained on normal videos cannot have the same performance as in unfiltered images. To apply the filtered videos for automated video analytics the respective preprocessing of the methods has to be adapted and we thus recommend a new training step based on feature extraction on the filtered images. However, based on the results of the subjective evaluation, the filter is especially appropriate for display in video surveillance systems that are guided and evaluated by human operators.</figDesc><table><row><cell></cell><cell>Objective evaluation</cell><cell></cell></row><row><cell></cell><cell cols="2">Adaptive edge filter Average (9)</cell></row><row><cell>Intelligibility</cell><cell>0.358355</cell><cell>0.502378</cell></row><row><cell>Privacy</cell><cell>0.693835</cell><cell>0.664903</cell></row><row><cell>Appropriateness</cell><cell>0.368627</cell><cell>0.56048</cell></row><row><cell cols="2">Subjective evaluation</cell><cell></cell></row><row><cell></cell><cell cols="2">Adaptive edge filter Average (9)</cell></row><row><cell>Intelligibility</cell><cell>0.678333</cell><cell>0.655741</cell></row><row><cell>Privacy</cell><cell>0.683750</cell><cell>0.683843</cell></row><row><cell>Appropriateness</cell><cell>0.532500</cell><cell>0.492130</cell></row></table></figure>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">ACKNOWLEDGMENTS</head><p>This work was supported by the European Commission under contracts FP7-261743 VideoSense.</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<monogr>
		<title/>
		<author>
			<persName><surname>References</surname></persName>
		</author>
		<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<monogr>
		<title level="m" type="main">Overview of the mediaeval 2013 visual privacy task</title>
		<author>
			<persName><forename type="first">A</forename><surname>Badii</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Einig</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Piatrik</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2013">2013</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">A computational approach to edge detection</title>
		<author>
			<persName><forename type="first">J</forename><surname>Canny</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Trans. Pattern Anal. Mach. Intell</title>
		<imprint>
			<biblScope unit="volume">8</biblScope>
			<biblScope unit="issue">6</biblScope>
			<biblScope unit="page" from="679" to="698" />
			<date type="published" when="1986">1986</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">A motion-enhanced hybrid probability hypothesis density filter for real-time multi-human tracking in video surveillance scenarios</title>
		<author>
			<persName><forename type="first">V</forename><surname>Eiselein</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Senst</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Keller</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Sikora</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">15th IEEE International Workshop on Performance Evaluation of Tracking and Surveillance (PETS)</title>
				<imprint>
			<date type="published" when="2013">2013</date>
			<biblScope unit="page" from="6" to="13" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Detection of static objects for the task of video surveillance</title>
		<author>
			<persName><forename type="first">R</forename><surname>Heras Evangelio</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Senst</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Sikora</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">IEEE Workshop on Applications of Computer Vision (WACV)</title>
				<imprint>
			<date type="published" when="2011">2011</date>
			<biblScope unit="page" from="534" to="540" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">PEViD: privacy evaluation video dataset</title>
		<author>
			<persName><forename type="first">P</forename><surname>Korshunov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Ebrahimi</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of SPIE</title>
				<meeting>SPIE</meeting>
		<imprint>
			<date type="published" when="2013">2013</date>
			<biblScope unit="volume">8856</biblScope>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
