<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Towards Measuring Risk Factors in Privacy Policies</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Najmeh</forename><forename type="middle">Mousavi</forename><surname>Nejad</surname></persName>
							<email>nejad@cs.uni-bonn.de</email>
						</author>
						<author>
							<persName><forename type="first">Damien</forename><surname>Graux</surname></persName>
							<email>damien.graux@iais.fraunhofer.de</email>
						</author>
						<author>
							<persName><forename type="first">Diego</forename><surname>Collarana</surname></persName>
							<email>diego.collarana.vargas@iais.fraunhofer.de</email>
						</author>
						<author>
							<affiliation key="aff0">
								<orgName type="institution" key="instit1">Fraunhofer IAIS</orgName>
								<orgName type="institution" key="instit2">University of Bonn Sankt Agustin</orgName>
								<address>
									<country key="DE">Germany</country>
								</address>
							</affiliation>
						</author>
						<author>
							<affiliation key="aff1">
								<orgName type="institution">Fraunhofer IAIS Sankt Agustin</orgName>
								<address>
									<country key="DE">Germany</country>
								</address>
							</affiliation>
						</author>
						<author>
							<affiliation key="aff2">
								<orgName type="institution">Fraunhofer IAIS Sankt Agustin</orgName>
								<address>
									<country key="DE">Germany</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Towards Measuring Risk Factors in Privacy Policies</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">D22B1B7112E5BE801F5E94709C26F1D6</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-23T21:01+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Privacy policy</term>
					<term>Deep learning</term>
					<term>Rule-based information extraction</term>
					<term>Risk factor</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>The ubiquitous availability of online services and mobile apps results in a rapid proliferation of contractual agreements in the form of privacy policies. Despite the importance of such consent forms, the majority of users tend to ignore them due to their content length and complexity. Thus, users might be consenting policies that are not aligned to regulations in laws such as the GDPR from the EU law. In this study, we propose a hybrid approach which measures a privacy policy's risk factor applying both supervised deep learning and rule-based information extraction. Benefiting from an annotated dataset of 115 privacy policies, a deep learning component is first able to predict high-level categories for each paragraph. Then, a rule-based module extracts pre-defined attributes and their values, based on high-level classes. Finally, a privacy policy's risk factor is computed based on these attribute values.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">INTRODUCTION</head><p>In the current digital era, almost everyone is exposed to accepting contractual agreements in the form of privacy policies. However, the majority of people skip privacy policies due to their length and complex terminology. According to a recent survey, from 543 university students, only 26% did not choose the 'quick join' routine, while joining a factious social network and unsurprisingly, their average reading time was only 73 seconds <ref type="bibr" target="#b1">[2]</ref>. Moreover, for the administrative state is it important to validate the compliance the privacy policies with a correspondent law. For example, the EU regulation General Data Protection Regulation (GDPR) states that the retention period must be specified and limited.</p><p>To assist end-users with consciously agreeing to the conditions, we can apply Natural Language Processing (NLP) and Information Extraction (IE) to present a privacy policy in a structured view. Our approach applies supervised deep learning using an annotated dataset (named OPP-115), to assign high-level classes to a privacy policy's paragraphs. Then, according to predicted classes, we define hand-coded rules based on experts annotations, to extract attributes values from each paragraph. Finally, having detailed information for each paragraph, a risk measurement function computes a risk factor based on extracted information. Consequently, a user could choose to stop using a website, if the predicted risk score is high. Additionally, this structured view can be also used by the administrative state to perform a shallow compliance checking.</p><p>OPP-115 is a widely-used dataset in the context of privacy policy analysis <ref type="bibr" target="#b4">[5]</ref>. It contains in-depth annotations for 115 privacy policies at paragraph level and each paragraph was annotated by 3 experts. There are two types of annotations: high-level classes which define 10 data practice categories; and low-level attributes which include mandatory and optional attributes. For instance, the high-level class First Party Collection/Use has 3 attributes: Collection Mode (explicit or implicit), Information Type (financial, health, contact, location, etc.) and Purpose (advertising, marketing, analytics, legal requirement, etc.).</p><p>The approach proposed in this paper, is built upon on our previous effort, which exploits OPP-115 and deep learning to solve a multi-label classification problem. We feed privacy policy's paragraphs along with the predicted classes into a rule-based IE component and retrieve attribute values. The rules are defined based on OPP-115 low-level annotations. Finally, all predicted categories and extracted information are passed into a risk measurement module and a risk factor will be computed based on hand-coded rules.</p><p>The paper is divided into the following sections: in Section 2, we provide an overview of existing effort on measuring risks in privacy policies; Section 3 presents our proposed approach and our evaluation scheme; and finally Section 4 will conclude this paper.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">RELATED WORK</head><p>In light of the, now enforced EU-wide, General Data Protection Regulation (GDPR) <ref type="bibr" target="#b3">[4]</ref>, there has been an increasing interest towards privacy policy analysis as this new set of regulations increases the constrains for companies holding customers data. Here, we provide a brief overview of studies that specifically addressed risk levels in privacy policies.</p><p>Polisis is an online service for automatic analysis of privacy policies <ref type="bibr" target="#b0">[1]</ref>. Along with classification and structured presentations of privacy policies, it assigns privacy icons which are based on the Disconnect<ref type="foot" target="#foot_0">1</ref> icons. These icons include Expected Use, Expected Collection, Precise Location, Data Retention and Children Privacy. For instance, Data Retention color assignments are: Green for retention periods of less than a year; Yellow, when the retention period is longer than one year; and Red, when there is no data retention policy provided. Polisis benefits from OPP-115 and employs supervised machine learning to extract high-level categories (in the above example, Data Retention) and attribute values of each category (e.g., Retention Period in this case). Finally, based on retrieved  attribute values and heuristic rules, privacy icons along with their colors are produced. Currently, Polisis's interface generates only a limited set of privacy icons. In future, we intend to further analyze privacy icons and extend them with the help of legal experts.</p><p>PrivacyCheck is an approach for automatic summarization of privacy policies using data mining <ref type="bibr" target="#b5">[6]</ref>. It answers 10 pre-defined questions concerning privacy and security of users' data and is also available as a Chrome browser extension. In order to train the model, a corpus containing 400 privacy policies was compiled and 7 privacy experts manually assigned risk levels (Green, Yellow, Red) to the 10 factors. First, a pre-processing step finds those paragraphs that have at least one keyword related to one of 10 factors. The methodology of selecting keywords was largely manual. Then, the selected paragraphs will be sent to a data mining server where 11 data mining models were trained, one for checking if the corresponding page is a privacy policy and one each for the 10 questions. The authors claim that on average, 60% of the times, PrivacyCheck finds the correct risk level. The limitation of PrivacyCheck is its lack of Inter Annotator Agreement (IAA) for the annotators. According to the paper, the quality control was performed by assigning each policy to two team members. However, only 15% of privacy policies were compared and their discrepancies were resolved which makes the training dataset less reliable.</p><p>PrivacyGuide is another summarization tool inspired by GDPR that classifies a privacy policy into 11 categories using NLP and machine learning and further measures the associated risk level of each class <ref type="bibr" target="#b2">[3]</ref>. Similar to previous studies, PrivacyGuide uses the three-level scale risk based on classification (i.e. Green, Yellow, Red). The 11 criteria and their associated risk levels were defined by GDPR experts. Based on these criteria, a privacy corpus was compiled with the help of 35 university students. Each participant assigned a privacy category to text snippets and classified them with a risk level. The author reported that the weighted average accuracy is 74% for classifying a privacy policy into one of the 11 classes and the accuracy of risk level detection is 90%. Although the results were encouraging, the dataset was not annotated by experts which is a fundamental criterion in legal text processing and analysis.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">PROPOSED APPROACH</head><p>In this section, we provide details of our approach for measuring a privacy policy's risk factor. Our proposed method leverages OPP-115 annotated dataset for training and evaluation <ref type="bibr" target="#b4">[5]</ref>. As discussed earlier, OPP-115 high-level annotations are divided into 10 classes: (9) International/Specific Audiences: practices that target a specific group of users (e.g., children, Europeans, etc.) (10) Other: additional practices not covered by the other categories.</p><p>In addition, each high-level category includes low-level attribute annotations. For instance, Data Retention category is further annotated with its attributes, which are: Retention Period, Retention Purpose and Information Type. The annotators provided either one or several values for each attribute along with the span of text based on which they have chosen that specific value(s). In the above example, Retention Period may have one of the following values: stated period, limited, indefinitely or unspecified.</p><p>Figure <ref type="figure" target="#fig_0">1</ref> shows the architecture of our proposed approach which consists of three main components: 1) a deep learning module is trained to predict high-level classes of a policy's paragraphs; 2) a rule-based pipeline in which the rules are defined based on lowlevel attribute annotations of OPP-115; and 3) a risk measurement function that assigns risk icons along with their corresponding colors (green, yellow, red), according to extracted information.</p><p>Following conventional ML practices, in the deep learning component, dataset splits are randomly partitioned into a ratio of 3:1:1 for training, validation and testing respectively; while maintaining a stratified set of labels. We further decomposed the Other category into its attributes: Introductory/Generic, Privacy Contact Information and Practice Not Covered. Therefore, considering that a paragraph in the dataset may be labeled with more than one category, we face a multi-label classification problem with 12 classes. The implementation of the ML component is completed and we achieve 79% micro-average for F1.</p><p>The high-level predicted classes are passed to the rule-based component where low-level attribute values will be extracted. The definition of rules are based on experts annotations in OPP-115 dataset. We intend to use 60% of low-level annotations for defining the rules, 20% for validating the defined rules and the remaining 20% for the final test. Having information about attribute values, the risk measurement module is able to assign appropriate risk icons along with their corresponding colors. As a proof-of-concept, we will found our risk measurement rules on Disconnect icons. Aforementioned in literature review, the Disconnect Data Retention color assignment are as follows: Green for retention period &lt;= 12 months; Yellow, for retention period &gt; 12 months; and Red, when there is no data retention policy provided. Algorithm 1 shows our interpretation of Data Retention icon. It is worth to mention that our interpretation is based on the available annotations from OPP-115 dataset. Hence, it is not the only representation that can be built from Disconnect icons and others may adopt their own understanding.</p><p>For the evaluation of our approach, we intend to generate risk factors according to OPP-115 experts annotations and use it as a goldstandard. We believe the final error will be close to sum of error rate in the deep learning module (predicting high-level classes) and the error which is caused due to incomplete set of rules in rule executor component. Considering the fact that we are now able to predict the correct high-level classes with 79% F1, with the careful definition of rules for extracting attribute values, it is predicted to gain a reasonable accuracy at the end of our pipeline.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">CONCLUSION</head><p>In this study, we proposed the application of Deep Learning models and Rule-Based Information Extraction to automatically present a structured view of risk factors in privacy policies. In particular, we presented a hybrid approach that takes advantage of the dataset OPP-115. This approach is of paramount importance to support users to consciously agree with terms and conditions of online services, and to perform shallow compliance checking where a highrisk score can be assigned to "indefinitely" and "unspecified" values. As next steps, we plan to implement the proposed architecture and run empirical evaluations to validate the presented hypothesis, i.e, users will be more motivated to read privacy policies when a color-coded structured view is presented to them.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: General Architecture.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>( 1 )</head><label>1</label><figDesc>First Party Collection/Use: how and why the information is collected. (2) Third Party Sharing/Collection: how the information may be used or collected by third parties. (3) User Choice/Control: choices and controls available to to users. (4) User Access/Edit/Deletion: if users can modify their information and how. (5) Data Retention: how long the information is stored. (6) Data Security: how is users' data secured. (7) Policy Change: if the service provider will change their policy and how the users are informed. (8) Do Not Track: if and how Do Not Track signals 2 is honored.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head></head><label></label><figDesc>Table 1 shows some sample rules for finding values of Retention Period attribute in Data Retention category. We found our rules definitions based on experts annotations. As shown in the table, the rules definition use the knowledge about high-level categories predicted by the deep learning component.</figDesc><table><row><cell cols="2">Algorithm 1 Sketch of risk measurement algorithm</cell></row><row><cell cols="2">Require: predicted high-level category, extracted attribute values</cell></row><row><cell cols="2">1: for all paragraphs in the privacy policy do</cell></row><row><cell>2:</cell><cell>cat eдory ← predicted high-level category</cell></row><row><cell>3:</cell><cell>if cat eдor y ∈ Data Retention then</cell></row><row><cell>4:</cell><cell>Ret ent ionP er iod ← extracted retention period</cell></row><row><cell>5:</cell><cell>if Ret ent ionP er iod ∈ (Stated Period, Limited) then</cell></row><row><cell>6:</cell><cell>Dat aRet ent ionI con ← Green</cell></row><row><cell>7:</cell><cell>else if Ret ent ionP er iod ∈ Indefinitely then</cell></row><row><cell>8:</cell><cell>Dat aRet ent ionI con ← Yellow</cell></row><row><cell>9:</cell><cell>else</cell></row><row><cell>10:</cell><cell>Dat aRet ent ionI con ← Red</cell></row><row><cell>11:</cell><cell>end if</cell></row><row><cell>12:</cell><cell>end if</cell></row><row><cell>13:</cell><cell>if cat eдory ∈ First Party Collection/Use then ...</cell></row><row><cell>14:</cell><cell>end if</cell></row><row><cell cols="2">15: end for</cell></row><row><cell cols="2">Ensure: risk icons and their corresponding colors</cell></row></table></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_0">https://disconnect.me/ Conference'17, July</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2017" xml:id="foot_1">, Washington, DC, USA Mousavi Nejad et al.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_2">https://en.wikipedia.org/wiki/Do_Not_Track</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_3">. The personal information collected about you through our online applications and in our communications with you is stored in our internal database.</note>
		</body>
		<back>
			<div type="annex">
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Conference'17, July 2017, Washington, DC, USA Indefinitely The posts and content you made will not be automatically deleted as part of the account removal process.</p><p>[store/keep/retain/maintain][Token]*[indefinitely] Indefinitely 1.This data is generally retained indefinitely.</p><p>2. The information we collect for statistical analysis and technical improvements is maintained indefinitely.</p><p>[store/keep/retain/maintain][Token]*[as long as][Token]+ Limited 1. We will retain your information for as long as your account is active or as needed to provide you services.</p><p>2. We will retain your personal information while you have an account and thereafter for as long as we need it for purposes not prohibited by applicable laws</p><p>If not one of the above conditions Unspecified 1. We receive and store certain types of information whenever you interact with us.</p></div>			</div>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<monogr>
		<title level="m" type="main">Polisis: Automated analysis and presentation of privacy policies using deep learning</title>
		<author>
			<persName><forename type="first">H</forename><surname>Harkous</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Fawaz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Lebret</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Schaub</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><forename type="middle">G</forename><surname>Shin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Aberer</surname></persName>
		</author>
		<idno>CoRR, abs/1802.02561</idno>
		<imprint>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">The biggest lie on the internet: Ignoring the privacy policies and terms of service policies of social networking services</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">A</forename><surname>Obar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Oeldorf-Hirsch</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Communication &amp; Society</title>
		<imprint>
			<biblScope unit="page" from="1" to="20" />
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
	<note>Information</note>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Privacyguide: Towards an implementation of the eu gdpr on internet privacy policy evaluation</title>
		<author>
			<persName><forename type="first">W</forename><forename type="middle">B</forename><surname>Tesfay</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Hofmann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Nakamura</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Kiyomoto</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Serna</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Fourth ACM International Workshop on Security and Privacy Analytics, IWSPA &apos;18</title>
				<meeting>the Fourth ACM International Workshop on Security and Privacy Analytics, IWSPA &apos;18<address><addrLine>New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2018">2018</date>
			<biblScope unit="page" from="15" to="21" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<monogr>
		<title level="m" type="main">The eu general data protection regulation (gdpr). A Practical Guide</title>
		<author>
			<persName><forename type="first">P</forename><surname>Voigt</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Von</surname></persName>
		</author>
		<author>
			<persName><surname>Bussche</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2017">2017</date>
			<publisher>Springer International Publishing</publisher>
			<pubPlace>Cham</pubPlace>
		</imprint>
	</monogr>
	<note>1st Ed</note>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">The creation and analysis of a website privacy policy corpus</title>
		<author>
			<persName><forename type="first">S</forename><surname>Wilson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Schaub</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">A</forename><surname>Dara</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Cherivirala</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><forename type="middle">G</forename><surname>Leon</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">S</forename><surname>Andersen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Zimmeck</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><forename type="middle">M</forename><surname>Sathyendra</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><forename type="middle">C</forename><surname>Russell</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics</title>
		<title level="s">Long Papers</title>
		<meeting>the 54th Annual Meeting of the Association for Computational Linguistics</meeting>
		<imprint>
			<date type="published" when="2016">2016</date>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="page" from="1330" to="1340" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Privacycheck: Automatic summarization of privacy policies using data mining</title>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">N</forename><surname>Zaeem</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">L</forename><surname>German</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><forename type="middle">S</forename><surname>Barber</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">ACM Trans. Internet Technol</title>
		<imprint>
			<biblScope unit="volume">18</biblScope>
			<biblScope unit="issue">4</biblScope>
			<biblScope unit="page">18</biblScope>
			<date type="published" when="2018-08">Aug. 2018</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
