<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">On the Impact of User Movement Simulations in the Evaluation of LBS Privacy-Preserving Techniques</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Sergio</forename><surname>Mascetti</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">DICo</orgName>
								<orgName type="institution">Università di Milano</orgName>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Dario</forename><surname>Freni</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">DICo</orgName>
								<orgName type="institution">Università di Milano</orgName>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Claudio</forename><surname>Bettini</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">DICo</orgName>
								<orgName type="institution">Università di Milano</orgName>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">X</forename><forename type="middle">Sean</forename><surname>Wang</surname></persName>
							<affiliation key="aff1">
								<orgName type="department">Dept. of CS</orgName>
								<orgName type="institution">University of Vermont</orgName>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Sushil</forename><surname>Jajodia</surname></persName>
							<affiliation key="aff2">
								<orgName type="department">CSIS</orgName>
								<orgName type="institution">George Mason University</orgName>
							</affiliation>
						</author>
						<title level="a" type="main">On the Impact of User Movement Simulations in the Evaluation of LBS Privacy-Preserving Techniques</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">B991661D03DDC2AF550574E8C9BB6098</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T02:58+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>The evaluation of privacy-preserving techniques for LBS is often based on simulations of mostly random user movements that only partially capture real deployment scenarios. We claim that benchmarks tailored to specific scenarios are needed, and we report preliminary results on how they may be generated through an agent-based contextaware simulator. We consider privacy preserving algorithms based on spatial cloaking and compare the experimental results obtained on two benchmarks: the first based on mostly random movements, and the second obtained from the context-aware simulator. The specific deployment scenario is the provisioning of a friend-finder-like service on weekend nights in a big city. Our results show that, compared to the contextaware simulator, the random user movement simulator leads to significantly different results for a spatial-cloaking algorithm, under-protecting in some cases, and over-protecting in others.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>Location-based services (LBS) are often cited as killer applications for the latest GPS-equipped 3G phones. These phones are slated to be massively distributed in 70 countries. While car navigation and identification of nearest points of interest are already widely used services, more interest are generating the socalled friend-finder services as a class of LBS that will change once more our way to interact. A friend-finder service reveals to a participating user the presence of other close-by participants belonging to a particular group (friends is only one example), possibly showing their position on a map. From a technical point of view, in contrast to services that find nearest points of interests, this service is characterized by a sequence of LBS requests instead of single ones, since a user may want to periodically check, while moving or even while staying in the same place, for close-by participants.</p><p>Sociological studies have shown that a large number of users perceive the release of their precise location, as part of a LBS request, as a possible privacy threat <ref type="bibr" target="#b0">[1]</ref>. Considering friend-finder services it is easy to identify two types of privacy threats: a) the association of the identity of the user with the specific group of persons he is interested in may reveal her religious, sexual, or political orientation, and b) the association of the identity of the user with her precise location may reveal what kind of places she has been to, or that she has not been where she was supposed to be at that time.</p><p>As formally shown in <ref type="bibr" target="#b1">[2]</ref> the likelihood of a privacy violation, and consequently the defense techniques to be applied, strongly depend on the knowledge that an adversary may have. In the friend-finder service the service provider (SP) may not be trusted or the communication channels may be insecure; then, the adversary's knowledge may include the precise identity and location information submitted with each request, and both the above privacy threats would become real privacy breaches. The substitution of identities with pseudonyms does not entirely solve the problem, if, for example, the adversary happens to know who is at the location at the time reported in the request (e.g., in the case the issuer of the request uses a fidelity card at a store). In some cases, the adversary may also be able to recognize sequences of requests as issued by the same anonymous user (e.g., by observing the same pseudonym or by spatio-temporal tracking) and use this information to re-identify the issuer.</p><p>Several defense techniques against both threats under different adversary models have been proposed, and may be applied to the friend-finder service; however, current proposals very rarely have formal assessments of the provided privacy preservation, and are generally supported by experimental results based either on real datasets of questionable significance for real LBS services (i.e., trucks or school bus traces) or on data simulations based on mostly random user movements that hardly match the specific deployment scenario of a LBS service.</p><p>In order to understand if the use of simulations based on mostly random user movements may be a real problem, or if it is actually useful and safe to use these simulations, we considered a typical deployment scenario for a friend-finder service: a large number of young people using the service on a weekend night in a large city like Milan, Italy. We performed a deep study, using different sources, including on-line surveys, of the parameters characterizing this scenario. We then used the Brinkhoff simulator <ref type="bibr" target="#b2">[3]</ref>, widely used in testing LBS privacy preservation, to generate, based on the parameters, a first dataset of user movements. A second dataset was created with a personalized version of the Siafu agent-based contextaware simulator <ref type="bibr" target="#b3">[4]</ref> which is able to capture much more details of our scenario. Then, based on a common metric for privacy and quality of service evaluation, we run a large number of tests on both datasets, considering different abilities of reidentification by the adversary, as well as different privacy preserving techniques.</p><p>Our results consistently show that (i) in some cases the evaluation on random movement simulations leads to the definition of overprotective techniques and (ii) in other cases, the techniques that are shown to meet privacy requirements based on those simulations do not meet them when tested with more realistic context-aware simulations.</p><p>We focus our technical treatment on protecting the association of the user with the request he has issued (e.g., with the group of people he is interested in, as in threat (a) described above), even if we believe that our arguments can be easily extended to techniques only aimed to protect the location.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Related work</head><p>We are not aware of related work in this area considering specifically the relevance of realistic simulations in LBS. There are however several studies on user movements with impact on many different application areas including epidemiology, transportation, computer networks, marketing, as well as LBS. A very interesting study supporting an argument against random movement simulations recently appeared <ref type="bibr" target="#b4">[5]</ref>. In the following we briefly report the main techniques currently proposed for LBS privacy preservation, identifying the ones similar to those tested in our experiments, and the ones using simulations to generate the datasets for experiments.</p><p>Privacy preserving solutions based on cryptographic techniques that totally hide the location information in requests, even to the SP, have been recently proposed <ref type="bibr" target="#b5">[6]</ref> for LBS based on 1-NN queries, and may be probably adapted for the service we consider. If proven to be correct, no simulation would be needed for these techniques since no information would leak from any request and the above privacy threats do not apply. However, this adaptation is still to be investigated, and there are some general concerns with these approaches regarding efficiency and flexibility.</p><p>A popular alternative technique is spatial cloaking, consisting in the generalization of the spatial information transmitted to the SP as part of a service request. By receiving generalized locations, the SP can only return approximate results on the presence of close-by group members and their positions; while it may be possible to have a trusted entity in the middle filtering the communication and improving the precision, the related overhead costs should be taken into account in evaluating the trade-off between generalization and quality of service. While in this paper we consider techniques based on spatial cloaking as in <ref type="bibr">[7-9, 2, 10]</ref>, other proposals have considered different techniques, including the generation of dummy requests, the use of incremental requests, or the substitution in the request of the position of the issuer with a region that does not include her (see among others <ref type="bibr" target="#b10">[11]</ref><ref type="bibr" target="#b11">[12]</ref><ref type="bibr" target="#b12">[13]</ref>).</p><p>Most of the proposals for LBS privacy have only considered requests in isolation while a few have also addressed the cases in which sequences of requests can be exploited by the adversary ( <ref type="bibr" target="#b13">[14,</ref><ref type="bibr" target="#b14">15]</ref> among others), as in the friend-finder service. A related problem is privacy-aware publication of trajectories <ref type="bibr" target="#b15">[16,</ref><ref type="bibr" target="#b16">17]</ref>; even if this has some aspects more similar to database publication than to service request privacy preservation, we believe that our results may be important for these studies as well.</p><p>Synthetic, mostly random, user movements obtained by the Brinkhoff simulator or other simulators have been used in most of the above cited papers as well as in our own previous work.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Organization</head><p>The rest of the paper is organized as follows. In Section 2 we formally define how we evaluate the privacy of LBS requests, or equivalently, how we measure the risk of a privacy violation upon issuing a request. In Section 3 we explain how the two datasets were obtained from the generators based on the parameters characterizing the deployment scenario. In Section 4 we briefly explain the privacy preservation algorithms being used and we report our experimental results. Section 5 concludes the paper.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">Privacy metric of generalized requests</head><p>As mentioned in the introduction, we are concerned with privacy protection via location generalization (also called spatial cloaking). In this section, we formalize the adversary model we consider in this paper, and give a metric to measure the privacy provided by a set of generalized requests against the adversary.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1">Requests, original requests, and generalized requests</head><p>We first formally define requests and generalized request for LBS. A request issued by a user without alteration is called an original request, and a generalized request is one that is sent to the service provider and has been altered from the original one for the purpose of privacy protection. Both kinds of requests are called requests and denoted r. A convention in this paper is to use r to denote generalized requests to emphasize its generalized nature, while use r to denote original requests, if not specified otherwise.</p><p>Either the client software or a trusted medium transforms (or generalizes) an original request to a generalized one. In this paper, we are not concerned about how the generalization has happened, but rather on the resulting generalized requests and their privacy properties. In the experimental section, we evaluate the performance of generalization algorithms based on the generalized requests they generate.</p><p>Each LBS request r, either original or generalized, is logically divided into three parts: IDdata, ST data, and SSdata, containing user identification data, location and time of the request, and other service parameters, respectively. In the sequel, the spatial and temporal components in ST data are denoted with Sdata and T data, respectively. In this paper, for the sake of simplicity, we consider space and time as discrete domains. However, our results can be easily extended to the case in which these two domains are continuous.</p><p>Each generalized request r must correspond to an original request r such that the difference between r and r is only in SData and furthermore, r.Sdata ⊆ r .Sdata, i.e., the spatial region of the generalized request must contain (or be equal to) the spatial region of the original request<ref type="foot" target="#foot_0">4</ref> . We use issuer(r) to denote the actual issuer of the (original or generalized) request r.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2">Adversary model</head><p>The objective here is to provide an adversary model that captures a general class of adversary models. In a sense, our adversary model is an adversary "metamodel". This adversary meta-model concerns two aspects of knowledge that an adversary might have: (1) knowledge of users' whereabouts (i.e., their locations), and (2) correlation of (generalized) requests. These two aspects cover the (explicit or implicit) assumptions appeared in the relevant literature.</p><p>For users' locations, we assume that the adversary has the knowledge expressed as the following Ident function:</p><p>Ident t : the Areas −→ the U ser sets, that is, given an area A, Ident t (A) is the set of users whom, through certain means, the adversary has identified to be located in area A at time t. In the following, when no confusion arises, we omit the time instant t. We further assume that this knowledge is correct in the sense that these identified users in reality are indeed in area A at the time.</p><p>For a given user i, if there exists an area A such that i ∈ Ident(A), then we say i is identified by the adversary. Furthermore, we say that i is identified in A. Note that there may be users who are also in A but the adversary does not identify them. This may happen either because the adversary is not aware of the presence of users in A, or because the adversary cannot identify these users even if he is aware of their presence. We do not distinguish these two cases in our adversary model as we shall see later that the distinction of the two cases does not make any perceptible difference in the ability of the adversary when the total population is large.</p><p>Clearly, in reality, there are lots of different sources of external information that can lead the adversary to estimate the location of users. Some may lead the adversary to know that a user is in a certain area, but not the exact location. For example, an adversary may know that Bob is in a pub (due to his use of a fidelity card at the pub), but may not know which room he is in. Some statistical analysis may be done to derive the probability that Bob is in a particular room, but this is beyond the scope of this paper.</p><p>The most conservative assumption regarding this capability of the adversary is that Ident(A) will give exactly all the users for each area A. It can be seen that if the privacy of the user is guaranteed in this most conservative assumption, then privacy is also guaranteed against any less precise Ident function. However, this conservative assumption is unlikely true in reality, while some observed that this assumption degenerates the quality of service unnecessarily. It will be interesting to see how much privacy and quality of service change with more realistic Ident functions. This is partly the goal of our paper.</p><p>As part of this adversary model regarding the location and users, we also assume another function:</p><formula xml:id="formula_0">N um t : the Areas −→ [0, ∞),</formula><p>that is, given an area A, N um t (A) gives an estimate of the number of users in the area at time t. This function can be derived from statistical information publicly available or through some kind of counting mechanism such as tickets to a theater. Again, when no confusion arises, we do not indicate the time instant t.</p><p>The second part of the adversary model is his ability to correlate requests. We formalize this with the following function L: L : the Requests −→ the Request sets, that is, given a (generalized) request r , L(r ) gives a set of requests such that the adversary has concluded, through certain means, are issued by the same user who issued the request r . In other words, all the requests in L(r ) are linked to r , although the adversary may still not know who the user is.</p><p>Note that L(r) may only give an (often small) subset of all the requests issued by the issuer of r. On the other hand, we assume that the function L is correct in the sense that each request in L(r) is indeed issued by the same user in reality. A set of requests is called a trace, denoted τ , if from the link function L we understand that all requests are issued by the same user. The requests in τ are implicitly ordered along the time dimension.</p><p>As in the case for Ident function, the most conservative assumption on correlation is that L(r) gives exactly all the (generalized) requests that are issued by the issuer of r. This is a very strong assumption that may lead to severely decrease quality of service when accompanied with the most conservative assumption about the Ident function. Again, a partial goal of this paper is to study the impact of a less conservative but more realistic assumption on L.</p><p>In <ref type="bibr" target="#b1">[2]</ref>, we proposed a formal framework to model LBS privacy attacks and defenses for the static case. The main idea is that the safety of a defense technique can be formally evaluated only if the context, i.e., the assumptions about the adversary's external knowledge, is explicitly stated. Following this methodology, in this paper, a context C H is given by three functions Ident, N um, and L, that is C H = (Ident, N um, L).</p><p>In the next section, we formalize the attack on the generalized requests that an adversary can perform in a context C H . A consequence of restricting to context C H is that, analogously to the related work in this area, we focus our attention on using only ST data as a quasiidentifier. Intuitively, a quasi-identifier in a request is a combination of values that can be used to provide more information on who the actual issuer of a request may be than without these values. For example, if the Ident function is given, the ST Data in the request is a quasi-identifier as it may provide information on the actual issuer, as shown in the next subsection. In principle, any information contained in a request should be carefully analyzed to see if it may serve as a quasi-identifier. For example, the IDdata part is an obvious target, and some service specific parameters may be used to link the request to users. However, these aspects are outside the scope of this paper.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.3">Privacy Evaluation</head><p>The general question for this subsection is, given a set of generalized requests and a context C H , how much privacy the users who issued these requests have.</p><p>We want to find the following function:</p><formula xml:id="formula_1">Att : the Request set × the U sers −→ [0, 1],</formula><p>Intuitively, given a (generalized) request r and a user i, Att(r , i) gives the probability that the adversary can derive from C H that i is the issuer of r among all the users. In the following of this section we show how to specify the attack function for context C H . Once the attack function is specified, we can use the following formula to evaluate the privacy value of a request:</p><formula xml:id="formula_2">P rivacy(r ) = 1 − Att C H (r , issuer(r ))<label>(1)</label></formula><p>Intuitively, this value is the probability that the attacker will not associate the issuer of request r to r . In order to specify the Att function, we introduce the function Inside(i, r ) that indicates the probability of user i to be located in r .Sdata at the time of the request. Intuitively, Inside(i, r ) = 1 if user i is identified by the adversary as one of the users that are located in r .Sdata at time r .T data, i.e., i ∈ Ident t (r .Sdata) when t = r .T data. On the contrary, Inside(i, r ) = 0 if i is recognized by the adversary as one of the users located outside r .Sdata at time r .T data, i.e., there exists an area A with A ∩ r .Sdata = ∅ such that i ∈ Ident(A). Finally, if neither of the above cases hold, then the adversary does not know where i is. There is still a probability that i is in r .Sdata. Theoretically, this probability is the number of users in r .Sdata that are not recognized by the adversary (i.e., N um(r .Sdata) − |Ident(r .Sdata)|) divided by all the users who are not recognized by the adversary anywhere (i.e., |I| − |Ident(Ω)|, where I is the set of all users, and Ω is the entire area for the application). Formally,</p><formula xml:id="formula_3">Inside(i, r ) =    1 if i ∈ Ident(r .Sdata) 0 if ∃A : A ∩ r .Sdata = ∅ and i ∈ Ident(A) N um(r .Sdata)−|Ident(r .Sdata)| |I|−|Ident(Ω)| otherwise<label>(2)</label></formula><p>We can now define the Att function in context C H . For the sake of presentation, let us first consider the attack in the snapshot context</p><formula xml:id="formula_4">C snap = (Ident, N um, L snap ),</formula><p>where for each generalized request r , L snap (r ) = {r }. In this special case, the probability of a user i of being the issuer of r is given by the probability of i being in r .Sdata at the time of the request, normalized among all the users in I. Formally, the attack can be defined as:</p><formula xml:id="formula_5">Att Csnap (r , i) = Inside(i, r ) i ∈I Inside(i , r )<label>(3)</label></formula><p>When the total population of users is large (relative to the number of users whose locations are known to the adversary), then the "otherwise" case in Formula 2 is very small, albeit nonzero. Intuitively, if a user i falls into this case, then the adversary cannot really distinguish this particular user from all other users who also fall into this case. For such a user i, we can simply give a value 1/|I| to Att Csnap (r , i). We could give 1/(|I| − N um(Ω)), but this does not make much impact in practice. Now it's easy to see that</p><formula xml:id="formula_6">Att Csnap (r , i) ≈    1/N um(r .Sdata) if Inside(i, r ) = 1 0 if Inside(i, r ) = 0 1/|I| otherwise<label>(4)</label></formula><p>The above formula makes intuitively sense. Indeed, if i is recognized as inside r .Sdata, without any other information, the adversary cannot distinguish him/her from any of the N um(r .Sdata) people in the area who might be the issuer. If i is recognized outside, then clearly i cannot be the issuer due to our definition of (generalized) requests. If i is not recognized anywhere (meaning he/she can be anywhere), then the attacker cannot distinguish him/her from any of the other people who are not recognized. Since we assume the total population is much greater than N um(Ω), the probability that i is the issuer is close to 1/|I|.</p><p>Example 1. Consider the situation shown in Figure <ref type="figure" target="#fig_1">1</ref>(a) in which there is the request r such that, at time r .T data, there are three users in r .Sdata: one of them is identified as i 1 , the other two are not identified. The adversary can also identify users i 2 and i 3 outside r .Sdata at time r .T data. Assume that the set I contains 100 users.  Clearly, i 2 and i 3 have zero probability of being the issuers, since they are identified outside r .Sdata and due to the assumption that the spatial region of any generalized request must contain the spatial region of the original request. On the contrary, the adversary is sure about the fact that i 1 is located in r .Sdata. By Equation 3, the attack associates i 1 to r with likelihood 1/( i ∈I Inside(i , r )). By Formula 2, for each user i in I \ {i 1 , i 2 , i 3 }, Inside(i, r ) = 2/100. Therefore, i ∈I Inside(i , r ) = 97 * 2/100 + 1 ≈ 3. Consequently, the probability of i 1 to be the issuer of r is approximately 1/3. Moreover, each user i ∈ I \ {i 1 , i 2 , i 3 } has a probability to be the issuer of about (2/100)/3 = 2/300.</p><p>In the general case L(r ) ⊇ {r }, we can evaluate, analogously to the snapshot case, the probability that a user is located in the generalized region of all the requests in the trace τ = L(r ). So, we can extend the Inside function to traces where, given a trace τ and a user i, Inside(i, τ ) is the probability that user i is located, for each request r in τ , in r .ST data. Then, the attack is</p><formula xml:id="formula_7">Att C H (r , i) = Inside(i, L(r )) i ∈I Inside(i , L(r ))<label>(5)</label></formula><p>We now turn to consider how to compute Inside(i, τ ). First consider some easy cases. If i ∈ Ident(r ) for all requests r ∈ τ , then Inside(i, τ ) = 1. If i ∈ Ident(A) and A ∩ r .Sdata = ∅ for an area A and at least one requests r ∈ τ , then Inside(i, τ ) = 0.</p><p>The rest of cases are difficult ones. To calculate Inside(i, τ ), we need to consider the likelihood of someone moving from one location to/from another in the specific times. In this paper, we advocate the following as a reasonable approach. We assume for each pair of locations A and B and two times t 1 and t 2 , we know the probability of a user i being in B at time t 2 conditioned on the fact that the user is in A at time t 1 . In formalism, consider two random variables X: "i is inside A at time t 1 " and Y : "i is inside B at time t 2 ", where A and B are two areas and t 1 and t 2 are two different times. We assume the adversary knows the value P (Y |X).</p><p>We note that P (Y |X) in general is not the same as P (Y ). Indeed, how likely user i is in B can depend on how likely the same user is in A. Take two extreme examples: if A and B are very far away and t 1 and t 2 are close to each other, then i cannot be in B at t 2 if i is in A at t 1 , i.e., P (Y |X) ≈ 0. On the other hand, if A and B are just two locations along a one-way road and the difference between times t 1 and t 2 matches the time needed to move from A to B with a normal moving speed, then P (Y |X) ≈ 1. In practice, this value can be derived from historical observations and experiences. Now, assume τ consists of the requests r 1 , . . . , r k . We form a Bayesian network for each user i with X 1 , . . ., X k as the nodes, where each X j corresponds to the random variable: "user i is in r j .Sdata at time r j .T data". In this network, for each node X h that satisfies the condition (denoted c) i ∈ Ident t (r h .Sdata) with t = r h .T data, we draw an arc towards each other node X h which does not satisfy condition C. In addition, for each pair of nodes r h and r h such that neither satisfy condition c, we draw an arc from X h to X h if the r h .T data &lt; r h .T data. (The resulting network is acyclic.) As we have assumed, we know the value P (X h |X h ) for each arc X h to X h . Denote by E the conjunctive fact that P (X h ) = 1 for each r h ∈ τ that satisfies condition c. What we want to find is P (X 1 , . . . , X k |E) = Inside(i, τ ). This is a well-studied belief revision problem, and many computation and approximation methods exist. (Note that if we apply this method to the easier cases mentioned earlier, we would arrive at the correct values.)</p><p>Example 2. Continue from Example 1 and assume a second request r (see Figure <ref type="figure" target="#fig_1">1(b)</ref>) is issued after r and that r is linked with r , so τ consists of these two requests. At time r .T data, there are 4 users inside r .Sdata, two of which are identified as i 1 and i 2 . No user is identified outside r .Sdata. From the above discussion, it follows that Inside(i 2 , τ ) = Inside(i 3 , τ ) = 0 since i 2 and i 3 are identified outside the first generalized request r . All the other users have a non-zero probability of being inside the generalized region of each request in the trace. In particular, Inside(i 1 , τ ) = 1 since i 1 is recognized in both requests. Consider a user i ∈ I \ {i 1 , i 2 , i 3 }, and denote X and Y being the assertion that "i is in r .Sdata at time r .T data" and "i is in r .Sdata at time r .T data". In this case, the Bayesian network for i has two nodes X r and X r , and there is an arc from X r to X r since r is issued after r is. Now let us assume P (X r |X r ) = 0.75, i.e., there is a 75% likelihood that someone in r .Sdata will move to r .Sdata at the specified times. Now compute Inside(i, τ ) = P (X r , X r ) = P (X r )P (X r |X r ) = 2/97 * 0.75. Now the sum of all the Inside(j, τ ) value is 1 + 0 + 0 + 97 * 2/97 * 0.75 = 2.5. The attack value under these assumptions then is as follows: For Att C H (r , i 1 ) = 1/2.5 = 40%, Att C H (r , i 2 ) = Att C H (r , i 3 ) = 0, and Att C H (r , i) = (2/97) * .75/2.5 ≈ 0.6% for all other 97 users i.</p><p>To make the situation more interesting, let us remove the fact that i 2 was recognized outside at time r .T data, and we want to figure out the value Inside(i 2 , τ ). In this case, let us assume P (X r |X r ) = 0.75, namely people who are in r .Sdata have a 75% likelihood to be from r .Sdata. Under the fact E that i 2 is in r .Sdata, then we know Inside(i 2 , τ ) = P (X r , X r |E) = 0.75. Then the sum of Inside values is 1+0.75+0+97 * 2/97 * 0.75 = 3.25. Hence, Att C H (r , i 1 ) ≈ 31%, Att C H (r , i 2 ) ≈ 23%, Att C H (r , i 3 ) = 0, and Att C H (r , i) = (2/97) * 0.75/3.25 ≈ 0.47% for each other 97 users i. This is an interesting exercise as it reveals that if we add i 2 to be possibly in r .Sdata (with 75% probability), then the likelihood that i 1 is the issuer decreases, which is intuitively correct.</p><p>It is worth noting that the definition of attack in context C H is a proper extension of the attack that can be defined in the conservative context in which the adversary knows the location and the identity of each user in each time instant. The historical attack in this context was first proposed in <ref type="bibr" target="#b13">[14]</ref>. The idea is that the only users that have non-zero probability of being the issuer of a trace of requests are those whose spatio-temporal location is contained in the generalized region of every request in the trace. It can be easily seen that, if each user can be identified at each time instant, then the Inside() function returns either 0 or 1 and hence the attack we specified for context C H assigns a zero probability to each user that is located outside the generalized region of any request in the trace.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">The MilanoByNight simulation</head><p>In order to evaluate privacy-preserving techniques applied to LBS, a dataset of users' movements is needed. In our experiments, we want to focus on privacy threats that arise when using a friend finder service, as described in Section 1. We suppose that this kind of service is primarily used by people during entertainment hours, especially at night. Therefore, the ideal dataset for our experiments should represent movements of people on a typical Friday or Saturday night in a big city, when users tend to move to entertainment places. To our knowledge, currently there are no datasets like this publicly available, specially considering that we want to have large scale, individual, and precise location data (i.e., with the same approximation of current consumer GPS technology). In this section we describe how we generated this user movement dataset.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1">Relevant Parameters</head><p>For our experiments we want to artificially generate movements for 100, 000 users on the road network of Milan<ref type="foot" target="#foot_1">5</ref> . The total area of the map is 324 km 2 , and the resulting average density is 308 users/km 2 . Very detailed digital vector maps of the city have been generously provided by the municipality of Milan. The simulation includes a total of 30, 000 home buildings and 1, 000 entertainment places; the first value is strictly related to the considered number of users, while the second is based on real data from public sources which also provide the geographical distribution of the places. Our simulation starts at 7 pm and ends at 1 am. During these hours, each user moves from house to an entertainment place, spends some time in that place, and possibly moves to another entertainment place or go back home.</p><p>All probabilities related to users' choices are modeled with a probability distributions. For this specific data generation, some of the important parameters of the simulation are:</p><p>-Source and destination. These are the locations essential to define movements. They may be homes or entertainment places. Some places in some districts are more popular than others. -StartingTime. The time at which a user leaves her home to go to the first entertainment place. -Permanence. How long will a user stay at one entertainment place? -NumPlaces. How many entertainment places will a user visit on one night?</p><p>In order to have a realistic model of these distributions, we prepared a survey to collect real users data. We are still collecting data, but the current parameters are based on interviews of more than 300 people in our target category.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2">Weaknesses of mostly random movement simulations</head><p>Many papers in the field of privacy preservation in LBS use artificial data generated by moving object simulators to evaluate their techniques. However, most of the simulators are usually not able to reproduce a realistic behavior of users. For example, objects generated by the Brinkhoff generator <ref type="bibr" target="#b2">[3]</ref> cannot be aggregated in certain places (e.g., entertainment places). Indeed, once an object is instantiated, the generator chooses a random destination point on the map; after reaching the destination, the object disappears from the dataset. For the same reason, it is not possible to reproduce simple movement patterns (e.g.: a user going out from her home to another place and then coming back home), nor to simulate that a user remains for a certain time in a place.</p><p>Despite these strong limitations, we made our best effort to use the Brinkhoff simulator to generate a set of user movements with characteristics as close as possible to those explained in Section 3.1. For example, in order to simulate entertainment places, some random points on the map, among those points on the trajectories of users, were picked. The simulation has the main purpose of understanding if testing privacy preservation over random movement simulations gives significantly different results with respect to more realistic simulations.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.3">Generation of user movements with a context simulator</head><p>In order to obtain a dataset consistent with the parameters specified in Section 3.1, we need a more sophisticated simulator. For our experiments, we have chosen to customize the Siafu context simulator <ref type="bibr" target="#b3">[4]</ref>. With a context simulator it is possible to design models for agents, places and context. Therefore, it is possible to define particular places of aggregation and make users dynamically choose which place to reach and how long to stay in that place. In our simulation homes are distributed almost uniformly on the map, with a minor concentration on the central zones of the city. Entertainment places are mostly concentrated in 5 zones of the city.</p><p>The distributions for StartingTime, Permanence and NumPlaces parameters introduced in Section 3.1 were modeled with the results of the survey. For example, the time of permanence in an entertainment place was modeled according to the following percentages derived from the survey: 9.17% of the users stays less than 1 hour, 34.20% stays between 1 and 2 hours, 32.92% stays between 2 and 3 hours, 16.04% stays between 3 and 4 hours, and 7.68% stays more than 4 hours.</p><p>Following these parameters, in our dataset users spend 50.87% of the time at home, 7.28% of the time moving from one place to another and 41.85% of the time in entertainment places. When a user moves from one place to another, she decides whether to go on foot or by car. In general, if an entertainment place is farther than 500 meters, people tend to move by car, and this is reflected in the simulation. The average speed of movements by car is 20 km/h, while the average speed on foot is 3.6 km/h. With our parameters 10.64% of movements are done on foot, while all the others are done by car.</p><p>In this section we show the results of our experimental evaluation. We first define how we evaluate the quality of service in Section 4.1, then we describe the experimental setting in Section 4.2 and the generalization algorithms we used in Section 4.3. Finally, in Sections 4.4 and 4.5 we show the impact of the simulation parameters and of the user movements, respectively, in the evaluation of the generalization algorithms.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1">Evaluation of the Quality of Service</head><p>Different metrics can be defined to measure QoS for different kind of services. For instance, for the friend-finder service we are considering, it would be possible to measure how many times the generalization leads the SP to return an incorrect result i.e., the issuer is not notified of a close-by friend or, vice versa, the issuer is notified for a friend that is not close-by. While this metric is useful for this specific application, we want to measure the QoS independently from the specific kind of service. For this reason, in this paper we evaluate how QoS degrades in terms of the perimeter of the generalized region. If the generalized region is too large, the service becomes useless. For this purpose, we introduce a new parameter, called maxP , that indicates this threshold in terms of the maximum perimeter. We assume that no request is sent to the SP with a perimeter larger than maxP .</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2">Experimental settings</head><p>In our experiments we used two datasets of users movements. The dataset AB (Agent-Based) was generated with the customized Siafu simulator as described in Section 3.3, while the dataset M RM (Mostly Random Movement) was created with the Brinkhoff simulator as described in Section 3.2. In both cases, we simulate LBS requests for the friend-finder service by choosing random users in the simulation, we compute for each request the generalization according to a given algorithm, we evaluate QoS as explained in Section 4.1 and privacy according to formula (1) presented in Section 2.</p><p>The most important parameters that characterize the simulations are reported in Table <ref type="table">1</ref>, with the values in bold denoting default values. The number of users indicates how many users are in the simulation, and the simulations are designed so that this number remains almost constant at each time instant. In every two minutes, each user has a probability P req of issuing a request. For technical reasons, the reported tests are based on a time frame of three hours over the total six hours of the MilanoByNight scenario. This implies that in the default case we consider a total of 45 requests (one every four minutes of the considered time frame). The parameter P id−in indicates the probability that a user is identified when she is located in a entertainment place while P id−out is the probability that a user is identified in any other location (e.g., while moving from home to a entertainment place). While we also perform experiments where the two probabilities are the same, our scenario suggests as much more realistic a higher value for P id−in (it is considered ten times higher than P id−out ). This is due to the fact that restaurants, pubs, movie theaters, and similar places are likely to have different ways to identify people (fidelity or membership cards, wifi hotspots, cameras, credit card payments, etc.) and in several cases more than one place is owned by the same company that may have an interest in collecting data about its customers.</p><p>Finally, P link indicates the probability that two consecutive requests can be identified as issued by the same user. <ref type="foot" target="#foot_2">6</ref> While we perform our tests considering a full range values, the specific default value reported in the table is due to a recent study on the ability of linking positions based on spatio-temporal correlation <ref type="bibr" target="#b17">[18]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Table 1. Parameter values</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Parameter</head><p>Values dataset AB, MRM number of users 10k, 20k, 30k, 40k, 50k, 60k, 70k, 80k, 90k, 100k Preq 0.1, 0.2, 0.3, 0.4, 0.5, 0.6, 0.7, 0.8, 0.9, 1.0 P id−in 0.1, 0.2, 0.3, 0.4, 0.5, 0.6, 0.7, 0.8, 0.9, 1.0 P id−out 0.01, 0.02, 0.03, 0.04, 0.05, 0.06, 0.07, 0.08, 0.09, 0.1</p><formula xml:id="formula_8">P link 0.1, 0<label>.2, 0.3, 0.4, 0.5, 0.6, 0.7, 0.8, 0.87, 0.9, 1.0</label></formula><p>The experimental results we show in this section are obtained by running the simulation for 100 issuers and then computing the average values.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.3">The Generalization Algorithms Used in the Experiments</head><p>In our experiments we evaluate the privacy and the QoS of requests generalized by using two algorithms previously proposed in the literature. The first one, called Grid, was presented in <ref type="bibr" target="#b18">[19]</ref>, and it is used as a representative of several algorithms aimed at guaranteeing k-anonymity in the snapshot case, i.e., these algorithms do not take into account link ability of the adversary. Intuitively, this particular algorithm partitions all users into blocks, each one having at least cardinality k. Then, it computes the generalized region as the minimum bounding rectangle (MBR) that covers the location of the users in the same block as the issuer.</p><p>The second algorithm, Greedy, was first proposed in <ref type="bibr" target="#b13">[14]</ref> and a similar idea was also described in <ref type="bibr" target="#b14">[15]</ref>. The use of Greedy is intended to represent algorithms aimed at preserving privacy in the historical case, i.e., the general C H context, assuming that the attacker may actually obtain and recognize traces of requests from the same issuer. This algorithm computes the generalization of the first request r in a trace using an algorithm for the snapshot case. While doing this, the set A of users located in the generalized region is stored. The generalized regions of the successive request r linked with r is then computed as the MBR of the location of the users in A at the time of r . In our implementation we use Grid as the snapshot algorithm to compute the generalization of the first request.</p><p>For the purpose of our tests, we modified the two algorithms above so that each generalized region has a perimeter always smaller than maxP . To achieve this, if the perimeter of the generalized region is larger than maxP , then the region is iteratively shrunk, until its perimeter is below maxP , by excluding from the MBR the user that is farther from the issuer. In the Greedy algorithm, when a user is excluded from the generalized region, then it is also excluded from the set A of users, and hence he is not used in the generalization of the successive requests. Eventually, when the set A contains the issuer only, a snapshot generalization is executed again and A is reinitialized.</p><p>In addition to the input request r, and the location of all the users in the system, the considered algorithms require two additional parameters: the value k, and the threshold maxP . In our tests, we used values for k between 10 and 60 (default is 10) and values for maxP between 1000 to 4000 meters (default is 1000 meters).</p><p>In our experimental results we also evaluated the privacy threat when no privacy preserving algorithm is applied. The label NoAlg is used in the figures to identify results in this particular case.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.4">Impact of Simulation Parameters in the Evaluation of the Generalization Algorithms</head><p>The objective of the first set of experimental results we present is to show which parameters of the simulation affect most the evaluation of the generalization algorithms. In these tests we used the AB dataset only. Figure <ref type="figure" target="#fig_3">2</ref>(a) shows that the average privacy obtained with Greedy and Grid is not significantly affected by the size of the total population. Indeed, both algorithms, independently from the total number of users, try to have generalized regions that cover the location of k users, so the privacy of the requests is not affected. However, when the density is high, the two algorithms can generalize to a small area, while when the density is low, a larger area is necessary to cover the location of k users (see Figure <ref type="figure" target="#fig_3">2(b)</ref>). On the contrary, the privacy obtained when no generalization is performed is significantly affected by the total population. Indeed, a higher density increases the probability of different users to be in the same location and hence it increases privacy also if the requests are not generalized.</p><p>A parameter that significantly affects the average privacy is the probability of identification of a user in a certain place. In Figure <ref type="figure" target="#fig_4">3</ref> we show the experimental results for different values of P id−in when, in each test, P id−out is set to  P id−in /10. As expected, considering a trace of requests, the higher is the probability of identifying users in one or more of the regions from which the requests in the trace were performed, the smaller is the level of privacy. Figure <ref type="figure">4</ref>(a) shows the impact of P link on the average privacy. As expected, high values of P link lead to small values of privacy. Our results show that the relation between the P link and privacy is not linear. Indeed, privacy depends almost linearly on the average length of the traces identified by the adversary (Figure <ref type="figure">4(b)</ref>). However, the average length of the traces grows almost exponentially with the value of P link (Figure <ref type="figure">5</ref>).</p><p>To summarize the first set of experiments, our findings show that many parameters of the simulation significantly affect the evaluation of the generalization algorithms. This implies that when a generalization algorithm is evaluated it is necessary to carefully estimate realistic values for the parameters of the simulation. Indeed, an error in the estimation may lead to misleading results.  </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.5">Impact of the User Movements on the Evaluation of the Generalization Algorithms</head><p>The objective of the second set of experiments is to answer an important question posed in this paper: what is the impact of the different simulated user movements on the evaluation of the Generalization Algorithms? We answer to this question with a set of tests performed on the two different datasets we obtained as described above.</p><p>The first set of tests, reported in in Figure <ref type="figure" target="#fig_8">6</ref>, compares the privacy achieved by the Greedy algorithm on the two datasets for different values of k and for different values of QoS. The experiments on M RM were repeated trying also larger values for the QoS threshold (maxP = 2000 and maxP = 4000), so three different versions of M RM appear in the figures. In order to focus on these parameters only, in these tests the probability of identification was set to the same value for any place (P id−in = P id−out = 0.1), and for the M RM dataset the issuer of the requests was randomly chosen only among those that stay in the simulation for 3 hours, ignoring the ones staying for much shorter time that inevitably are part of this dataset. This setting allowed us to compare the results on the two datasets using the same average length of traces identified by the adversary.</p><p>Figure <ref type="figure" target="#fig_8">6</ref>(a) shows that the average privacy of the algorithm evaluated on the AB dataset is much higher than on the M RM dataset. This is mainly motivated by the fact that in AB users tend to concentrate in a few locations (the entertainment places) and this enhances privacy. This is also confirmed by a similar test performed without using any generalization of locations; we obtained values constantly higher for the AB dataset (the average privacy is 0.67 in AB and 0.55 in M RM ).</p><p>In Figure <ref type="figure" target="#fig_8">6</ref>(b) we show the QoS achieved by the algorithm in the two datasets with respect to the average privacy achieved. This result confirms that the level of privacy evaluated on the AB dataset using small values of k and maxP for the algorithm cannot be observed on the M RM dataset even with much higher values for these parameters.</p><p>From the experiments shown in Figure <ref type="figure" target="#fig_8">6</ref> we can conclude that if the M RM dataset is used as a benchmark to estimate the values of k and maxP that are necessary to provide a desired average level of privacy, then the results will suggest the use of values that are over-protective. As a consequence, it is possible that the service will exhibit a much lower QoS than the one that could be achieved with the same algorithm.  The above results may still support the safety of using M RM , since according to what we have seen above a technique achieving a certain level of privacy may only do better in a real scenario. However, our second set of experiments shows that this is not the case.</p><p>In Figure <ref type="figure" target="#fig_10">7</ref> we show the results we obtained by varying the probability of identification. For this test, we considered two sets of issuers in the M RM data set. One set is composed by users that stay in the simulation for 3 hours, (M RM long traces, in Figure <ref type="figure" target="#fig_10">7</ref>), while the other contains issuers randomly chosen in the entire set of users (M RM all traces, in Figure <ref type="figure" target="#fig_10">7</ref>), hence including users staying in the simulation for a much shorter time.</p><p>In Figure <ref type="figure" target="#fig_10">7</ref>(a) and 7(b) we can observe that the execution on the M RM dataset leads to evaluate a privacy level that is higher than the one obtained on the AB dataset. In particular, the evaluation of the Grid algorithm using the M RM dataset (Figure <ref type="figure" target="#fig_10">7</ref>(b)), would suggest that the algorithm is able to provide a high privacy protection. However, when evaluating the same algorithm using the more realistic dataset AB, this conclusion seems to be incorrect. In this case, the evaluation on the M RM dataset may lead to underestimate the privacy risk, and hence to deploy services based on generalization algorithms that may not provide the minimum required level of privacy.  </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">Conclusions and open issues</head><p>In this paper we claim that the experimental evaluation of LBS privacy preserving techniques should be based on user movement datasets obtained through simulations tailored to the specific deployment scenario of the target services. Our results support our thesis for the class of LBS known as friend-finder services, for techniques based on spatial cloaking, and for adversary models that include the possibility for the adversary to occasionally recognize people in certain locations. We believe that these results can be generalized to other LBS, techniques and adversary models. For example, as a future work, it would be interesting to also evaluate some defense techniques that generalize the issuer's location to an area that does not necessarily contain the issuer's location. Moreover, in our experiments we only considered the first of the two privacy threats presented in the introduction. We do have some ideas on how to extend them to consider the second, location privacy, as well. Finally, we believe a significant effort should be devoted to the development of new flexible and efficient context-aware user movement simulators, as well as to the collection of real data, possibly even in an aggregated form, to properly tune the simulations. In our opinion this is a necessary step to have significant common benchmarks to evaluate LBS privacy preserving techniques.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head></head><label></label><figDesc>(a) First request, r .(b) Second request, r .</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Fig. 1 .</head><label>1</label><figDesc>Fig. 1. Example of attack</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_3"><head>Fig. 2 .</head><label>2</label><figDesc>Fig. 2. Performance evaluation for different values of the total population.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_4"><head>Fig. 3 .</head><label>3</label><figDesc>Fig. 3. Average privacy for different values of P id−in (P id−out = P id−in /10).</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_5"><head></head><label></label><figDesc>Average privacy as a function of P link . Average privacy as a function of the average trace length.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_6"><head>Fig. 4 .Fig. 5 .</head><label>45</label><figDesc>Fig. 4. Performance evaluation for different values of P link .</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_7"><head></head><label></label><figDesc>Average privacy as a function of k. Average privacy as a function of the average perimeter.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_8"><head>Fig. 6 .</head><label>6</label><figDesc>Fig. 6. Evaluation of the Greedy algorithm using AB and M RM data sets. P id−in = P id−out = 0.1</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_9"><head></head><label></label><figDesc>Grid algorithm.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_10"><head>Fig. 7 .</head><label>7</label><figDesc>Fig. 7. Average privacy using AB and M RM data sets. P id−out = P id−in /10.</figDesc></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="4" xml:id="foot_0">Here "region" can be a point.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="5" xml:id="foot_1">100, 000 is an estimation of the number of people participating in the service we consider.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="6" xml:id="foot_2">The limitation to consecutive requests is because in our specific scenario we assume linking is performed mainly through spatio-temporal correlation.</note>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgments</head><p>The authors would like to thank Stefano Varesi for his contribution in writing the code that was used for our simulations. This work was partially supported by National Science Foundation (NSF) under grant N. CNS-0716567, and by Italian MIUR under grant InterLink II04C0EC1D.</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Location-based services for mobile telephony: a study of users privacy concerns</title>
		<author>
			<persName><forename type="first">L</forename><surname>Barkhuus</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Dey</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. of the 9th International Conference on Human-Computer Interaction</title>
				<meeting>of the 9th International Conference on Human-Computer Interaction</meeting>
		<imprint>
			<publisher>IOS Press</publisher>
			<date type="published" when="2003">2003</date>
			<biblScope unit="page" from="709" to="712" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Anonymity in location-based services: towards a general framework</title>
		<author>
			<persName><forename type="first">C</forename><surname>Bettini</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Mascetti</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><forename type="middle">S</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Jajodia</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. of the 8th International Conference on Mobile Data Management</title>
				<meeting>of the 8th International Conference on Mobile Data Management</meeting>
		<imprint>
			<publisher>IEEE Computer Society</publisher>
			<date type="published" when="2007">2007</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">A framework for generating network-based moving objects</title>
		<author>
			<persName><forename type="first">T</forename><surname>Brinkhoff</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">GeoInformatica</title>
		<imprint>
			<biblScope unit="volume">6</biblScope>
			<biblScope unit="issue">2</biblScope>
			<biblScope unit="page" from="153" to="180" />
			<date type="published" when="2002">2002</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">A generic large scale simulator for ubiquitous computing</title>
		<author>
			<persName><forename type="first">M</forename><surname>Martin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Nurmi</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">3rd Annual International Conference on Mobile and Ubiquitous Systems: Networking &amp; Services</title>
				<imprint>
			<publisher>IEEE Computer Society</publisher>
			<date type="published" when="2006-07">July 2006</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Understanding individual human mobility patterns</title>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">C</forename><surname>Gonzalez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">A</forename><surname>Hidalgo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">L</forename><surname>Barabasi</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Nature</title>
		<imprint>
			<biblScope unit="volume">453</biblScope>
			<biblScope unit="page" from="779" to="782" />
			<date type="published" when="2008-06">June 2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Private queries in location based services: Anonymizers are not necessary</title>
		<author>
			<persName><forename type="first">G</forename><surname>Ghinita</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Kalnis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Khoshgozaran</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Shahabi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><forename type="middle">L</forename><surname>Tan</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. of SIGMOD</title>
				<meeting>of SIGMOD</meeting>
		<imprint>
			<publisher>ACM Press</publisher>
			<date type="published" when="2008">2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Anonymous usage of location-based services through spatial and temporal cloaking</title>
		<author>
			<persName><forename type="first">M</forename><surname>Gruteser</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Grunwald</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. of the 1st International Conference on Mobile Systems, Applications and Services (MobiSys)</title>
				<meeting>of the 1st International Conference on Mobile Systems, Applications and Services (MobiSys)</meeting>
		<imprint>
			<publisher>The USENIX Association</publisher>
			<date type="published" when="2003">2003</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">The new casper: query processing for location services without compromising privacy</title>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">F</forename><surname>Mokbel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">Y</forename><surname>Chow</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><forename type="middle">G</forename><surname>Aref</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. of the 32nd International Conference on Very Large Data Bases</title>
				<meeting>of the 32nd International Conference on Very Large Data Bases</meeting>
		<imprint>
			<publisher>VLDB Endowment</publisher>
			<date type="published" when="2006">2006</date>
			<biblScope unit="page" from="763" to="774" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Protecting location privacy with personalized k-anonymity: Architecture and algorithms</title>
		<author>
			<persName><forename type="first">B</forename><surname>Gedik</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Liu</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Transactions on Mobile Computing</title>
		<imprint>
			<biblScope unit="volume">7</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page" from="1" to="18" />
			<date type="published" when="2008">2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Preventing location-based identity inference in anonymous spatial queries</title>
		<author>
			<persName><forename type="first">P</forename><surname>Kalnis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Ghinita</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Mouratidis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Papadias</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Transactions on Knowledge and Data Engineering</title>
		<imprint>
			<biblScope unit="volume">19</biblScope>
			<biblScope unit="issue">12</biblScope>
			<biblScope unit="page" from="1719" to="1733" />
			<date type="published" when="2007">2007</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Protection of location privacy using dummies for location-based services</title>
		<author>
			<persName><forename type="first">H</forename><surname>Kido</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Yanagisawa</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Satoh</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. of the 21st International Conference on Data Engineering Workshops</title>
				<meeting>of the 21st International Conference on Data Engineering Workshops</meeting>
		<imprint>
			<publisher>IEEE Computer Society</publisher>
			<date type="published" when="2005">2005</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Spacetwist: Managing the tradeoffs among location privacy, query performance, and query accuracy in mobile services</title>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">L</forename><surname>Yiu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">S</forename><surname>Jensen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Huang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Lu</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. of the 24th International Conference on Data Engineering</title>
				<meeting>of the 24th International Conference on Data Engineering</meeting>
		<imprint>
			<publisher>IEEE Computer Society</publisher>
			<date type="published" when="2008">2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Location privacy protection through obfuscation-based techniques</title>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">A</forename><surname>Ardagna</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Cremonini</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Damiani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">D C</forename><surname>Di Vimercati</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Samarati</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. of the 21st Conference on Data and Applications Security</title>
		<title level="s">Lecture Notes in Computer Science</title>
		<meeting>of the 21st Conference on Data and Applications Security</meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2007">2007</date>
			<biblScope unit="volume">4602</biblScope>
			<biblScope unit="page" from="47" to="60" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">Protecting privacy against location-based personal identification</title>
		<author>
			<persName><forename type="first">C</forename><surname>Bettini</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><forename type="middle">S</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Jajodia</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. of the 2nd workshop on Secure Data Management</title>
				<meeting>of the 2nd workshop on Secure Data Management</meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2005">2005</date>
			<biblScope unit="volume">3674</biblScope>
			<biblScope unit="page" from="185" to="199" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">Location anonymity in continuous location-based services</title>
		<author>
			<persName><forename type="first">T</forename><surname>Xu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Cai</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. of ACM International Symposium on Advances in Geographic Information Systems</title>
				<meeting>of ACM International Symposium on Advances in Geographic Information Systems</meeting>
		<imprint>
			<publisher>ACM Press</publisher>
			<date type="published" when="2007">2007</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">Never walk alone: Uncertainty for anonymity in moving objects databases</title>
		<author>
			<persName><forename type="first">O</forename><surname>Abul</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Bonchi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Nanni</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. of the 24th International Conference on Data Engineering</title>
				<meeting>of the 24th International Conference on Data Engineering</meeting>
		<imprint>
			<publisher>IEEE Computer Society</publisher>
			<date type="published" when="2008">2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">Privacy preservation in the publication of trajectories</title>
		<author>
			<persName><forename type="first">M</forename><surname>Terrovitis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Mamoulis</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. of the 9th International Conference on Mobile Data Management</title>
				<meeting>of the 9th International Conference on Mobile Data Management</meeting>
		<imprint>
			<publisher>IEEE Computer Society</publisher>
			<date type="published" when="2008">2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">Tracking moving objects in anonymized trajectories</title>
		<author>
			<persName><forename type="first">N</forename><surname>Vyahhi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Bakiras</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Kalnis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Ghinita</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. of 19th International Conference on Database and Expert Systems Applications</title>
				<meeting>of 19th International Conference on Database and Expert Systems Applications</meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2008">2008</date>
		</imprint>
	</monogr>
	<note>to Appear</note>
</biblStruct>

<biblStruct xml:id="b18">
	<analytic>
		<title level="a" type="main">Spatial generalization algorithms for lbs privacy preservation</title>
		<author>
			<persName><forename type="first">S</forename><surname>Mascetti</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Bettini</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Freni</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><forename type="middle">S</forename><surname>Wang</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Location Based Services</title>
		<imprint>
			<biblScope unit="volume">2</biblScope>
			<biblScope unit="issue">1</biblScope>
			<date type="published" when="2008">2008</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
