<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Evaluating Differential Privacy Approaches for Query Obfuscation in Information Retrieval ⋆ Discussion Paper</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Guglielmo</forename><surname>Faggioli</surname></persName>
							<affiliation key="aff0">
								<orgName type="institution">University of Padova</orgName>
								<address>
									<settlement>Padova</settlement>
									<country key="IT">Italy</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Nicola</forename><surname>Ferro</surname></persName>
							<affiliation key="aff0">
								<orgName type="institution">University of Padova</orgName>
								<address>
									<settlement>Padova</settlement>
									<country key="IT">Italy</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Evaluating Differential Privacy Approaches for Query Obfuscation in Information Retrieval ⋆ Discussion Paper</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">8A27DC8F3ABCFC8FEBA7261EB89AF106</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T18:23+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Protecting the privacy of a user while they interact with an Information Retrieval (IR) system is crucial. This becomes more challenging when the IR system is not cooperative in satisfying the user's privacy needs. Recent advancements in Natural Language Processing (NLP) have demonstrated Differential Privacy's (DP) effectiveness in safeguarding text privacy for tasks like spam detection and sentiment analysis, even under the assumption of a non-cooperative system. Our investigation explores if DP methods, originally designed for specific NLP tasks, can effectively obscure queries in IR. Our analyses show that using the Vickrey DP mechanism, employing the Mahalanobis norm with a privacy budget ranging from 𝜖 = 10 to 12.5, provides cutting-edge privacy protection and enhances effectiveness. Unlike previous methods, DP allows users to fine-tune their desired level of privacy by adjusting the privacy budget 𝜖. This flexibility offers a balance between how effective the system is and how much privacy is maintained, unlike the more rigid nature of previous approaches.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>Information Retrieval (IR) systems are a commodity used for many tasks, including searching for personal information, such as symptoms and diseases, political opinions, or egosurfing. Such searches can be used to profile the user and can put at risk their privacy. For example, an insurance company might try to access the user's queries to determine if they have any disease, or a malicious employee of a search engine might access the query log to blackmail them. To alleviate this, obfuscation approaches hide the sensitive information need by breaking it down into multiple non-sensitive queries. To this end, some approaches rely on replacing words with generalizations, i.e., hypernyms <ref type="bibr" target="#b1">[2]</ref>. Other strategies use a local corpus to determine which words, by co-occurring in the documents with those in the query, induce the same ranked list <ref type="bibr" target="#b2">[3]</ref>. We investigate for the first time whether Differential Privacy (DP) mechanisms, originally designed for specific Natural Language Processing (NLP) tasks, can effectively be used in IR to obfuscate queries. DP <ref type="bibr" target="#b3">[4]</ref> is a state-of-the-art framework meant to release privately sensitive information. The general idea is to use a randomized mechanism that introduces noise into the computation. Thanks to this, the user can "plausibly deny" the output: it is impossible to prove that the output corresponds to the input of the user and is not due to the randomness of the mechanism. DP is particularly effective in the NLP domain. A line of research <ref type="bibr" target="#b4">[5,</ref><ref type="bibr" target="#b5">6,</ref><ref type="bibr" target="#b6">7]</ref> operationalizes DP to release text by obfuscating each word individually. Such mechanisms work as follows: i) each word in the text is mapped to a non-contextual embedding space; ii) the embeddings are perturbed with noise drawn from a specific distribution; iii) each word is replaced with the word closest to the noisy embedding. A major advantage of DP is that it allows setting the privacy budget based on the needs of the user. This is different from current obfuscation mechanisms in IR, which are either active or not and cannot be tuned based on the user's needs. In this work, we focus on three of DP mechanisms: the Calibrated Multivariate Perturbation (CMP) <ref type="bibr" target="#b4">[5]</ref>, the Mahalanobis <ref type="bibr" target="#b5">[6]</ref> and the Vickrey <ref type="bibr" target="#b6">[7]</ref>. These approaches were originally devised and tested for NLP tasks that include text classification and sentiment analysis. We assume the IR system to not preserve user privacy, and to possibly be malicious. In our use case, users are the ones concerned about their privacy. They do not want to reveal their real information needs and prefer to transmit obfuscated queries to the IR system while still retrieving relevant documents. Therefore, to operationalize our mechanism, we assume each user to locally obfuscate their query and transmit the obfuscated query, or possibly multiple queries, to the IR system instead of their real query. Our goal is to determine if the DP mechanisms introduced above can successfully obfuscate users' information needs while still retrieving relevant documents.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Approaches</head><p>All the approaches described in this work are based on a relaxation of classical DP, called Metric-DP. To achieve traditional DP in a metric space, an obfuscation mechanism should have an equal probability of obfuscating any pair of points as the same point, irrespective of their distance. While this grants the highest level of privacy, it also requires high levels of noise, decreasing the utility of the data. In the case of metric spaces, it is often sufficient if the probability of obfuscating two points with the same one is proportional to the distance between the two points. Alternatively, the proportion of sampling a certain noise is inversely proportional to the norm of the noise itself. To this end, a relaxation of DP, called Metric-DP, has been introduced. Metric-DP <ref type="bibr" target="#b7">[8,</ref><ref type="bibr" target="#b8">9,</ref><ref type="bibr" target="#b9">10]</ref> is defined as follows: given a privacy budget 𝜖 and a distance measure 𝑑 : R 𝑝 × R 𝑝 → [0, ∞), a randomized mechanism ℳ : R 𝑝 → R 𝑝 defined over a geometric space is Metric-DP iff, for any three points in the space 𝑤, 𝑤 ′ , 𝑤 ˆ∈ R 𝑝 , the following holds: 𝑃 𝑟{ℳ(𝑤)=𝑤 ^} 𝑃 𝑟{ℳ(𝑤 ′ )=𝑤 ^} ≤ exp(𝜖𝑑{𝑤, 𝑤 ′ }) If the 𝑑{𝑤, 𝑤 ′ } is small, 𝑤 and 𝑤 ′ are more likely to be obfuscated with the same point. Vice-versa, far apart points might be obfuscated with different points, without violating privacy constraints.</p><p>We describe here the three major DP efforts for obfuscating text in the NLP scenario, which we evaluate for the IR task. More in detail, these approaches take as input a sequence of words. Each word is mapped into a non-contextual embedding, such as GloVe <ref type="bibr" target="#b10">[11]</ref>. Then, the embedding is obfuscated by adding some appositely sampled noise to it. To ensure that Metric-DP is achieved, the noise vector 𝑧 is expected to be sampled from a distribution 𝑓 such that the probability of observing 𝑧 is 𝑓 (𝑧) ∝ exp(−𝜖||𝑧||), i.e., the probability of sampling a noise with norm ||𝑧|| is inversely proportional to ||𝑧||. Finally, the closest word to the noisy embedding is used to obfuscate the corresponding word in the original text. We propose to use these approaches in the IR scenario to perturb the queries instead of the documents, as done for NLP tasks.</p><p>The Calibrated Multivariate Perturbation (CMP) mechanism, defined by Feyisetan et al. <ref type="bibr" target="#b4">[5]</ref>, is based on sampling a noise vector for each term in the query following an n-dimensional Laplace distribution. Such sampling works by sampling two vectors: i) an n-dimensional unitary vector 𝑝 ∈ R 𝑛 that represents the direction of the perturbation. ii) the radius of the perturbation 𝑟 ∈ R + is sampled from a Gamma distribution. To sample 𝑝, a vector 𝑁 ∈ R 𝑛 is sampled from a multivariate normal distribution, with location 0 and identity covariance matrix I 𝑛 : 𝑁 ∼ 𝒩 (0, I 𝑛 ). Then 𝑝 = 𝑁/||𝑁 || 2 . The radius 𝑟 of the noise is sampled from a Gamma distribution with shape 𝑛 and scale 1  𝜖 as 𝑟 ∼ 𝐺𝑎𝑚(𝑛, 1 𝜖 ). It is possible to observe that, the larger the privacy requirement, i.e., the smaller the 𝜖, the bigger the noise. The noise 𝑧 is defined as 𝑧 = 𝑝 • 𝑟. To perturb a word 𝑤, the noise vector 𝑧 is added to the original word embedding 𝜑(𝑤) ∈ R 𝑛 , and the word closest to the noisy word embedding is used as obfuscation. Feyisetan et al. <ref type="bibr" target="#b4">[5]</ref> demonstrate that for any word sequence 𝒲 𝑙 of length 𝑙 ≥ 1 and any 𝜖 &gt; 0, CMP satisfies 𝜖𝑑-privacy with respect to 𝑑, where 𝑑 is the Euclidean distance.</p><p>The second mechanism investigated is the Mahalanobis (Mhl) mechanism. Xu et al. <ref type="bibr" target="#b5">[6]</ref> noticed how the perturbation induced by CMP mechanism tends to be weak, especially for high 𝜖. They hypothesize that sampling the direction of the perturbation on a circumference (||𝑝|| 2 = 1) increases the risk of sampling a point on an empty region. Therefore, Xu et al. adapt the CMP mechanism by transforming the direction of the noise from a circumference to an ellipsis whose orientation can be set to be towards the other embeddings. To do so, it is necessary to modify the sampling mechanism, so that, instead of sampling 𝑝 such that ||𝑝|| 2 = 1, 𝑝 is sampled so that ||𝑝|| 𝑀 = 1 where || • || 𝑀 is the Mahalanobis norm. To ensure that the noise 𝑧 is sampled such that its probability distribution is 𝑓 (𝑧) ∝ exp(−𝑒||𝑧|| 𝑀 ) a vector 𝑁 is sampled from the multivariate normal distribution 𝑁 ∼ 𝒩 (0, 𝐼 𝑛 ). Then, 𝑝 is such that 𝑝 = Σ 1/2 • (𝑁/||𝑁 || 2 ), where Σ ∈ R 𝑛×𝑛 is the covariance matrix of all the word embeddings. This forces the noise towards more populated areas. The sampling of the norm of the noise 𝑟 is the same as for CMP.</p><p>Finally, we investigate the Vickrey (Vkr) mechanism. The Mhl still tends to obfuscate a word with itself for large 𝜖. To reduce the probability of masking a token with itself, Xu et al. <ref type="bibr" target="#b6">[7]</ref> define the Vickrey DP mechanism (we refer to it as Vkr). Vkr is based on two steps. In the first step, a noisy vector is sampled using any of the mechanisms described above: we can instantiate Vkr with either Mhl mechanism (Vkr 𝑀 ℎ𝑙 ) or the CMP mechanism (Vkr 𝐶𝑀 𝑃 ). In the second step, with probability 𝑃 𝑟 the word corresponding to the closest embedding to the noisy vector is used as the obfuscation word. Vice versa, with probability 1 − 𝑃 𝑟 the word corresponding to the second closest embedding is used as obfuscation. The probability 𝑃 𝑟 is defined as 𝑃 𝑟(𝑡, 𝑣 ˆ) =</p><formula xml:id="formula_0">(1−𝑡)||𝜑(𝑢 2 )−𝑣 ^|| 2 𝑡||𝜑(𝑢 1 )−𝑣 ^|| 2 +(1−𝑡)||𝜑(𝑢 2 )−𝑣 ^|| 2 ,</formula><p>where 𝜑(𝑢 1 ) and 𝜑(𝑢 2 ) are respectively the closest and second closest word embeddings to 𝑣 ˆ, the perturbed embedding of 𝑤, and 𝑡 is an additional free parameter. We set 𝑡 = 0.75, being the best performing <ref type="bibr" target="#b6">[7]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Evaluation</head><p>We consider two different collections TREC Robust '04 and TREC Deep Learning (DL '19). As word embeddings, we used GloVe <ref type="bibr" target="#b10">[11]</ref> with 300 dimensions trained on the Common Crawl. In terms of retrieval models, we consider a sparse bag-of-word model, BM25, and a dense bi-encoder, Contriever <ref type="bibr" target="#b11">[12]</ref>. To set a baseline, we compare the DP approaches aforementioned with two non-DP obfuscation approaches originally devised explicitly for the IR task. We take into consideration the seminal work by Arampatzis et al. <ref type="bibr" target="#b1">[2]</ref>, labeled AED, and the recent state-of-the-art solution by Fröbe et al. <ref type="bibr" target="#b2">[3]</ref>, labeled FSH. For each approach and for each query, we generate 20 obfuscation queries. Table <ref type="table" target="#tab_0">1</ref> shows, as a proxy of the privacy achieved by the mechanisms, the average similarity between the original query and the obfuscation queries generated to hide it. We compute the similarity as the dot product between the MiniLM <ref type="bibr" target="#b12">[13]</ref> representations of the original query and the obfuscated ones. As expected from a DP mechanism, the higher the 𝜖 the higher the similarity between the queries -with 𝜖 = 50 for both Robust '04 and DL '19, CMP and Mhl achieve similarity higher than 95%. This indicates that overall the generated queries are almost identical to the original ones and there is no substantial privacy protection. FSH, which explicitly removes synonyms and hypernyms from the queries, is particularly safe and corresponds to a DP Vkr 𝐶𝑀 𝑃 mechanism with 𝜖 ∈ <ref type="bibr" target="#b4">[5,</ref><ref type="bibr" target="#b9">10]</ref> or a Vkr 𝑀 ℎ𝑙 with 𝜖 ∈ [10, 12.5] for the Robust '04, and DP Vkr 𝐶𝑀 𝑃 and a Vkr 𝑀 ℎ𝑙 mechanism with 𝜖 ∈ <ref type="bibr" target="#b4">[5,</ref><ref type="bibr" target="#b9">10]</ref> for the DL '19. The privacy achieved by AED can be achieved with 𝜖 in the range [10; 12.5] by CMP and Mhl on both collections. 𝜖 values that grant a comparable level of privacy are much higher for Vkr-based mechanisms, especially Vkr 𝑀 ℎ𝑙 , on both collections -this means that the Vkr mechanisms are substantially more secure from a privacy perspective.</p><p>As both CMP and Mhl are less effective from a privacy perspective, we focus the following analyses on the Vkr mechanism, with 𝜖 ∈ {10, 12.5, 15}. More in detail, we compare these DP mechanisms with AED and FSH, based on three axes: i) the obfuscation; ii) the pooled recall; iii) the nDCG@10 observed if we re-rank the documents pooled by the obfuscation queries. We define the obfuscation as 1 minus the similarity of the original query and the obfuscated one. The pooled recall is obtained by transmitting to the IR system 20 obfuscated queries: for each ranked list in response to an obfuscated query, we select the first 100 documents and merge all the sets of documents obtained. We compute the recall on this new set of documents. Finally, to compute nDCG@10, we rerank the pooled documents using a different IR model (we use TAS-B to avoid biasing toward any IR model) and evaluate the quality of this ranked list. For each approach, these measures are reported on a radar plot where, as a rule of thumb, a larger area corresponds to more desirable results. Figure <ref type="figure" target="#fig_1">1</ref> reports the radar plots, showing the performance of different obfuscation approaches over the three axes mentioned above. We notice that the area corresponding to the AED approach (in red) is encompassed within the area corresponding to Vkr 𝑀 ℎ𝑙 with 𝜖 = 15 (green). In fact, on the Robust '04 collection, AED  achieves nDCG@10 of 0.410 and 0.424 for BM25 and Contriever respectively, recall of 0.420 and 0.419, and obfuscation of 0.513. Vice versa Vkr 𝑀 ℎ𝑙 with 𝜖 = 15 obtains nDCG@10 of 0.416 and 0.431, recall of 0.493 and 0.462, and obfuscation of 0.618. The exception is DL '19 with Contriever as the IR system, where AED has higher recall than Vkr 𝑀 ℎ𝑙 (0.497 against 0.418). Nevertheless, this larger recall does not correspond to much larger nDCG@10, indicating that Vkr 𝑀 ℎ𝑙 is preferable over AED, as it has comparable nDCG@10 (0.604 for Vkr 𝑀 ℎ𝑙 against 0.607 for AED), with improved obfuscation (0.785 against 0.491). When it comes to FSH (purple), the behaviour depends on the collection. In the DL '19, using Vkr 𝑀 ℎ𝑙 with 𝜖 = 10 (blue) provides an edge over FSH: they have comparable obfuscation (0.916 the former, 0.923 the latter), but Vkr 𝑀 ℎ𝑙 has much larger nDCG@10 (0.254 compared to 0.064). On the Robust '04 collection, to observe an improvement in terms of nDCG@10, it is necessary to use Vkr 𝑀 ℎ𝑙 with 𝜖 = 12.5 (nDCG@10 of 0.349 and 0.355 for BM25 and Contriever respectively) to overcome FSH in terms of nDCG@10 (0.140 and 0.194). While Vkr 𝑀 ℎ𝑙 with 𝜖 = 12.5 exhibits nDCG@10 performance slightly lower than AED, it also has obfuscation (0.719) relatively close to FSH, which has obfuscation of 0.797, much closer than AED, with obfuscation 0.513. As a general guideline, we propose to use Vkr 𝑀 ℎ𝑙 as the obfuscation mechanism, with 𝜖 chosen in the interval <ref type="bibr" target="#b9">[10,</ref><ref type="bibr">15]</ref>, depending on the optimal trade-off between privacy and effectiveness, as chosen by the user.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Conclusion and Future Work</head><p>In this work, we analyzed for the first time the performance of three DP mechanisms, originally designed for NLP, in the proxy query obfuscation IR task. We evaluated these mechanisms on the IR setting by considering three aspects: their obfuscation capabilities, their effectiveness in terms of recall, and their ability in allowing to retrieve highly relevant documents. Our findings highlight that the Vickrey mechanism with 𝜖 ∈ [10, 12.5] achieves higher privacy guarantees, with improved effectiveness, than current state-of-the-art approaches. Furthermore, lower or higher levels of 𝜖 allow for better satisfy the user, either in terms of privacy or accuracy, depending on their inclinations. As a future work, we plan to investigate how to perturb dense representations of the queries and combine them with generative language models to produce obfuscation queries with the same dense representation, but different terms.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: Performance of different obfuscation mechanisms over three axes: pooled recall, nDCG@10 of the reranked documents, obfuscation (obf), measured as 1-similarity. Cnt. stands for "Contriever".</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1</head><label>1</label><figDesc>Average MiniLM sentence similarity between the original query and 20 obfuscation queries generated with different approaches.</figDesc><table><row><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell cols="2">Robust '04</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell>DL '19</cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell>𝜖</cell><cell>1</cell><cell>5</cell><cell>10</cell><cell>12.5</cell><cell>15</cell><cell>17.5</cell><cell>20</cell><cell>50</cell><cell>No DP</cell><cell>1</cell><cell>5</cell><cell>10</cell><cell>12.5</cell><cell>15</cell><cell>17.5</cell><cell>20</cell><cell>50</cell><cell>No DP</cell></row><row><cell cols="9">CMP 0.074 0.100 0.396 0.672 0.871 0.961 0.987 0.996</cell><cell></cell><cell cols="8">0.024 0.032 0.214 0.458 0.681 0.824 0.903 0.952</cell><cell></cell></row><row><cell cols="9">Mhl 0.077 0.095 0.244 0.427 0.627 0.794 0.907 0.996</cell><cell></cell><cell cols="8">0.020 0.034 0.119 0.241 0.427 0.610 0.750 0.951</cell><cell></cell></row><row><cell cols="9">Vkr𝐶𝑀𝑃 0.077 0.100 0.278 0.412 0.511 0.578 0.622 0.760</cell><cell></cell><cell cols="8">0.028 0.032 0.137 0.211 0.308 0.372 0.413 0.565</cell><cell></cell></row><row><cell cols="9">Vkr𝑀ℎ𝑙 0.076 0.096 0.188 0.282 0.382 0.472 0.533 0.746</cell><cell></cell><cell cols="8">0.023 0.026 0.084 0.149 0.215 0.284 0.333 0.553</cell><cell></cell></row><row><cell>AED</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell>0.487</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell>0.509</cell></row><row><cell>FSH</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell>0.203</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell>0.077</cell></row></table></figure>
		</body>
		<back>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Query Obfuscation for Information Retrieval through Differential Privacy</title>
		<author>
			<persName><forename type="first">G</forename><surname>Faggioli</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Ferro</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Procs. of ECIR 2024</title>
				<meeting>s. of ECIR 2024</meeting>
		<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Enhancing deniability against query-logs</title>
		<author>
			<persName><forename type="first">A</forename><surname>Arampatzis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><forename type="middle">S</forename><surname>Efraimidis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Drosatos</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Procs. of ECIR 2011</title>
				<meeting>s. of ECIR 2011</meeting>
		<imprint>
			<date type="published" when="2011">2011</date>
			<biblScope unit="page" from="117" to="128" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Efficient query obfuscation with keyqueries</title>
		<author>
			<persName><forename type="first">M</forename><surname>Fröbe</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><forename type="middle">O</forename><surname>Schmidt</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Hagen</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Procs. of WI-IAT &apos;21</title>
				<meeting>s. of WI-IAT &apos;21</meeting>
		<imprint>
			<date type="published" when="2021">2021</date>
			<biblScope unit="page" from="154" to="161" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">The algorithmic foundations of differential privacy</title>
		<author>
			<persName><forename type="first">C</forename><surname>Dwork</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Roth</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Found. Trends Theor. Comput. Sci</title>
		<imprint>
			<biblScope unit="volume">9</biblScope>
			<biblScope unit="page" from="211" to="407" />
			<date type="published" when="2014">2014</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Privacy-and utility-preserving textual analysis via calibrated multivariate perturbations</title>
		<author>
			<persName><forename type="first">O</forename><surname>Feyisetan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Balle</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Drake</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Diethe</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Procs. of WSDM &apos;20</title>
				<meeting>s. of WSDM &apos;20</meeting>
		<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="178" to="186" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">A differentially private text perturbation method using regularized mahalanobis metric</title>
		<author>
			<persName><forename type="first">Z</forename><surname>Xu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Aggarwal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Feyisetan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Teissier</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Procs. of the Second Workshop on Privacy in NLP</title>
				<meeting>s. of the Second Workshop on Privacy in NLP</meeting>
		<imprint>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<monogr>
		<title level="m" type="main">On a utilitarian approach to privacy preserving text generation</title>
		<author>
			<persName><forename type="first">Z</forename><surname>Xu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Aggarwal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Feyisetan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Teissier</surname></persName>
		</author>
		<idno>CoRR abs/2104.11838</idno>
		<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Geo-indistinguishability: differential privacy for location-based systems</title>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">E</forename><surname>Andrés</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Bordenabe</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Chatzikokolakis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Palamidessi</surname></persName>
		</author>
		<idno type="DOI">10.1145/2508859.2516735</idno>
	</analytic>
	<monogr>
		<title level="m">ACM SIGSAC Conference on Computer and Communications Security, CCS&apos;13</title>
				<editor>
			<persName><forename type="first">A</forename><surname>Sadeghi</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">V</forename><forename type="middle">D</forename><surname>Gligor</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">M</forename><surname>Yung</surname></persName>
		</editor>
		<meeting><address><addrLine>Berlin, Germany</addrLine></address></meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2013-11-04">2013. November 4-8, 2013. 2013</date>
			<biblScope unit="page" from="901" to="914" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Broadening the scope of differential privacy using metrics</title>
		<author>
			<persName><forename type="first">K</forename><surname>Chatzikokolakis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Andrés</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Bordenabe</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Palamidessi</surname></persName>
		</author>
		<idno type="DOI">10.1007/978-3-642-39077-7_5</idno>
		<idno>doi:</idno>
		<ptr target="10.1007/978-3-642-39077-7\_5" />
	</analytic>
	<monogr>
		<title level="m">Privacy Enhancing Technologies -13th International Symposium, PETS 2013</title>
		<title level="s">Lecture Notes in Computer Science</title>
		<editor>
			<persName><forename type="first">E</forename><forename type="middle">D</forename><surname>Cristofaro</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">M</forename><forename type="middle">K</forename><surname>Wright</surname></persName>
		</editor>
		<meeting><address><addrLine>Bloomington, IN, USA</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2013">July 10-12, 2013. 2013</date>
			<biblScope unit="volume">7981</biblScope>
			<biblScope unit="page" from="82" to="102" />
		</imprint>
	</monogr>
	<note>Proceedings</note>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">A framework of metrics for differential privacy from local sensitivity</title>
		<author>
			<persName><forename type="first">P</forename><surname>Laud</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Pankova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Pettai</surname></persName>
		</author>
		<idno type="DOI">10.2478/popets-2020-0023</idno>
	</analytic>
	<monogr>
		<title level="j">Proc. Priv. Enhancing Technol</title>
		<imprint>
			<biblScope unit="volume">2020</biblScope>
			<biblScope unit="page" from="175" to="208" />
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Glove: Global vectors for word representation</title>
		<author>
			<persName><forename type="first">J</forename><surname>Pennington</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Socher</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">D</forename><surname>Manning</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Procs. of EMNLP 2014</title>
				<meeting>s. of EMNLP 2014</meeting>
		<imprint>
			<date type="published" when="2014">2014</date>
			<biblScope unit="page" from="1532" to="1543" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Unsupervised dense information retrieval with contrastive learning</title>
		<author>
			<persName><forename type="first">G</forename><surname>Izacard</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Caron</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Hosseini</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Riedel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Bojanowski</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Joulin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Grave</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Trans. Mach. Learn. Res</title>
		<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">MiniLM: Deep Self-Attention Distillation for Task-Agnostic Compression of Pre-Trained Transformers</title>
		<author>
			<persName><forename type="first">W</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Wei</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Dong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Bao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Yang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Zhou</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Procs. of NeurIPS &apos;20</title>
				<meeting>s. of NeurIPS &apos;20</meeting>
		<imprint>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
