<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <article-id pub-id-type="doi">10.1137/1.9781611977653.CH99</article-id>
      <title-group>
        <article-title>QuIPU: Evaluating Actual Privacy of Obfuscated Queries.⋆</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Discussion Paper</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Francesco Luigi</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>De Faveri</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Guglielmo</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Faggioli</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Nicola</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Ferro</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Department of Information Engineering, University of Padova</institution>
          ,
          <addr-line>Padova</addr-line>
          ,
          <country country="IT">Italy</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2025</year>
      </pub-date>
      <volume>1</volume>
      <fpage>16</fpage>
      <lpage>19</lpage>
      <abstract>
        <p>When Information Retrieval (IR) models are applied to and trained on sensitive and personal information, users' privacy is at risk. While mechanisms have been presented to safeguard user privacy, the efectiveness of these privacy protections is generally evaluated by studying the relations between performance on a downstream task and the parameters of the mechanisms, e.g., the privacy budget  in Diferential Privacy ( DP). This often causes a partial understanding between formal privacy and the privacy experienced by the user, the actual privacy. In this paper, we discuss the Query Inference for Privacy and Utility (QuIPU) framework, a novel evaluation methodology designed to assess actual privacy based on the risk that an “honest-but-curious” IR system may correctly guess the original query from the obfuscated queries received. The QuIPU framework constitutes the ifrst endeavour to quantify actual privacy for IR tasks, extending beyond the partial comparison of formal privacy parameters. Our findings show that formal privacy parameters do not necessarily correspond to actual privacy, resulting in cases where, despite identical privacy parameters, two systems reveal difering actual privacy levels.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Evaluation Measures</kwd>
        <kwd>Diferential Privacy</kwd>
        <kwd>Information Retrieval</kwd>
        <kwd>Information Security</kwd>
        <kwd>Privacy Risks</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Privacy is a fundamental right preserved by the Universal Declaration of Human Rights and its Article
12. Natural Language Processing (NLP) and Information Retrieval (IR) algorithms are trained and tested
using textual datasets consisting of queries, documents, reviews and posts on online social media. In
such a large amount of textual data, personal user profiles, personal opinions on diferent matters, such
as politics and religions [
        <xref ref-type="bibr" rid="ref2 ref3">2, 3</xref>
        ], along with sensitive information needs, can raise privacy concerns from
the user interactions with such systems. Specifically, through the analysis of browser search histories
and obtained documents, malicious actors can reveal private information, including an individual’s
salary and medical conditions [
        <xref ref-type="bibr" rid="ref4 ref5">4, 5</xref>
        ]. Heuristic strategies [
        <xref ref-type="bibr" rid="ref6 ref7">6, 7</xref>
        ] have been proposed for preserving privacy
in document retrieval tasks. From another view, progress in NLP have demonstrated the potential of
Diferential Privacy ( DP) [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] in the release of privacy-preserving text for diferent applications, including
text classification [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ], authorship anonymization [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ], and query obfuscation [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ].
      </p>
      <p>
        Obfuscating a query concerns safeguarding the original information need of a user in such a manner
that the resulting obfuscated queries can retrieve relevant documents while not fully disclosing those
information needs. For example, the query “how tumour grows” may be transformed into the
obfuscated alternatives “how cancer grows”, “how infection spreads”, “how leukemia evolves”.
Focusing on the mechanisms’ privacy parameters represents a preliminary way to measure privacy.
Several attempts to assess the privacy provided have been proposed by adapting information security
measures based on entropy [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ], syntactic and semantic similarities [
        <xref ref-type="bibr" rid="ref13 ref14">13, 14</xref>
        ] between original and
obfuscated texts. However, all these measures limit the actual privacy evaluation reached by the
mechanism [
        <xref ref-type="bibr" rid="ref15 ref16 ref17">15, 16, 17</xref>
        ]. After observing the obfuscated queries “how cancer grows”, “how infection
spreads”, “how leukemia evolves”, an adversarial system can quickly discover the actual user
information need about cancer diseases and spreading using available query logs to generate potential
guesses of the original query. Nevertheless, some privacy budget parameters would lead to obfuscated
queries, giving mathematical guarantees of privacy, and most of the state-of-the-art measures would
consider such queries as properly obfuscated. In addition, since such queries still retain similar semantic
meaning to the original, they would probably produce high retrieval utility, giving a false impression
of privacy and achieving high retrieval results. Therefore, limiting the privacy analysis to the formal
mechanism parameters does not quantify the user’s risk in submitting the obfuscated queries [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ].
      </p>
      <p>
        In this study, we discuss the Query Inference for Privacy and Utility (QuIPU) framework [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], an
evaluation model developed to measure the actual privacy and utility trade-of provided by a mechanism
to safeguard potential information leakage in an obfuscation protocol. The QuIPU score is computed by
assessing the risk represented by a malicious adversary trying to infer the original user need correctly.
Therefore, to estimate such score, the Query Inference Attack (QuIA), a variation of the inference attacks
known as Membership Inference Attack (MIA) [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ] tailored for a query obfuscation protocol, is used
against the obfuscated queries submitted to the system. Such an attack considers the risk probability
that the original query is successfully inferred from a query log by a IR system after analyzing the
alternative queries received and obfuscated using diferent configurations, i.e., using diferent formal
privacy parameters of an obfuscation mechanism. The measure takes into account the trade-of between
privacy and utility, extending beyond the configuration parameters of the obfuscated mechanisms by
calculating a modified version of the Area Under the Curve ( AUC) on the risk versus utility trend. Our
ifndings show that formal privacy does not necessarily imply actual privacy, explicitly showing that
there is a high probability of a correct query guess for low values of the privacy parameter.
      </p>
      <p>The paper is structured by first presenting the Related Works and Background in Section 2, introducing
the diferent measures used in the privacy evaluation and the background on the Query Obfuscation
protocol. Section 3 presents the formal definition of QuIPU, showing the phases completed to evaluate
the actual textual privacy provided by a mechanism. Finally, Section 4 reports the results and discussion
of the formal and actual privacy analysis performed on diferent obfuscation mechanisms.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related Works and Background</title>
      <sec id="sec-2-1">
        <title>2.1. Related Works</title>
        <p>
          Diferent metrics have been proposed to organize available privacy measures [
          <xref ref-type="bibr" rid="ref20 ref21">20, 21</xref>
          ]. Wagner and
Eckhof [
          <xref ref-type="bibr" rid="ref20">20</xref>
          ] systematically classified over eighty privacy metrics, ofering a comprehensive framework
for assessing privacy across diferent domains, e.g., communication, databases, and social networks.
The survey highlights the significance of identifying the specific aspect of privacy that a metric aims to
quantify, suggesting nine guiding aspects for selecting the appropriate privacy measures. Specifically, the
authors stress the importance of considering the adversary’s knowledge and capability when evaluating
privacy. In addition, Sousa and Kern [
          <xref ref-type="bibr" rid="ref21">21</xref>
          ] described how diferent mechanisms developed for NLP tasks
provide privacy for textual data and which can be the threats in such scenarios. Moreover, Habernal [
          <xref ref-type="bibr" rid="ref22">22</xref>
          ]
discussed the importance of not relying strictly on formal analysis of DP and its application on NLP
tasks but to push research towards concrete measurements of the privacy provided to texts.
        </p>
        <p>
          Traditional methods for evaluating privacy primarily focus on estimating the failure rates of
obfuscation mechanisms [
          <xref ref-type="bibr" rid="ref12">12</xref>
          ] or assessing the similarities between original and obfuscated texts [
          <xref ref-type="bibr" rid="ref11 ref23">23, 11</xref>
          ]. On
the one hand, uncertainty measures such as  and  [
          <xref ref-type="bibr" rid="ref24 ref25">24, 25</xref>
          ] estimate the probability that a term 
remains unchanged after obfuscation and the minimum cardinality of the set of words to which  is
mapped by the mechanism, respectively. However, such measures do not capture if the mechanism
changes the original term with a closely related one. On the other hand, the similarity between the
original and obfuscated texts is commonly estimated using metrics like the Jaccard index or cosine
similarity between sentence embeddings computed by a Transformer, drawing inspiration from the
use of BERTScores used to assess the quality of generated texts [
          <xref ref-type="bibr" rid="ref14">14</xref>
          ]. Meisenbacher et al. [
          <xref ref-type="bibr" rid="ref23">23</xref>
          ]
proposed the  -PUC score to compute an  -weighted mean between uncertainty, similarity measures, and
utility preserved. This score is tuned by the tuning parameter  , which adjusts the focus on utility
or privacy, allowing the user to decide whether to prioritize the former or the latter. However, none
of the above measures ofer insights into the actual privacy aforded to the texts, nor do they assess
the adversarial potential to infer the original meaning of the obfuscated text. Specifically, previous
studies [
          <xref ref-type="bibr" rid="ref26">26</xref>
          ] criticized the reliance on formal privacy analysis solely based on the privacy budget 
parameter. DP mechanisms employing configurations where  &gt; 1 lack a comprehensive analysis
of actual privacy guarantees, raising concerns about the suficiency of privacy protection methods
employed1. In addition, Damie et al. [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ] introduced a novel indicator to assess the risk of successful
query recovery attacks within searchable encryption protocols. The study revealed that, even without
additional background knowledge, an adversary can obtain the original queries with a success rate of
85%, encouraging analysis of privacy measures employed considering real attack scenarios.
        </p>
      </sec>
      <sec id="sec-2-2">
        <title>2.2. Background</title>
        <p>
          The research community has widely studied application of privacy measure when performing NLP
and IR tasks [
          <xref ref-type="bibr" rid="ref27 ref28 ref29">27, 28, 29</xref>
          ]. The scenario discussed in this study assumes that the users are willingly
paying part of the utility during the document retrieval phase to defend the privacy of their search
activity with the Information Retrieval system. The system is considered non-cooperative, as it does not
actively contribute to protecting user privacy, e.g., it does not provide any private online API to mask
the information need of the user. Figure 1 illustrates the general query obfuscation protocol, the focus
of application of QuIPU in IR [
          <xref ref-type="bibr" rid="ref11 ref30">11, 30</xref>
          ]. The process considers two distinct domains: on the user (safe)
side, the original query is generated by the user and privatized using an obfuscation mechanism, i.e., an
algorithm that, given an original sensitive query , generates  non-sensitive obfuscated queries that
(theoretically) prevent the unveiling of the original information need. These obfuscated queries are
sent to the IR system without explicitly disclosing their information need, after the initial obfuscation
process. During this step, the user sets the parameters, i.e., formal privacy guarantees, considering the
utility lost on the tasks [
          <xref ref-type="bibr" rid="ref31 ref32">31, 32</xref>
          ]. On the (unsafe) IR system side, relevant documents are retrieved by
the “honest” system considering the obfuscated queries received, thus putting such documents on a
lower rank. Finally, the documents are returned to the users for the post-processing described above.
To prevent a “curious” IR system from discovering the actual query, the obfuscation methods employed
are divided into two families of mechanisms, either based on heuristics or -DP.
        </p>
        <p>
          Heuristics Obfuscation. To protect privacy in IR tasks, non-formal obfuscation methods were
proposed [
          <xref ref-type="bibr" rid="ref6 ref7">6, 7</xref>
          ]. Arampatzis et al. [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ] employed the WordNet [
          <xref ref-type="bibr" rid="ref33">33</xref>
          ] database to replace original terms
within the query using synonyms, hypernyms, and holonyms. The obfuscation was performed based
1The DP configurations with  &gt; 1 deviate from the “theoretically secure” privacy setting, i.e., strong assurance about the
formal privacy introduced, see DP definition [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ].
on a hierarchical degree, i.e., the level parameter, aligned with the desired obfuscation the user aims to
achieve. Such an approach was further extended by Fröbe et al. [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ]. More in detail, the obfuscation
approach retrieves locally the top- documents from a local corpus. Then, using a sliding window, the
sequences of  terms within such documents are taken as candidate obfuscation queries, removing
those queries that contain synonyms and holonyms. Using the top- documents retrieved locally as
pseudo-relevant, the queries submitted are the ones that achieve the higher nDCG.
        </p>
        <p>
          Diferential Privacy ( DP) Obfuscation. Dwork et al. [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ] introduced the -DP framework to
formalize the privacy guarantees when releasing data. Given a privacy budget  ∈ R+, and any pair of
neighbouring datasets , ′, i.e., datasets that difer for only one entry, an obfuscation mechanism ℳ
is DP if it holds the inequality Pr [ℳ() ∈ ] ≤  · Pr [ℳ(′) ∈ ] ∀ ⊂ Im(ℳ). DP introduces
calibrated noise levels during output computation using the privacy budget , which controls the balance
between data privacy and utility. The adoption of the DP framework for metric spaces, and therefore for
NLP tasks, has been proposed in [
          <xref ref-type="bibr" rid="ref34">34</xref>
          ]. Metric-DP extends the traditional DP definition by ensuring that
the probability of obfuscating two distinct points , ′ is proportional to the distance (, ′) between
them. The DP formal framework has enabled the privacy research community to propose diferent
strategies based on noisy sampling [
          <xref ref-type="bibr" rid="ref30 ref35 ref36 ref37">35, 36, 37, 30</xref>
          ] and perturbed word embeddings [
          <xref ref-type="bibr" rid="ref24 ref25">24, 25, 38</xref>
          ].
        </p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. The QuIPU Evaluation Framework</title>
      <p>We present the Query Inference for Privacy and Utility (QuIPU) framework: we report the threat model
for an obfuscation protocol and the settings of QuIA. Finally, we report the risk evaluation of the attack.</p>
      <sec id="sec-3-1">
        <title>3.1. Defining the Threat Model</title>
        <p>In this scenario, the adversary is represented by the IR system, which aims to understand the original
user information need. In the query obfuscation protocol, the sweet spot for inferring the original
queries is represented by the ones the system receives. The mechanism parameters, e.g., the  privacy
budget parameter of the DP obfuscation mechanisms, do not guarantee with absolute certainty that the
original text is changed (or changed enough). Therefore, such queries may cause a leakage of the real
information need. In addition, for the same parameters, diferent obfuscation strategies may produce
texts with diferent obfuscation degrees. For instance, the efect of the parameter  depends on the
specific mechanism used [ 39, 40]. As a result, two DP mechanisms (one embedding-perturbation based
and the other sampling-based) are both parametrized with  = 3 and could lead to a situation where
one method achieves an actual obfuscation while the other achieves only formal obfuscation. Therefore,
the IR system aims to extract as much information as possible from the received queries, previously
obfuscated on the user side, using this knowledge to infer the real text.</p>
        <p>Consequently, the threat of a successful query inference stems not only from the obfuscation failure
of the mechanism but also from the additional knowledge about the queries possessed by the adversary.
The IR system might exploit its queries from the logs [41, 42]: by producing a classification on the
information needs carried by the obfuscated queries received and the information in its logs, it aims
to improve the chances of a correct guess of the original user query. Note that if the original query is
not an extremely long tail one, it is reasonable to assume that the original information need has been
previously submitted to the IR system, and thus, the attack can succeed with high probability.</p>
        <p>Finally, a critical remark must be made regarding using cryptographic primitives in the protocol of the
scenario we are analysing. Eavesdroppers or man-in-the-middle adversaries do not significantly threaten
the user or the system. Cryptography can be employed while exchanging queries and documents
between the client, i.e., the user, and server, i.e., the IR system, ensuring confidentiality among the
internal parties of the protocol and security against external auditors. However, confidentiality does
not imply privacy: if the IR system aims to disclose the user’s original query, cryptography techniques
alone are insuficient to safeguard privacy concerning an internal adversary.</p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Introducing the Query Inference Attack (QuIA)</title>
        <sec id="sec-3-2-1">
          <title>Algorithm 1: The Query Inference Attack (QuIA).</title>
          <p>Result: Ranked list of query logs ℒ.</p>
          <p>Data: obf (obf. queries ′), logs (query log ),  (transformer encoder).
1 Encoding  (obf) = { (′) ∈ R} and  ︀( logs︀) = { () ∈ R};
6 return ℒ;
3 Compute  = [︀ cos (ˆ,  ()) ,  () ∈ 
4 Define ℒ = [︀ (, ) ,  ∈ ,  ∈ logs︀] ;
2 Define ˆ as the centroid of the vectors in  (obf);</p>
          <p>︀( logs︀)] ;
5 Sort ℒ in descending order considering the similarity score ;</p>
          <p>
            The class of attacks known as Membership Inference Attack (MIA) was introduced by Shokri et
al. [
            <xref ref-type="bibr" rid="ref19">19</xref>
            ] to investigate the information leakage stemming from the output of machine learning models.
The attack is defined under the assumption that the attacker sees a data record but has no information
about either the model parameters or the actual model architecture, i.e., a so-called black-box scenario.
The attack is successful when the attacker can correctly guess whether the data is in the training set.
          </p>
          <p>In an obfuscation protocol, the Query Inference Attack (QuIA) uses the received obfuscated queries
and the query logs to generate a ranked list of queries from the logs based on the similarity with the
information need. Similarly to the black-box of the MIA scenario, the assumption is that the IR system
does not know the obfuscation mechanism used on the user side and the privacy parameters of the
obfuscation mechanism. Algorithm 1 reports the pseudo-code of the attack: the system receives the set
of obfuscated queries obf and knows its query logs logs. Firstly, it uses a Transformer [43] encoder 
to obtain the embeddings of the queries in the sets2. Once the texts in obf are encoded, it calculates
the centroid ˆ of the vectors in  (obf), to capture the average contextual similarities among the
obfuscated queries received. The system computes the cosine similarity between the embeddings of the
queries from the logs  () ∈ 
︀(
logs︀) and the query ˆ to understand which queries from the logs
most closely represents the average information need carried by the obfuscated queries and saves it
into the list . The algorithm generates a ranked list ℒ of the queries in the logs  ∈ logs by sorting
(, ) in descending order based on the similarities  ∈ . In case of inefective obfuscations, then
most likely, the higher a query from the logs is ranked in ℒ, the more it fits the user information need.</p>
        </sec>
      </sec>
      <sec id="sec-3-3">
        <title>3.3. Assessing the QuIPU Risk Modelling</title>
        <p>Privacy is strictly linked with the definition of risk [ 44], i.e., the possibility that an action or event
generates consequences that have an impact on what users value, in this scenario, disclosing sensitive
information. The higher the risk, the lower the privacy. For example, DP obfuscation mechanisms ofer
the possibility that privacy and utility can be balanced by tuning the privacy budget . However, the
framework does not provide any assurance against inference attacks [45]. To overcome this limitation,
we need a formal definition of the risk against inference in the obfuscation protocol. After the
QuIA
algorithm has returned the ranked list ℒ, the IR system is tasked to guess the original query. This
inference is based on the computed ranking, which considers the similarities between the obfuscated
queries received (potentially leaking information) and the system’s query logs (auxiliary knowledge for
a correct guess of the original user query). At this point, the IR system strategy to guess the correct
query is sequential: knowing that the first query is the most similar to the average information need
carried by the obfuscated queries, it represents the best choice for the guess. If the first query in the
logs ℒ is the correct query, the attack is successful, and there is a 100% risk of correct inference. On the
other hand, if the first one is not the correct guess, the adversary tries with the second query in the list,
and so on, until the original query is guessed, decreasing the risk of success. Therefore, the risk  of
2 (obf) ,  (logs) indicate the sets of text embeddings, and with  (′),  () the singular vector embedding of the queries.
successful QuIA in  guesses can be defined as the probability that the IR system correctly guesses ¯ as
the original query , seeing the sets obf and logs, i.e.,  = P ︀[ {¯= } ∩ { ≤ } |obf, logs︀] , with
 the maximum number of guessing attempts the IR system is willing to take. The upper bound for
the value of  is determined by the size of the set logs. However, determining the precise threshold 
and assessing the risk the user faces is impossible without access to the IR system’s internal data and
kind of attack. Therefore, we propose to model the malicious IR system with three kinds of attackers,
representing relevant use cases: i) the “lazy” attacker, i.e., the one that looks only at the top position of
the ranked list ℒ and makes only one guess; ii) the “active” attacker, i.e., an adversary that selects the
top- queries and checks only with them if the guess is correct; and iii) the “motivated” attacker, i.e.,
the one that tries all the queries until the original one has been found. To model the probability of the
risk a user faces against each of such attackers, we propose to use proxy indicators computed on where
the original query appears in the ranked list ℒ: Precision at 1 ( @1) for the lazy attacker, Recall at 
(@) for the active attacker, and Reciprocal Rank (RR) for the motivated attacker.</p>
        <p>Drawing inspiration from the usual ROC AUC, Figure 2 illustrates the evaluation plane that links
the risk  of a successful QuIA and the utility  measure considering a set of queries obf() – an
efectiveness measure such as nDCG in the IR case – obfuscated by a certain formal parameter  – the
 parameter in case of DP. In the risk-utility plane, the Risk-Utility Boundary line, i.e., the diagonal,
describes two regions where the risk-utility trend  (, ) can be: i) above the line indicates that the
utility  exceeds the risk , and ii) below the line, where  is less than . Therefore, the QuIPU score in
Equation 1 considers the pairs (, ) estimated by submitting the set of obfuscated queries obf().</p>
        <p>QuIPU = 2(+ + − ) = 2
∫︁
 (, )  + 2
∫︁
 (, )</p>
        <p>(1)
+ −
where  represents an infinitesimal variation on the Risk-Utility Boundary line, and the factor 2 is
introduced to map the score from [︀ − 2 2</p>
        <p>
          1 , 1 ]︀ to [
          <xref ref-type="bibr" rid="ref1">− 1, 1</xref>
          ] interval. The integrals are calculated with respect
to the diagonal of the plane, such that regions where the curve lies below this diagonal, i.e., − , are
assigned negative values, indicating that the risk  is greater than the utility . Conversely, positive
values are computed for regions where the utility  exceeds the risk , i.e., +.
        </p>
        <p>In Figure 2, four points are defined. The No Utility Point shows when the risk and utility are
reduced to 0. It depicts the situation where the obfuscation mechanism fully modifies the original query,
completely stopping a QuIA. However, the user completely renounces the efectiveness of the task, i.e.,
the submitted queries failed to retrieve any relevant documents. The No Privacy Point illustrates the
efect of not using the obfuscation protocol. The queries are not obfuscated, meaning the original query
is fully exposed to the IR system, resulting in 100% risk. Yet, utility is fully achieved, as the system
uses the original query to retrieve the full list of relevant documents. Finally, the Optimal Point and
the Trash Point present the best and worst cases theoretically obtainable. In the first, the obfuscation
mechanism provides complete protection against Query Inference attacks, i.e., 0% of risk, maintaining
maximum utility. The user’s information need are entirely met during the retrieval without exposing
any information about the original query. The second is the opposite of the optimal point, i.e., the
mechanism neither obfuscates the query nor can these queries retrieve any relevant documents. This
case can happen in a “fully-dishonest” scenario, a.e., a phishing IR system [46].</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Empirical Evaluation</title>
      <p>We present the experimental setup and compare the results observed for the privacy analysis using the
parameters and QuIPU score. Further analysis, the data used, and the code are publicly available3.</p>
      <sec id="sec-4-1">
        <title>4.1. Experiments Setup</title>
        <p>
          We test the QuIPU framework on three diferent TREC collections: Deep Learning (DL’19) [ 47] and
Deep Learning (DL’20) [48], based on the MS MARCO passages corpus, containing 43 and 54 queries
respectively, and the Robust ’04 (Robust ’04) [49] which relies on disks 4 and 5 of the TIPSTER corpus
and contains 249 queries. As obfuscation mechanisms, we consider those available in the pyPANTERA
framework [50], i.e., CMP, Mahalanobis, their Vickrey Variants, CusText, SanText, TEM, and WBB.
As privacy budget , we followed the parametrization reported in the original papers, which is also
the one used by the pyPANTERA package and other recent experiments [
          <xref ref-type="bibr" rid="ref11">11</xref>
          ]. In detail, we select
 ∈ {1, 5, 10, 12.5, 15, 17.5, 20, 25, 30, 50}. The heuristics obfuscation mechanisms, i.e., AEA [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ] and
FEA [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ], using diferent synonyms levels {3, 4, 5}, and sliding windows sizes {12, 14, 16}, respectively.
We generated 50 obfuscation variants for each query and mechanism configuration. Finally, the IR
system used as re-ranker in the post-retrieval phase of the protocol is the neural dense model Contriever,
as used in [
          <xref ref-type="bibr" rid="ref11 ref30">11, 30</xref>
          ]. The honest aspect of the IR system, i.e., the part that performs the document
retrieval task, is also the Contriever model while, the curious part of the system uses as encoder model
distilbert-base-uncased [51].We use two models for the tasks to obtain unbiased results, in line
with [
          <xref ref-type="bibr" rid="ref11 ref30">11, 30</xref>
          ]. To simulate a realistic scenario for the curious IR to perform the QuIA, we use as query
logs the AOL collection4 from which 750k queries were selected and added to the original ones.
        </p>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. Privacy Analysis Using the Mechanisms’ Parameters</title>
        <p>
          The traditional privacy analysis evaluates the utility as a function of the formal privacy parameters,
e.g., . Figure 3 reports the results of the nDCG@10 vs the formal privacy parameters on the three
diferent collections analysed. Note that the x-axis, representing the PrivacyParameter, considers both
the values for the  parameter of the DP mechanisms and the parameters of the heuristics [
          <xref ref-type="bibr" rid="ref6 ref7">6, 7</xref>
          ]. From
this traditional perspective, it emerges that with lower values of the privacy parameters, mechanisms
based on a DP strategy, like TEM or SanText, achieve higher efectiveness for low values of the privacy
parameter . On the other hand, obfuscation mechanisms based on the embedding obfuscation strategies
perform with high efectiveness only if the formal parameter is high. Finally, the Heuristics show high
nDCG@10 for AEA and the worst results for the FEA mechanism. These results show a misleading
sense of privacy: high results do not imply actual privacy, i.e., the submitted queries are the originals.
        </p>
      </sec>
      <sec id="sec-4-3">
        <title>4.3. Privacy Analysis Using the QuIPU Score</title>
        <p>Table 1 reports the QuIPU scores obtained analysing the Risk vs. Utility on each set of obfuscated
queries of the collections. The results show three distinct patterns that can be traced back to the three
diferent obfuscation strategies. The Sampling-based mechanisms show weaker defences against the
three attackers, and especially against the “active” one, i.e., QuIPU score more negative. In contrast,
the Embedding Perturbation mechanisms are designed to protect user information from attackers,
yielding higher QuIPU scores even against a “motivated” attacker. This suggests that, when using as
DP obfuscation mechanisms, if the user wants to achieve strong actual privacy guarantees against the
QuIA, it should select an obfuscation relying on changing the word embeddings of the queries. Finally,
the heuristics strategies obtain a null QuIPU score due to the stable risk and utility achieved. FEA
reaches a slightly positive QuIPU score against the three attackers, implying that it is impractical for an
attacker to guess the original query even if “motivated” to do so.</p>
        <sec id="sec-4-3-1">
          <title>3https://github.com/Kekkodf/QuIPU_Framework 4https://ir-datasets.com/aol-ia.html</title>
          <p>0.7
0.6
00.5
1
CG@0.4
D
tyn0.3
il
itU0.2
in terms of  = nDCG@10, and the risk  of a successful Query Inference considering diferent adversary models.
Positive values correspond to a better Utility-Privacy trade-of, cf. Section 3.</p>
          <p>Obfuscation</p>
          <p>Strategy
Sampling
Embedding
Perturbation
Heuristics</p>
          <p>Mechanism
CusText
SanText
TEM
WBB
CMP
Mahalanobis
VickreyCMP
VickreyMhl
AEA
FEA</p>
          <p>Lazy Attacker</p>
          <p>Active Attacker</p>
          <p>Motivated Attacker
DL’19</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Conclusion and Future Work</title>
      <p>Assessing the privacy guarantees provided to users during IR tasks remains an open challenge. Relying
solely on a formal privacy analysis considering the mechanism parameters is insuficient for concretely
evaluating the privacy of obfuscation mechanisms. In this study, we introduced the QuIPU framework,
a new benchmark designed to assess actual privacy provided to queries in an obfuscation protocol. We
empirically evaluated the risk that an “honest-but-curious” IR system can accurately infer the original
query from the obfuscated ones received using its queries from the logs. Our findings demonstrate that
strong formal privacy guarantees do not necessarily imply actual privacy protection. In future work,
we plan to explore additional proxy measures to investigate their correlation with the QuIPU score. In
addition, we intend to explore the capabilities of Large Language Models in determining whether or
not a query has been suficiently obfuscated, adopting such models as defensive mechanisms against a
successful QuIA.</p>
    </sec>
    <sec id="sec-6">
      <title>Declaration on Generative AI</title>
      <p>During the preparation of this work, the authors used Grammarly for Readability and Spelling checks.
After using this tool, the authors reviewed and edited the content as needed and took full responsibility
for the publication’s content.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>F.</given-names>
            <surname>L. De Faveri</surname>
          </string-name>
          , G. Faggioli,
          <string-name>
            <given-names>N.</given-names>
            <surname>Ferro</surname>
          </string-name>
          ,
          <article-title>Measuring actual privacy of obfuscated queries in information retrieval</article-title>
          ,
          <source>in: Advances in Information Retrieval: 47th European Conference on Information Retrieval</source>
          ,
          <string-name>
            <surname>ECIR</surname>
          </string-name>
          <year>2025</year>
          , Lucca, Italy, April 6-
          <issue>10</issue>
          ,
          <year>2025</year>
          , Proceedings,
          <string-name>
            <surname>Part</surname>
            <given-names>I</given-names>
          </string-name>
          , SpringerVerlag, Berlin, Heidelberg,
          <year>2025</year>
          , p.
          <fpage>49</fpage>
          -
          <lpage>66</lpage>
          . URL: https://doi.org/10.1007/978-3-
          <fpage>031</fpage>
          -88708-
          <issue>6</issue>
          _4. doi:
          <volume>10</volume>
          .1007/978-3-
          <fpage>031</fpage>
          -88708-
          <issue>6</issue>
          _
          <fpage>4</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>H.</given-names>
            <surname>Le</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Maragh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Ekdale</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>High</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Havens</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Shafiq</surname>
          </string-name>
          ,
          <article-title>Measuring political personalization of google news search</article-title>
          ,
          <source>in: The World Wide Web Conference</source>
          , WWW '19,
          <string-name>
            <surname>Association</surname>
          </string-name>
          for Computing Machinery, New York, NY, USA,
          <year>2019</year>
          , p.
          <fpage>2957</fpage>
          -
          <lpage>2963</lpage>
          . URL: https://doi.org/10.1145/3308558.3313682. doi:
          <volume>10</volume>
          .1145/3308558.3313682.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>E.</given-names>
            <surname>Mustafaraj</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Lurie</surname>
          </string-name>
          ,
          <string-name>
            <surname>C. Devine,</surname>
          </string-name>
          <article-title>The case for voter-centered audits of search engines during political elections</article-title>
          ,
          <source>in: Proceedings of the 2020 Conference on Fairness, Accountability, and Transparency</source>
          , FAT* '
          <volume>20</volume>
          ,
          <string-name>
            <surname>Association</surname>
          </string-name>
          for Computing Machinery, New York, NY, USA,
          <year>2020</year>
          , p.
          <fpage>559</fpage>
          -
          <lpage>569</lpage>
          . URL: https://doi.org/10.1145/3351095.3372835. doi:
          <volume>10</volume>
          .1145/3351095.3372835.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>S.</given-names>
            <surname>Bavadekar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. M.</given-names>
            <surname>Dai</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Davis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Desfontaines</surname>
          </string-name>
          , I. Eckstein,
          <string-name>
            <given-names>K.</given-names>
            <surname>Everett</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Fabrikant</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Flores</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Gabrilovich</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Gadepalli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Glass</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Huang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Kamath</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Kraft</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Kumok</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Marfatia</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Mayer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Miller</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Pearce</surname>
          </string-name>
          ,
          <string-name>
            <given-names>I. M.</given-names>
            <surname>Perera</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Ramachandran</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Raman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Roessler</surname>
          </string-name>
          , I. Shafran,
          <string-name>
            <given-names>T.</given-names>
            <surname>Shekel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Stanton</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Stimes</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Sun</surname>
          </string-name>
          , G. Wellenius,
          <string-name>
            <given-names>M.</given-names>
            <surname>Zoghi</surname>
          </string-name>
          ,
          <string-name>
            <surname>Google</surname>
            <given-names>COVID</given-names>
          </string-name>
          <article-title>-19 search trends symptoms dataset: Anonymization process description (version 1</article-title>
          .0), CoRR abs/
          <year>2009</year>
          .01265 (
          <year>2020</year>
          ). URL: https://arxiv.org/abs/
          <year>2009</year>
          .01265. arXiv:
          <year>2009</year>
          .01265.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>M.</given-names>
            <surname>Damie</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Hahn</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Peter</surname>
          </string-name>
          ,
          <article-title>A highly accurate query-recovery attack against searchable encryption using non-indexed documents</article-title>
          , in: M. D.
          <string-name>
            <surname>Bailey</surname>
          </string-name>
          , R. Greenstadt (Eds.),
          <source>30th USENIX Security Symposium, USENIX Security</source>
          <year>2021</year>
          ,
          <year>August</year>
          11-
          <issue>13</issue>
          ,
          <year>2021</year>
          ,
          <string-name>
            <given-names>USENIX</given-names>
            <surname>Association</surname>
          </string-name>
          ,
          <year>2021</year>
          , pp.
          <fpage>143</fpage>
          -
          <lpage>160</lpage>
          . URL: https://www.usenix.org/conference/usenixsecurity21/presentation/damie.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>A.</given-names>
            <surname>Arampatzis</surname>
          </string-name>
          , G. Drosatos,
          <string-name>
            <given-names>P.</given-names>
            <surname>Efraimidis</surname>
          </string-name>
          ,
          <article-title>A versatile tool for privacy-enhanced web search</article-title>
          , in: P. Serdyukov,
          <string-name>
            <given-names>P.</given-names>
            <surname>Braslavski</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. O.</given-names>
            <surname>Kuznetsov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Kamps</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. M.</given-names>
            <surname>Rüger</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Agichtein</surname>
          </string-name>
          ,
          <string-name>
            <given-names>I.</given-names>
            <surname>Segalovich</surname>
          </string-name>
          , E. Yilmaz (Eds.),
          <source>Advances in Information Retrieval - 35th European Conference on IR Research</source>
          , ECIR
          <year>2013</year>
          , Moscow, Russia, March
          <volume>24</volume>
          -27,
          <year>2013</year>
          . Proceedings, volume
          <volume>7814</volume>
          of Lecture Notes in Computer Science, Springer,
          <year>2013</year>
          , pp.
          <fpage>368</fpage>
          -
          <lpage>379</lpage>
          . URL: https://doi.org/10.1007/978-3-
          <fpage>642</fpage>
          -36973-5_
          <fpage>31</fpage>
          . doi:
          <volume>10</volume>
          .1007/978-3-
          <fpage>642</fpage>
          -36973-5\_
          <fpage>31</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>M.</given-names>
            <surname>Fröbe</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E. O.</given-names>
            <surname>Schmidt</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Hagen</surname>
          </string-name>
          ,
          <article-title>Eficient query obfuscation with keyqueries</article-title>
          , in: J.
          <string-name>
            <surname>He</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Unland</surname>
            ,
            <given-names>E. S.</given-names>
          </string-name>
          <string-name>
            <surname>Jr.</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          <string-name>
            <surname>Tao</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          <string-name>
            <surname>Purohit</surname>
          </string-name>
          , W. van den Heuvel, J. Yearwood, J. Cao (Eds.), WI-IAT '21: IEEE/WIC/ACM International Conference on Web Intelligence,
          <source>Melbourne VIC Australia, December 14 - 17</source>
          ,
          <year>2021</year>
          , ACM,
          <year>2021</year>
          , pp.
          <fpage>154</fpage>
          -
          <lpage>161</lpage>
          . URL: https://doi.org/10.1145/3486622.3493950. doi:
          <volume>10</volume>
          .1145/3486622.3493950.
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>C.</given-names>
            <surname>Dwork</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>McSherry</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Nissim</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Smith</surname>
          </string-name>
          ,
          <article-title>Calibrating noise to sensitivity in private data analysis</article-title>
          , in: S. Halevi, T. Rabin (Eds.),
          <source>Theory of Cryptography</source>
          , Springer Berlin Heidelberg, Berlin, Heidelberg,
          <year>2006</year>
          , pp.
          <fpage>265</fpage>
          -
          <lpage>284</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>O.</given-names>
            <surname>Feyisetan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Kasiviswanathan</surname>
          </string-name>
          ,
          <article-title>Private release of text embedding vectors</article-title>
          , in: Y.
          <string-name>
            <surname>Pruksachatkun</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Ramakrishna</surname>
            ,
            <given-names>K.-W.</given-names>
          </string-name>
          <string-name>
            <surname>Chang</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Krishna</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Dhamala</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Guha</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          Ren (Eds.),
          <source>Proceedings of the First Workshop on Trustworthy Natural Language Processing</source>
          , Association for Computational Linguistics, Online,
          <year>2021</year>
          , pp.
          <fpage>15</fpage>
          -
          <lpage>27</lpage>
          . URL: https://aclanthology.org/
          <year>2021</year>
          .trustnlp-
          <volume>1</volume>
          .3. doi:
          <volume>10</volume>
          .18653/v1/
          <year>2021</year>
          .trustnlp-
          <volume>1</volume>
          .3.
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>H.</given-names>
            <surname>Bo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. H. H.</given-names>
            <surname>Ding</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B. C. M.</given-names>
            <surname>Fung</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Iqbal</surname>
          </string-name>
          ,
          <article-title>ER-AE: diferentially private text generation for authorship anonymization</article-title>
          , in: K.
          <string-name>
            <surname>Toutanova</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Rumshisky</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <string-name>
            <surname>Zettlemoyer</surname>
            , D. HakkaniTür, I. Beltagy,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Bethard</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Cotterell</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Chakraborty</surname>
          </string-name>
          , Y. Zhou (Eds.),
          <source>Proceedings of the</source>
          <year>2021</year>
          <article-title>Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT 2021, Online</article-title>
          , June 6-11,
          <year>2021</year>
          , Association for Computational Linguistics,
          <year>2021</year>
          , pp.
          <fpage>3997</fpage>
          -
          <lpage>4007</lpage>
          . URL: https://doi.org/10.18653/v1/
          <year>2021</year>
          .naacl-main.
          <volume>314</volume>
          . doi:
          <volume>10</volume>
          .18653/V1/
          <year>2021</year>
          .NAACL-MAIN.
          <year>314</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>G.</given-names>
            <surname>Faggioli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Ferro</surname>
          </string-name>
          ,
          <article-title>Query obfuscation for information retrieval through diferential privacy</article-title>
          , in: N.
          <string-name>
            <surname>Goharian</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          <string-name>
            <surname>Tonellotto</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          <string-name>
            <surname>He</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Lipani</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          <string-name>
            <surname>McDonald</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Macdonald</surname>
          </string-name>
          , I. Ounis (Eds.),
          <source>Advances in Information Retrieval - 46th European Conference on Information Retrieval</source>
          ,
          <string-name>
            <surname>ECIR</surname>
          </string-name>
          <year>2024</year>
          , Glasgow, UK, March
          <volume>24</volume>
          -28,
          <year>2024</year>
          , Proceedings,
          <string-name>
            <surname>Part</surname>
            <given-names>I</given-names>
          </string-name>
          , volume
          <volume>14608</volume>
          of Lecture Notes in Computer Science, Springer,
          <year>2024</year>
          , pp.
          <fpage>278</fpage>
          -
          <lpage>294</lpage>
          . URL: https://doi.org/10.1007/978-3-
          <fpage>031</fpage>
          -56027-9_
          <fpage>17</fpage>
          . doi:
          <volume>10</volume>
          .1007/978-3-
          <fpage>031</fpage>
          -56027-9\_
          <fpage>17</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>S.</given-names>
            <surname>Clauß</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Schifner</surname>
          </string-name>
          ,
          <article-title>Structuring anonymity metrics</article-title>
          , in: A.
          <string-name>
            <surname>Juels</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Winslett</surname>
            ,
            <given-names>A</given-names>
          </string-name>
          . Goto (Eds.),
          <source>Proceedings of the 2006 Workshop on Digital Identity Management</source>
          , Alexandria,
          <string-name>
            <surname>VA</surname>
          </string-name>
          , USA, November 3,
          <year>2006</year>
          , ACM,
          <year>2006</year>
          , pp.
          <fpage>55</fpage>
          -
          <lpage>62</lpage>
          . URL: https://doi.org/10.1145/1179529.1179539. doi:
          <volume>10</volume>
          .1145/1179529.1179539.
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Kang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Niu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Tong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , W. Wang,
          <article-title>Input perturbation: A new paradigm between central and local diferential privacy</article-title>
          , CoRR abs/
          <year>2002</year>
          .08570 (
          <year>2020</year>
          ). URL: https://arxiv.org/abs/
          <year>2002</year>
          . 08570. arXiv:
          <year>2002</year>
          .08570.
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>T.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Kishore</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K. Q.</given-names>
            <surname>Weinberger</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Artzi</surname>
          </string-name>
          , Bertscore:
          <article-title>Evaluating text generation with BERT</article-title>
          ,
          <source>in: 8th International Conference on Learning Representations, ICLR</source>
          <year>2020</year>
          ,
          <string-name>
            <given-names>Addis</given-names>
            <surname>Ababa</surname>
          </string-name>
          , Ethiopia,
          <source>April 26-30</source>
          ,
          <year>2020</year>
          , OpenReview.net,
          <year>2020</year>
          . URL: https://openreview.net/forum?id= SkeHuCVFDr.
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>J.</given-names>
            <surname>Domingo-Ferrer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Sánchez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Blanco-Justicia</surname>
          </string-name>
          ,
          <article-title>The limits of diferential privacy (and its misuse in data release and machine learning</article-title>
          ),
          <source>Commun. ACM</source>
          <volume>64</volume>
          (
          <year>2021</year>
          )
          <fpage>33</fpage>
          -
          <lpage>35</lpage>
          . URL: https: //doi.org/10.1145/3433638. doi:
          <volume>10</volume>
          .1145/3433638.
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>J.</given-names>
            <surname>Mattern</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Weggenmann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Kerschbaum</surname>
          </string-name>
          ,
          <article-title>The limits of word level diferential privacy</article-title>
          , in: M.
          <string-name>
            <surname>Carpuat</surname>
            , M. de Marnefe,
            <given-names>I. V. M.</given-names>
          </string-name>
          <string-name>
            <surname>Ruíz</surname>
          </string-name>
          (Eds.),
          <source>Findings of the Association for Computational Linguistics: NAACL</source>
          <year>2022</year>
          , Seattle, WA,
          <string-name>
            <surname>United</surname>
            <given-names>States</given-names>
          </string-name>
          ,
          <source>July 10-15</source>
          ,
          <year>2022</year>
          , Association for Computational Linguistics,
          <year>2022</year>
          , pp.
          <fpage>867</fpage>
          -
          <lpage>881</lpage>
          . URL: https://doi.org/10.18653/v1/
          <year>2022</year>
          .findings-naacl.
          <volume>65</volume>
          . doi:
          <volume>10</volume>
          .18653/V1/
          <year>2022</year>
          .FINDINGS-NAACL.
          <year>65</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>F. L. D.</given-names>
            <surname>Faveri</surname>
          </string-name>
          , G. Faggioli,
          <string-name>
            <given-names>N.</given-names>
            <surname>Ferro</surname>
          </string-name>
          ,
          <article-title>Beyond the parameters: Measuring actual privacy in obfuscated texts</article-title>
          , in: K. Roitero,
          <string-name>
            <given-names>M.</given-names>
            <surname>Viviani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Maddalena</surname>
          </string-name>
          , S. Mizzaro (Eds.),
          <source>Proceedings of the 14th Italian Information Retrieval Workshop</source>
          , Udine, Italy, September 5-
          <issue>6</issue>
          ,
          <year>2024</year>
          , volume
          <volume>3802</volume>
          <source>of CEUR Workshop Proceedings, CEUR-WS.org</source>
          ,
          <year>2024</year>
          , pp.
          <fpage>53</fpage>
          -
          <lpage>57</lpage>
          . URL: https://ceur-ws.
          <source>org/</source>
          Vol-
          <volume>3802</volume>
          /paper5.pdf.
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>G.</given-names>
            <surname>Duncan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Keller-McNulty</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Stokes</surname>
          </string-name>
          ,
          <article-title>Disclosure risk vs. data utility: The ru confidentiality map</article-title>
          ,
          <source>A Los Alamos National Laboratory Technical Report LA-UR-01-6428</source>
          (
          <year>2001</year>
          )
          <fpage>1</fpage>
          -
          <lpage>30</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>R.</given-names>
            <surname>Shokri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Stronati</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Song</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Shmatikov</surname>
          </string-name>
          ,
          <article-title>Membership inference attacks against machine learning models</article-title>
          ,
          <source>in: 2017 IEEE Symposium on Security and Privacy</source>
          ,
          <string-name>
            <surname>SP</surname>
          </string-name>
          <year>2017</year>
          , San Jose, CA, USA, May
          <volume>22</volume>
          -26,
          <year>2017</year>
          , IEEE Computer Society,
          <year>2017</year>
          , pp.
          <fpage>3</fpage>
          -
          <lpage>18</lpage>
          . URL: https://doi.org/10.1109/SP.
          <year>2017</year>
          .
          <volume>41</volume>
          . doi:
          <volume>10</volume>
          .1109/SP.
          <year>2017</year>
          .
          <volume>41</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>I.</given-names>
            <surname>Wagner</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Eckhof</surname>
          </string-name>
          ,
          <article-title>Technical privacy metrics: A systematic survey</article-title>
          ,
          <source>ACM Comput. Surv</source>
          .
          <volume>51</volume>
          (
          <year>2018</year>
          )
          <volume>57</volume>
          :
          <fpage>1</fpage>
          -
          <lpage>57</lpage>
          :
          <fpage>38</fpage>
          . URL: https://doi.org/10.1145/3168389. doi:
          <volume>10</volume>
          .1145/3168389.
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>S.</given-names>
            <surname>Sousa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Kern</surname>
          </string-name>
          ,
          <article-title>How to keep text private? A systematic review of deep learning methods for privacy-preserving natural language processing</article-title>
          ,
          <source>Artif. Intell. Rev</source>
          .
          <volume>56</volume>
          (
          <year>2023</year>
          )
          <fpage>1427</fpage>
          -
          <lpage>1492</lpage>
          . URL: https://doi.org/10.1007/s10462-022-10204-6. doi:
          <volume>10</volume>
          .1007/S10462-022-10204-6.
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <surname>I. Habernal</surname>
          </string-name>
          ,
          <article-title>When diferential privacy meets NLP: the devil is in the detail</article-title>
          , in: M.
          <string-name>
            <surname>Moens</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          <string-name>
            <surname>Huang</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <string-name>
            <surname>Specia</surname>
          </string-name>
          , S. W. Yih (Eds.),
          <source>Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, EMNLP</source>
          <year>2021</year>
          , Virtual Event / Punta Cana, Dominican Republic,
          <fpage>7</fpage>
          -
          <issue>11</issue>
          <year>November</year>
          ,
          <year>2021</year>
          , Association for Computational Linguistics,
          <year>2021</year>
          , pp.
          <fpage>1522</fpage>
          -
          <lpage>1528</lpage>
          . URL: https: //doi.org/10.18653/v1/
          <year>2021</year>
          .emnlp-main.
          <volume>114</volume>
          . doi:
          <volume>10</volume>
          .18653/V1/
          <year>2021</year>
          .EMNLP-MAIN.
          <year>114</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <given-names>S. J.</given-names>
            <surname>Meisenbacher</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Nandakumar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Klymenko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Matthes</surname>
          </string-name>
          ,
          <article-title>A comparative analysis of word-level metric diferential privacy: Benchmarking the privacy-utility trade-of</article-title>
          , in: N.
          <string-name>
            <surname>Calzolari</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Kan</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          <string-name>
            <surname>Hoste</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Lenci</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Sakti</surname>
          </string-name>
          , N. Xue (Eds.),
          <source>Proceedings of the 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation, LREC/COLING</source>
          <year>2024</year>
          ,
          <volume>20</volume>
          -25 May,
          <year>2024</year>
          , Torino, Italy, ELRA and ICCL,
          <year>2024</year>
          , pp.
          <fpage>174</fpage>
          -
          <lpage>185</lpage>
          . URL: https://aclanthology.org/
          <year>2024</year>
          . lrec-main.
          <volume>16</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [24]
          <string-name>
            <given-names>O.</given-names>
            <surname>Feyisetan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Balle</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Drake</surname>
          </string-name>
          , T. Diethe,
          <article-title>Privacy- and utility-preserving textual analysis via calibrated multivariate perturbations</article-title>
          , in: J.
          <string-name>
            <surname>Caverlee</surname>
            ,
            <given-names>X. B.</given-names>
          </string-name>
          <string-name>
            <surname>Hu</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Lalmas</surname>
          </string-name>
          , W. Wang (Eds.),
          <source>Proceedings of the 13th International Conference on Web Search and Data Mining, ACM</source>
          ,
          <year>2020</year>
          , pp.
          <fpage>178</fpage>
          -
          <lpage>186</lpage>
          . doi:
          <volume>10</volume>
          .1145/3336191.3371856.
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          [25]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Aggarwal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Feyisetan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Teissier</surname>
          </string-name>
          ,
          <article-title>A diferentially private text perturbation method using regularized mahalanobis metric</article-title>
          ,
          <source>in: Proceedings of the Second Workshop on Privacy in NLP, Association for Computational Linguistics</source>
          ,
          <year>2020</year>
          . doi:
          <volume>10</volume>
          .18653/v1/
          <year>2020</year>
          .privatenlp-
          <volume>1</volume>
          .2.
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          [26]
          <string-name>
            <given-names>A.</given-names>
            <surname>Blanco-Justicia</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Sánchez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Domingo-Ferrer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Muralidhar</surname>
          </string-name>
          ,
          <article-title>A critical review on the use (and misuse) of diferential privacy in machine learning</article-title>
          ,
          <source>ACM Comput. Surv</source>
          .
          <volume>55</volume>
          (
          <year>2023</year>
          )
          <volume>160</volume>
          :
          <fpage>1</fpage>
          -
          <lpage>160</lpage>
          :
          <fpage>16</fpage>
          . URL: https://doi.org/10.1145/3547139. doi:
          <volume>10</volume>
          .1145/3547139.
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          [27]
          <string-name>
            <given-names>S.</given-names>
            <surname>Zimmerman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Thorpe</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Fox</surname>
          </string-name>
          , U. Kruschwitz,
          <article-title>Investigating the interplay between searchers' privacy concerns and their search behavior</article-title>
          , in: B.
          <string-name>
            <surname>Piwowarski</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Chevalier</surname>
            , É. Gaussier,
            <given-names>Y.</given-names>
          </string-name>
          <string-name>
            <surname>Maarek</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Nie</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Scholer</surname>
          </string-name>
          (Eds.),
          <source>Proceedings of the 42nd International ACM SIGIR Conference on Research and Development in Information Retrieval</source>
          ,
          <string-name>
            <surname>SIGIR</surname>
          </string-name>
          <year>2019</year>
          , Paris, France,
          <source>July 21-25</source>
          ,
          <year>2019</year>
          , ACM,
          <year>2019</year>
          , pp.
          <fpage>953</fpage>
          -
          <lpage>956</lpage>
          . URL: https://doi.org/10.1145/3331184.3331280. doi:
          <volume>10</volume>
          .1145/3331184.3331280.
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          [28]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <article-title>A survey on diferential privacy for unstructured data content</article-title>
          ,
          <source>ACM Comput. Surv</source>
          .
          <volume>54</volume>
          (
          <year>2022</year>
          )
          <volume>207</volume>
          :
          <fpage>1</fpage>
          -
          <lpage>207</lpage>
          :
          <fpage>28</fpage>
          . URL: https://doi.org/10.1145/3490237. doi:
          <volume>10</volume>
          .1145/3490237.
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          [29]
          <string-name>
            <given-names>O.</given-names>
            <surname>Klymenko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Meisenbacher</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Matthes</surname>
          </string-name>
          ,
          <article-title>Diferential privacy in natural language processing the story so far</article-title>
          , in: O.
          <string-name>
            <surname>Feyisetan</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Ghanavati</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Thaine</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          <string-name>
            <surname>Habernal</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Mireshghallah</surname>
          </string-name>
          (Eds.),
          <source>Proceedings of the Fourth Workshop on Privacy in Natural Language Processing</source>
          , Association for Computational Linguistics, Seattle, United States,
          <year>2022</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>11</lpage>
          . URL: https://aclanthology.org/
          <year>2022</year>
          .privatenlp-
          <volume>1</volume>
          .1. doi:
          <volume>10</volume>
          .18653/v1/
          <year>2022</year>
          .privatenlp-
          <volume>1</volume>
          .1.
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          [30]
          <string-name>
            <surname>F. L. De Faveri</surname>
            , G. Faggioli,
            <given-names>N.</given-names>
          </string-name>
          <string-name>
            <surname>Ferro</surname>
          </string-name>
          , Words Blending Boxes.
          <article-title>Obfuscating Queries in Information Retrieval using Diferential Privacy</article-title>
          ,
          <source>CoRR abs/2405</source>
          .09306 (
          <year>2024</year>
          ). URL: https://doi.org/10.48550/ arXiv.2405.09306. doi:
          <volume>10</volume>
          .48550/ARXIV.2405.09306. arXiv:
          <volume>2405</volume>
          .
          <fpage>09306</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref31">
        <mixed-citation>
          [31]
          <string-name>
            <given-names>C.</given-names>
            <surname>Clifton</surname>
          </string-name>
          , T. Tassa,
          <article-title>On syntactic anonymity and diferential privacy</article-title>
          , in: C. Y.
          <string-name>
            <surname>Chan</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Lu</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          <string-name>
            <surname>Nørvåg</surname>
          </string-name>
          , E. Tanin (Eds.),
          <source>Workshops Proceedings of the 29th IEEE International Conference on Data Engineering, ICDE</source>
          <year>2013</year>
          , Brisbane, Australia, April 8-
          <issue>12</issue>
          ,
          <year>2013</year>
          , IEEE Computer Society,
          <year>2013</year>
          , pp.
          <fpage>88</fpage>
          -
          <lpage>93</lpage>
          . URL: https://doi.org/10.1109/ICDEW.
          <year>2013</year>
          .
          <volume>6547433</volume>
          . doi:
          <volume>10</volume>
          .1109/ICDEW.
          <year>2013</year>
          .
          <volume>6547433</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref32">
        <mixed-citation>
          [32]
          <string-name>
            <given-names>J.</given-names>
            <surname>Hsu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Gaboardi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Haeberlen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Khanna</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Narayan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B. C.</given-names>
            <surname>Pierce</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Roth</surname>
          </string-name>
          ,
          <article-title>Diferential privacy: An economic method for choosing epsilon</article-title>
          ,
          <source>in: IEEE 27th Computer Security Foundations Symposium, CSF</source>
          <year>2014</year>
          , Vienna, Austria,
          <fpage>19</fpage>
          -
          <issue>22</issue>
          <year>July</year>
          ,
          <year>2014</year>
          , IEEE Computer Society,
          <year>2014</year>
          , pp.
          <fpage>398</fpage>
          -
          <lpage>410</lpage>
          . URL: https://doi.org/10.1109/CSF.
          <year>2014</year>
          .
          <volume>35</volume>
          . doi:
          <volume>10</volume>
          .1109/CSF.
          <year>2014</year>
          .
          <volume>35</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref33">
        <mixed-citation>
          [33]
          <string-name>
            <given-names>G. A.</given-names>
            <surname>Miller</surname>
          </string-name>
          ,
          <article-title>Wordnet: A lexical database for english</article-title>
          ,
          <source>Commun. ACM</source>
          <volume>38</volume>
          (
          <year>1995</year>
          )
          <fpage>39</fpage>
          -
          <lpage>41</lpage>
          . URL: https://doi.org/10.1145/219717.219748. doi:
          <volume>10</volume>
          .1145/219717.219748.
        </mixed-citation>
      </ref>
      <ref id="ref34">
        <mixed-citation>
          [34]
          <string-name>
            <given-names>K.</given-names>
            <surname>Chatzikokolakis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. E.</given-names>
            <surname>Andrés</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N. E.</given-names>
            <surname>Bordenabe</surname>
          </string-name>
          ,
          <string-name>
            <surname>C.</surname>
          </string-name>
          <article-title>Palamidessi, Broadening the scope of diferential privacy using metrics</article-title>
          , in: E. D.
          <string-name>
            <surname>Cristofaro</surname>
          </string-name>
          , M. K. Wright (Eds.),
          <string-name>
            <surname>Privacy Enhancing</surname>
          </string-name>
          Technologies - 13th
          <source>International Symposium, PETS</source>
          <year>2013</year>
          ,
          <article-title>Bloomington</article-title>
          , IN, USA, July
          <volume>10</volume>
          -
          <issue>12</issue>
          ,
          <year>2013</year>
          . Proceedings, volume
          <volume>7981</volume>
          of Lecture Notes in Computer Science, Springer,
          <year>2013</year>
          , pp.
          <fpage>82</fpage>
          -
          <lpage>102</lpage>
          . URL: https://doi.org/10.1007/978-3-
          <fpage>642</fpage>
          -39077-
          <issue>7</issue>
          _5. doi:
          <volume>10</volume>
          .1007/978-3-
          <fpage>642</fpage>
          -39077-7\_5.
        </mixed-citation>
      </ref>
      <ref id="ref35">
        <mixed-citation>
          [35]
          <string-name>
            <given-names>S.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Mo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.-Y.</given-names>
            <surname>Nie</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Cui</surname>
          </string-name>
          ,
          <article-title>A customized text sanitization mechanism with diferential privacy</article-title>
          , in: A.
          <string-name>
            <surname>Rogers</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Boyd-Graber</surname>
          </string-name>
          , N. Okazaki (Eds.),
          <source>Findings of the Association for Computational Linguistics: ACL</source>
          <year>2023</year>
          ,
          <article-title>Association for Computational Linguistics</article-title>
          , Toronto, Canada,
          <year>2023</year>
          , pp.
          <fpage>5747</fpage>
          -
          <lpage>5758</lpage>
          . URL: https://aclanthology.org/
          <year>2023</year>
          .findings-acl.
          <volume>355</volume>
          . doi:
          <volume>10</volume>
          .18653/v1/
          <year>2023</year>
          .findings-acl.
          <volume>355</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref36">
        <mixed-citation>
          [36]
          <string-name>
            <given-names>X.</given-names>
            <surname>Yue</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Du</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Sun</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. S. M.</given-names>
            <surname>Chow</surname>
          </string-name>
          ,
          <article-title>Diferential privacy for text analytics via natural text sanitization</article-title>
          , in: C.
          <string-name>
            <surname>Zong</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Xia</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          <string-name>
            <surname>Li</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Navigli</surname>
          </string-name>
          (Eds.),
          <article-title>Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021</article-title>
          ,
          <article-title>Association for Computational Linguistics</article-title>
          , Online,
          <year>2021</year>
          , pp.
          <fpage>3853</fpage>
          -
          <lpage>3866</lpage>
          . URL: https://aclanthology.org/
          <year>2021</year>
          .findings-acl.
          <volume>337</volume>
          . doi:
          <volume>10</volume>
          .18653/ v1/
          <year>2021</year>
          .findings-acl.
          <volume>337</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref37">
        <mixed-citation>
          [37]
          <string-name>
            <given-names>R. S.</given-names>
            <surname>Carvalho</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Vasiloudis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Feyisetan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <article-title>TEM: high utility metric diferential privacy on text</article-title>
          , in: S.
          <string-name>
            <surname>Shekhar</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          <string-name>
            <surname>Zhou</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          <string-name>
            <surname>Chiang</surname>
          </string-name>
          , G. Stiglic (Eds.),
          <source>Proceedings of the 2023 SIAM International Conference on Data Mining, SDM</source>
          <year>2023</year>
          ,
          <article-title>Minneapolis-St</article-title>
          . Paul Twin Cities, MN, USA, April
          <volume>27</volume>
          -
          <issue>29</issue>
          ,
          <year>2023</year>
          , SIAM,
          <year>2023</year>
          , pp.
          <fpage>883</fpage>
          -
          <lpage>890</lpage>
          . URL: https://doi.org/10.1137/1.9781611977653.
          <year>ch99</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>