<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Federated Recommender Systems with Learning to Rank</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Vito Walter Anelli</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Yashar Deldjoo</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Tommaso Di Noia</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Antonio Ferrara</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Fedelucio Narducci</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Politecnico di Bari</institution>
          ,
          <addr-line>Via E. Orabona, 4, 70126 Bari</addr-line>
          ,
          <country country="IT">Italy</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Recommendation services are extensively adopted in several user-centered applications as a tool to alleviate the information overload problem and help users in orienteering in a vast space of possible choices. In such scenarios, data ownership is a crucial concern since users may not be willing to share their sensitive preferences (e.g., visited locations, read books, bought items) with a central server. Unfortunately, data harvesting and collection is at the basis of modern, state-of-the-art approaches to recommendation. To address this issue, we extend Federated Pair-wise Learning (FPL), an architecture in which users collaborate in training a central factorization model while controlling the amount of sensitive data leaving their devices. The proposed approach implements pair-wise learning-to-rank optimization by following the Federated Learning principles, conceived originally to mitigate the privacy risks of traditional machine learning.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Federated Learning</kwd>
        <kwd>Recommender Systems</kwd>
        <kwd>Information Retrieval</kwd>
        <kwd>Learning to Rank</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Collaborative filtering (CF) models have been mainstream research in the recommender system
(RS) community over the last two decades thanks to their performance accuracy [
        <xref ref-type="bibr" rid="ref1 ref2">1, 2</xref>
        ]. Among
them, a prominent class uses the matrix factorization (MF) approach as the inference model.
The MF model’s main aim is to uncover user and item latent representations whose linear
interaction explains observed feedback. To date, the majority of existing MF models are trained
in a centralized fashion causing several concerns about the privacy of user data.
      </p>
      <p>
        The consequent data scarcity dilemma can thereby jeopardize the training of MF models.
Training high-quality MF models strongly relies on suficient in-domain interaction data to
ensure that enough co-occurrence information exists to shape similar behavioral/preference
patterns in a user community. Although cross-domain recommendation approaches allow
combating the issue of data scarcity, their applicability largely depends upon the availability of
data providers that can collect/supply cross-domain in their platform (e.g., Amazon). However,
these approaches remain out of focus in this work. In recent years, federated learning (FL) was
proposed by Google as a mean to ofer a privacy-by-design solution [
        <xref ref-type="bibr" rid="ref3 ref4 ref5">3, 4, 5</xref>
        ] for machine-learned
models. Federated learning aims to meet ML privacy shortcomings by horizontally distributing
the model’s training over user devices; thus, clients exploit private data without sharing them [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ].
Despite its original formulation, the FL concept is extended to a more comprehensive idea
of privacy-preserving, decentralized, collaborative ML techniques [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] where diferent data
partitions share the same feature space (horizontal federation) or not (vertical federation). Weiss
et al. [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] state that privacy can be preserved by limiting data collection, which is one of the
main privacy concerns [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]. Indeed, the accuracy of RS based on the CF paradigm is strictly
dependent on the amount of user preferences available.
      </p>
      <p>
        Our idea is to put users in control of their sensitive data by allowing them to choose the
amount of information to share with the server. Hence, if data collection from the server side is
reduced, other threats related to retention, sales, and unauthorized data browsing are limited.
The proposed system extends FPL [
        <xref ref-type="bibr" rid="ref10 ref9">9, 10</xref>
        ] (short for Federated Pair-wise Learning), is a federated
factorization model for collaborative recommendation1. It extends state-of-the-art factorization
approaches to build a RS that puts users in control of their sensitive data. Users participating in
the federation process can decide if and to which extent they are willing to disclose their sensitive
private data (i.e., what they liked/consumed). FPL mainly leverages not-sensitive information
(e.g., places the user has not visited) – which can be large and non-sensitive – to reach a
competitive accuracy and, at the same time, respect a satisfactory balance between accuracy and
privacy. We have carried out extensive experiments on real-world datasets [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] in the Point of
Interest (PoI) domain by considering the accuracy of recommendation and diversity metrics. The
experimental evaluation shows that FPL can provide high-quality recommendations, putting
the user in control of the amount of sensitive data to share.
      </p>
    </sec>
    <sec id="sec-2">
      <title>2. Approach</title>
      <p>
        In this section, after a brief introduction of background technologies, we extend FPL [
        <xref ref-type="bibr" rid="ref10 ref9">9, 10</xref>
        ]
(depicted in Fig. 1). To the best of our knowledge, FPL is the first attempt to put pair-wise
optimization in federated recommender systems and give the users the possibility to select the
trade-of between data disclosure and the recommendation utility.
      </p>
      <sec id="sec-2-1">
        <title>2.1. Background Technologies</title>
        <p>
          Federated Learning. Federated learning (FL) is a paradigm initially envisioned by Google [
          <xref ref-type="bibr" rid="ref12 ref3 ref5">3,
12, 5</xref>
          ] to train a machine-learning model from data distributed among a loose federation of users’
devices (e.g., personal mobile phones). The rationale is to face the increasing issues of ownership
and locality of data to mitigate the privacy risks (and leaks) resulting from centralized machine
learning [
          <xref ref-type="bibr" rid="ref13 ref14">13, 14</xref>
          ]. In particular, given Θ denoting the parameters of a machine learning model,
we consider a learning scenario where the objective is to minimize a generic loss function (Θ).
FL is a learning paradigm in which the users  ∈  of a federation collaborate to solve the
learning problem under the coordination of a central server  without sharing or exchanging
1A public implementation of FPL is available at https://github.com/sisinflab/FedBPR/.
their raw data with . From an algorithmic point of view, we start with  sharing Θ with the
federation of devices. Then, specific methods solve a local optimization problem on the single
device, i.e., using its data, and exploiting Θ. Afterwards, the client shares the parameters of its
local model with . The parameters provided by the clients are used to update Θ, which is sent
back to the devices in a new iteration step.
        </p>
      </sec>
      <sec id="sec-2-2">
        <title>Factorization Models and Pair-Wise Recommendation A recommendation problem</title>
        <p>over a set of users  and a set of items ℐ is defined as the activity of finding for each user  ∈ 
an item  ∈ ℐ that maximizes a utility function  :  × ℐ → R. In this context, X ∈ R||×|ℐ|
is the user-item matrix containing for each  an explicit or implicit feedback (e.g., rating or
check-in, respectively) of user  ∈  for item  ∈ ℐ. In the work at hand, an implicit feedback
scenario is considered — i.e., feedback is, e.g., purchases, visits, clicks, views, check-ins —, with
X containing binary values. Therefore,  = 1 and  = 0 denote either user  has consumed
or not item , respectively.</p>
        <p>
          In FPL, the underlying data model is a Factorization model, inspired by MF [
          <xref ref-type="bibr" rid="ref15">15</xref>
          ], a
recommendation model that became popular in the last decade thanks to its state-of-the-art recommendation
accuracy [
          <xref ref-type="bibr" rid="ref16">16</xref>
          ]. This technique aims to build a model Θ in which each user  and each item  is
represented by the embedding vectors p and q, respectively, in the shared latent space R .
The algorithm relies on the assumption that X can be factorized such that the dot product
between p and q can explain any observed user-item interaction , and that any non-observed
interaction can be estimated as ^(Θ) = (Θ) + p (Θ) · q(Θ) where  is a term denoting
the bias of the item . Among pair-wise approaches for learning-to-rank the items of a catalog,
Bayesian Personalized Ranking (BPR) [
          <xref ref-type="bibr" rid="ref17">17</xref>
          ] is one of the most broadly adopted, thanks to its
capabilities to correctly rank with acceptable computational complexity. In detail, given a training
set defined by  = {(, , ) |  = 1 ∧  = 0}, BPR solves the optimization problem via
the criterion mΘax ∑︀(,,)∈ ln  (^ (Θ)) −  ‖Θ‖2, where ^ (Θ) = ^(Θ) − ^ (Θ) is a
real value modeling the relation between user , item  and item ,  (· ) is the sigmoid function,
and  is a model-specific regularization parameter to prevent overfitting.
        </p>
        <p>Pair-wise optimization can be applied to a wide range of recommendation models, included
factorization. Hereafter, we denote the model Θ = ⟨P, Q, b⟩, where P ∈ R||×  is a matrix
whose -th row corresponds to the vector p, and Q ∈ R|ℐ|×  is a matrix in which the -th
row corresponds to the vector q. Finally, b ∈ R|ℐ| is a vector whose -th element corresponds
to the value .</p>
      </sec>
      <sec id="sec-2-3">
        <title>2.2. FPL: Federated Pair-wise Learning for Recommendation</title>
        <p>
          Following the aforementioned federated learning principles, let us assume that users in 
consume items from a catalog ℐ and give feedback about them.  is aware of the whole catalog
ℐ, while only user  knows her own set of consumed items. Given these conditions, the
classic BPR-MF learning procedure [
          <xref ref-type="bibr" rid="ref17">17</xref>
          ] for model training can not be applied to the federated
learning scheme [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ]. Instead, we propose a novel learning paradigm (depicted in Figure 1) that
is executed for a number  of epochs and works by rounds of communication that envisages
Distribution→Computation→Transmission→Aggregation sequences between the server
and the clients.
        </p>
        <p>Training data Ku</p>
        <p>Client u</p>
        <p>Client-side</p>
        <p>User-Factor
Vector (pu)</p>
        <p>BPR
Optimization
Δ(
Δ(
)
)</p>
        <p>Item-Factor</p>
        <p>Matrix (Q)
Δ(
)
Δ(
)
Δ(
Δ(
)
)
Δ(
)</p>
        <p>Server S</p>
        <p>Server-side
Σ</p>
        <p>In the FPL setting, a global model Θ is built on  such that Θ = ⟨Q, b⟩, where Q ∈ R|ℐ|× 
and b ∈ R|ℐ| are the item-factor matrix and the bias vector (introduced in Section 2.1). On the
other hand, on each device in the federation, FPL builds a model Θ = ⟨p⟩, which corresponds
to the representation of user  in a latent space of dimensionality  . It is noteworthy that, in
FPL, only user  holds the embedding vector p; therefore, each user  autonomously computes
her personalized item ranking, by combining the global model Θ, sent by  to the devices in
the federation, with her local model Θ. In such a setting, each user  holds her own private
feedback dataset x ∈ Rℐ , which — compared with a centralized recommender system —
corresponds to the -th row of matrix X. Each FPL client  hosts a user-specific training set
 :  × ℐ × ℐ defined by  = {(, , ) |  = 1 ∧  = 0}, where  represents the -th
element of . Please note that, in the following, we refer to + = ∑︀∈ |{ |  = 1}| as
the number of positive interactions.</p>
        <p>The number of rounds of communication performed in each learning epoch is a parameter
denoted by the symbol rpe (round-per-epoch). Each round of communication is envisioned as a
four-step protocol, described in the following.</p>
        <p>1. Distribution.  randomly selects a subset of users  − ⊆ 
Θ.
and delivers them the model
2. Computation. Each user  generates  triples (, , ) from her dataset  and for each
of them performs BPR stochastic optimization to compute the updates for the local p
vector of Θ, and for p, , p , and  of the received Θ, following:
Δ =</p>
        <p>− ^ 
1 + − ^ ·  ^ − ,
with

 ^ =
⎧(q − q ) if  = p,
⎪
⎪
⎪⎪⎪⎪p if  = q,
⎨</p>
        <p>− p
⎪⎪⎪⎪1 if  = ,
⎪
⎪⎩− 1 if  =  .</p>
        <p>
          if  = q , (1)
It is worth noticing that Rendle [
          <xref ref-type="bibr" rid="ref17">17</xref>
          ] suggests, in a centralized scenario, to adopt a uniform
distribution (over ) to choose the training triples randomly. The purpose is to avoid data
is traversed item-wise or user-wise, since this may lead to slow convergence. Conversely,
in a federated approach, we required to train the model user-wise since the training of
each round of communication is performed separately on each client  knowing only
data in . This is the reason why, in FPL, the designer can control of the number of
triples  used for training, to tune the degree of local computation — i.e., how much the
sampling is user-wise traversing.
3. Transmission. The clients in  − send back to  a portion of the updates for the computed
item factor vector and item bias. More in detail, since the training output of a triple (, , )
in BPR lets the server distinguish the consumed item  from the non-consumed one 
(for example just by analyzing the positive and the negative sign of Δ and Δ ), while
they show the same absolute value, we argue that sending all the updates computed
by  may allow  to reconstruct  thus raising a privacy issue. Since our primary
goal is to put users in control of their data, FPL proposes a solution to overcome this
vulnerability. By sending the sole update (Δq , Δ ) of each training triples (, , ), user
 would share with  indistinguishably negative or missing values, which are assumed to
be non-sensitive data. Furthermore, in FPL we introduce the parameter  , which allows
users to control of the number of consumed items to share with the central server . The
parameter  works as a probability that clients send also a specific positive item update
(Δq, Δ) in addition to (Δq , Δ ).
4. Global aggregation.  aggregates all the received updates in Q and b to build the new
model Θ ← Θ +  ∑︀∈− ΔΘ, with  being the learning rate (each row of the
matrix Q and each element of b is updated by summing up the contribution of all clients
in  − for the corresponding item).
        </p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Experimental Setup</title>
      <p>In this section, we introduce the experimental setting designed to analyze the performance
of FPL. To this extent, we introduce the choice of the datasets with a brief analysis of their
characteristics. Then, we describe the state-of-the-art algorithms we have involved. For the
sake of reproducibility, for each method, we report the explored hyper-parameters in a specific
section. Lastly, we present the evaluation protocol, and the metrics considered in the study.</p>
      <sec id="sec-3-1">
        <title>3.1. Datasets</title>
        <p>
          The evaluation of FPL needs to meet some particular constraints: the availability of transaction
data to obtain a reliable experimental setting and a domain that guarantees the presence of
data the user may prefer to protect. In our view, the optimal domain would be that of the
Point-of-Interest (PoI), which concerns data that users usually perceive as sensitive. Among
the many available datasets, we think that a very good candidate is the Foursquare dataset [
          <xref ref-type="bibr" rid="ref11">11</xref>
          ].
Indeed, it is often considered as a reference for evaluating PoI recommendation models.
        </p>
        <p>
          To mimic a federation of devices in a single country, we have extracted check-ins for three
countries, namely Brazil, Canada, and Italy. Since our only constraint was to obtain datasets
with diferent size/sparsity characteristics, we took the liberty of choosing three countries
of recent RecSys conference venues. To fairly evaluate FPL against the baselines, we have
kept users with more than 20 interactions2. Moreover, we have split the datasets by adopting
a realistic temporal hold-out 80-20 splitting on a per-user basis [
          <xref ref-type="bibr" rid="ref18 ref19">18, 19</xref>
          ]. Table 1 shows the
characteristics of the resulting training sets.
        </p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Baselines</title>
        <p>
          To evaluate the eficacy of FPL, we have conducted the experiments by considering
nonpersonalized methods (random and most popular recommendation), and diferent
recommendation approaches, including the centralized BPR-MF implementation [
          <xref ref-type="bibr" rid="ref17">17</xref>
          ], VAE [
          <xref ref-type="bibr" rid="ref20">20</xref>
          ], and
FCF [21], which is, to date, the only federated recommendation approach based on MF (since
no source code is available, we reimplemented and considered it in the reader’s interest). To
evaluate the impact of feedback deprivation on recommendation accuracy, we have
evaluated diferent values of  in the range [0.0, 1.0]. We remember that  = 0.0 means  is not
sharing any update (Δq, Δ) with  regarding her positive items feedback, while  = 1.0
means  is sharing the updates on all positive items. Hence, we have considered four diferent
configurations regarding computation and communication:
• sFPL: it aims to reproduce the stochastic learning approach of centralized factorization
model with pair-wise learning, where the central model is updated sequentially; therefore,
we set | − | = 1 to involve just one random client per round, and it extracts solely one
triple (, , ) from its dataset ( = 1) for the training phase;
2The limitations of the Collaborative Filtering in a cold-start user setting are well-known in the literature.
• sFPL+: we increase client local computation by raising to + the number of triples 
||
extracted from  by each client involved in the round of communication;
• pFPL: we enable parallelism by involving all clients in each round of communication
( − =  ); we keep  = 1;
• pFPL+: we extend pFPL by letting each client sample  = + triples from ; the
||
rationale is that the overall training samples are exactly +, as in centralized BPR-MF.
In Rendle et al. [
          <xref ref-type="bibr" rid="ref17">17</xref>
          ], authors suggest to set the number of triples in one epoch of BPR to +,
which corresponds to the number of optimizations steps. A particular choice is to randomly
sampling  = + triples per user. To make a federated training epoch of FPL comparable
to BPR and amo|ng| diferent configurations, we set rpe to obtain always the same number of
interactions  between clients and server in one epoch. This value is equal to the overall number
of optimization steps in one epoch of the centralized pair-wise learning. In detail, we set  = +
and then rpe =  · | − | which results in  = + for sFPL, and  = + for pFPL.
|− |
        </p>
      </sec>
      <sec id="sec-3-3">
        <title>3.3. Reproducibility</title>
        <p>For the splitting strategy, we have adopted a temporal hold-out 80/20 to separate our datasets
in training and test set. Moreover, to find the most promising learning rate  , we have further
split the training set, adopting a temporal hold-out 80/20 strategy on a user basis to extract
her validation set. VAE has been trained by considering three autoencoder topologies, with
the following number of neurons per layer: 200-100-200, 300-100-300, 600-200-600. We have
chosen candidate models by considering the best models after training for 50, 100, and 200
epochs, respectively. For the factorization models, we have performed a grid search in
BPRMF for  ∈ {0.005, 0.05, 0.5} varying the number of latent factors in {10, 20, 50}. Then, to
ensure a fair comparison, we have exploited the same learning rate and number of latent factors
to train FPL and FCF, and we explored the models in the range of {10, . . . , 50} iterations.
We have set user- and positive item-regularization parameter to 210 of the learning rate. The
negative item-regularization parameter is 2010 of the learning rate, as suggested in mymedialite3
implementation as well as by Anelli et al. [22].</p>
      </sec>
      <sec id="sec-3-4">
        <title>3.4. Evaluation Metrics</title>
        <p>We have evaluated the performance of FPL under the accuracy and diversity perspective. The
accuracy of the models is measured by exploiting Precision ( @ ) and Recall (@ ). They
respectively represent, for each user, the proportion of relevant recommended items in the
recommendation list, and the fraction of relevant items that have been altogether suggested.
We have assessed the statistical significance of results by adopting Student’s paired T-test
considering p-values &lt; 0.054. The results are in general statistically significant but the
differences among BPR-MF, sFPL, and pFPL, which is a very important result. To measure the
diversity of recommendations, we have measured the Item Coverage (@ ), and the Gini
3http://www.mymedialite.net/
4The complete results are available at https://github.com/sisinflab/fpl-results/.</p>
        <p>Random 0.00013
Top-Pop 0.01909
BPR-MF 0.07702
VAE *
FCF
sFPL
sFPL+
pFPL
pFPL+</p>
        <p>PR
Best  obtained for each the proposed FPL variations for Brazil, Canada, and Italy are: sFPL = (0.5, 0.1, 0.4), pFPL = (0.8, 0.1, 1)
* VAE does not always produce recommendations for all the users. For Italy, the reported results cover the 14% of the users.
For this reason, it is not marked with bold in the table.</p>
        <p>0.1
0.5</p>
        <p>(a) Brazil
1
F
0.06
0.05
0.04
1
0.1
0.5</p>
        <p>(c) Italy
1
F
sFPL+, light blue is pFPL, light green is pFPL+.</p>
        <p>
          Index (@ ).   provides the number of diverse items recommended to users. It also
conveys the sense of the degree of personalization [23].  measures distributional inequality,
i.e., how unequally diferent items a RS provides users with [ 24]. A higher value of  [
          <xref ref-type="bibr" rid="ref18">18</xref>
          ]
corresponds to higher personalization.
        </p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Discussion</title>
      <p>The main goal of the experiments is assessing whether it is possible to obtain a recommendation
performance comparable to a centralized pair-wise learning approach while allowing the users
to control their data. In this respect, Table 2 shows the accuracy and diversity results of the
comparison between the state-of-the-art baselines and the four configurations of FPL presented
in Section 3. By focusing on accuracy metrics, we may notice that User-kNN outperforms
the other approaches in the three datasets, while the performance of Item-kNN and BPR-MF
approximately settle in the same range of values. This is possibly due to the user-item ratio [25],
that favors the user-based schemes (see Table 1). On the other hand, it is important to investigate
the diferences of FPL with respect to BPR-MF, which is a pair-wise centralized approach, being
FPL the first federated pair-wise recommender based on a factorization model. The performance
of BPR-MF against FPL, in the configuration sFPL, shows how precision and recall in sFPL
are slightly outperforming BPR-MF for the three datasets. This result is surprising since the
two methods share the sequential training, but sFPL exploits a  reduced to 0.5, 0.1, and 0.4,
respectively, for Brazil, Canada, and Italy. This behavior is more evident in Figure 2, where the
harmonic mean between Precision and Recall (F1) is plotted for diferent values of  . If we look
at the dark blue line, we may observe how the best result does not correspond to  = 1. In
the last three rows of Table 2, we explore an increasing of the local computation (sFPL+), or
an increased parallelism (pFPL), or a combination of both (pFPL+). In detail, we observe that
sFPL+ takes advantage of the increased local computation, and FPL significantly outperforms
BPR-MF for the three datasets; for instance, for Canada, we observe an interesting increase in
precision. Instead, when comparing pFPL with sFPL, we observe that the increased parallelism
does not afect the performance significantly. Even then, the increased local computation boosts
the Precision and Recall performance, up to 24% for precision in the Italy dataset. The results
confirm that the proposed system can generate recommendations with a quality that is
comparable with the centralized pair-wise learning approach. Moreover, the increased local computation
causes a considerable improvement in the accuracy of recommendation. On the other side, the
training parallelism does not significantly afects results. Finally, when the local computation
is combined with parallelism, the results show a further improvement.</p>
      <p>Afterwards, we varied  in the range {0.1, . . . , 1.0} to assess how removal of the updates
for consumed items afects the final recommendation accuracy, and we plotted the accuracy
performance by considering F1 in Figure 2. As previously observed, the best performance rarely
corresponds to  = 1. On the contrary, a general trend can be observed: the training reaches a
peak for a certain value of  — depending on the dataset —, and then the system performance
decays in accuracy when increasing the amount of shared positive updates. In rare cases, e.g.,
sFPL, and pFPL for Brazil dataset, the decay is absent, but results that are very close for diferent
values of  . The general behavior suggests that the system learning exploits the updates of
positive items to absorb information about popularity. This consideration is coherent with the
mathematical formulation of the learning procedure, and it is also supported by the observation
that for Canada and Italy FPL reaches the peak before with respect to Brazil. Indeed, Canada
and Italy datasets are less sparse than Brazil, and the increase of information about positive
items may lead to push up too much the popular items (this is a characteristic of pair-wise
learning), while the same behavior in Brazil can be observed for values of  very close to 1.
The same mathematical background, for sFPL+ and pFPL+ with Brazil dataset, which is very
sparse, explains the higher value of  needed to reach good performance. Here, the lack of
positive information with a vast catalog of items, confuses the training that cannot exploit item
popularity. Now, we can positively assert that user can receive high-quality recommendations
also when decides to disclose a small amount of her sensitive data. However, it should be noted
that the more the dataset is sparse, the more the amount of sensitive data should be large.</p>
    </sec>
    <sec id="sec-5">
      <title>5. Conclusion and Future Work</title>
      <p>Inspired by the potential ubiquity of the federated learning paradigm, we extend FPL, a novel
federated learning framework that exploits pair-wise learning for factorization models. To
this purpose, we have designed a model to leave the user-specific information of the original
factorization model in the clients’ devices. With FPL, a user may be completely in control
of her sensitive data and could share no positive feedback with the centralized server. The
framework can be envisioned as a general factorization model in which clients can tune the
amount of information shared among devices. We have conducted an exploratory, but extensive,
experimental evaluation to analyze the degree of accuracy, the diversity of the recommendation
results, the trade-of between accuracy, and amount of shared transactions. We have assessed
that the proposed model shows performance comparable with several state-of-the-art baselines
and the classic centralized factorization model with pair-wise learning. The evaluation shows
that clients may share a small portion of their data with the server and still receive
highperformance recommendations. We believe that the proposed privacy-oriented paradigm may
open the doors to a new class of ubiquitous recommendation engines.
[21] M. Ammad-ud-din, E. Ivannikova, S. A. Khan, W. Oyomno, Q. Fu, K. E. Tan, A. Flanagan,
Federated collaborative filtering for privacy-preserving personalized recommendation
system, CoRR abs/1901.09888 (2019). arXiv:1901.09888.
[22] V. W. Anelli, T. D. Noia, E. D. Sciascio, C. Pomo, A. Ragone, On the discriminative power
of hyper-parameters in cross-validation and how to choose them, in: Proc. of the 13th
ACM Conf. on Recommender Systems, ACM, 2019, pp. 447–451.
[23] G. Adomavicius, Y. Kwon, Improving aggregate recommendation diversity using
rankingbased techniques, IEEE Trans. Knowl. Data Eng. 24 (2012) 896–911.
[24] P. Castells, N. J. Hurley, S. Vargas, Novelty and diversity in recommender systems, in:
F. Ricci, L. Rokach, B. Shapira (Eds.), Recommender Systems Handbook, Springer, 2015, pp.
881–918.
[25] G. Adomavicius, J. Zhang, Impact of data characteristics on recommender systems
performance, ACM Trans. Management Inf. Syst. 3 (2012) 3:1–3:17. URL: https://doi.org/10.1145/
2151163.2151166. doi:10.1145/2151163.2151166.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>B.</given-names>
            <surname>McFee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Barrington</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G. R. G.</given-names>
            <surname>Lanckriet</surname>
          </string-name>
          ,
          <article-title>Learning content similarity for music recommendation</article-title>
          ,
          <source>IEEE Trans. Audio, Speech &amp; Language Processing</source>
          <volume>20</volume>
          (
          <year>2012</year>
          )
          <fpage>2207</fpage>
          -
          <lpage>2218</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>J.</given-names>
            <surname>Yuan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Shalaby</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Korayem</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Lin</surname>
          </string-name>
          ,
          <string-name>
            <surname>K.</surname>
          </string-name>
          <article-title>AlJadda</article-title>
          , J. Luo,
          <article-title>Solving cold-start problem in large-scale recommendation engines: A deep learning approach</article-title>
          ,
          <source>in: 2016 IEEE Int. Conf. on Big Data, BigData 2016</source>
          ,
          <string-name>
            <surname>Washington</surname>
            <given-names>DC</given-names>
          </string-name>
          , USA, December 5-
          <issue>8</issue>
          ,
          <year>2016</year>
          , IEEE Computer Society,
          <year>2016</year>
          , pp.
          <fpage>1901</fpage>
          -
          <lpage>1910</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>J.</given-names>
            <surname>Konecný</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>McMahan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Ramage</surname>
          </string-name>
          ,
          <article-title>Federated optimization: Distributed optimization beyond the datacenter</article-title>
          ,
          <source>CoRR abs/1511</source>
          .03575 (
          <year>2015</year>
          ). arXiv:
          <volume>1511</volume>
          .
          <fpage>03575</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>V. W.</given-names>
            <surname>Anelli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Deldjoo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T. Di</given-names>
            <surname>Noia</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Ferrara</surname>
          </string-name>
          ,
          <article-title>Towards efective device-aware federated learning</article-title>
          ,
          <source>in: International Conference of the Italian Association for Artificial Intelligence</source>
          , Springer,
          <year>2019</year>
          , pp.
          <fpage>477</fpage>
          -
          <lpage>491</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <surname>H. B. McMahan</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          <string-name>
            <surname>Moore</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Ramage</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Hampson</surname>
          </string-name>
          , et al.,
          <article-title>Communication-eficient learning of deep networks from decentralized data</article-title>
          ,
          <source>arXiv preprint arXiv:1602.05629</source>
          (
          <year>2016</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>Q.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <surname>Y. Tong,</surname>
          </string-name>
          <article-title>Federated machine learning: Concept and applications</article-title>
          ,
          <source>ACM TIST 10</source>
          (
          <year>2019</year>
          )
          <volume>12</volume>
          :
          <fpage>1</fpage>
          -
          <lpage>12</lpage>
          :
          <fpage>19</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>S.</given-names>
            <surname>Weiss</surname>
          </string-name>
          ,
          <article-title>The need for a paradigm shift in addressing privacy risks in social networking applications</article-title>
          ,
          <source>in: IFIP International Summer School on the Future of Identity in the Information Society</source>
          , Springer,
          <year>2007</year>
          , pp.
          <fpage>161</fpage>
          -
          <lpage>171</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>A. J. P.</given-names>
            <surname>Jeckmans</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Beye</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Erkin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. H.</given-names>
            <surname>Hartel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. L.</given-names>
            <surname>Lagendijk</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Tang</surname>
          </string-name>
          ,
          <article-title>Privacy in recommender systems</article-title>
          , in: N.
          <string-name>
            <surname>Ramzan</surname>
            , R. van Zwol,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Lee</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          <string-name>
            <surname>Clüver</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          Hua (Eds.), Social Media Retrieval,
          <source>Computer Communications and Networks</source>
          , Springer,
          <year>2013</year>
          , pp.
          <fpage>263</fpage>
          -
          <lpage>281</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>V. W.</given-names>
            <surname>Anelli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Deldjoo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T. Di</given-names>
            <surname>Noia</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Ferrara</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Narducci</surname>
          </string-name>
          ,
          <article-title>How to put users in control of their data in federated top-n recommendation with learning to rank</article-title>
          ,
          <source>in: Proceedings of the 36th Annual ACM Symposium on Applied Computing</source>
          , SAC '21,
          <string-name>
            <surname>Association</surname>
          </string-name>
          for Computing Machinery, New York, NY, USA,
          <year>2021</year>
          , p.
          <fpage>1359</fpage>
          -
          <lpage>1362</lpage>
          . URL: https://doi.org/10. 1145/3412841.3442010. doi:
          <volume>10</volume>
          .1145/3412841.3442010.
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>V. W.</given-names>
            <surname>Anelli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Deldjoo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T. D.</given-names>
            <surname>Noia</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Ferrara</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Narducci</surname>
          </string-name>
          ,
          <article-title>Federank: User controlled feedback with federated recommender systems</article-title>
          , in: D.
          <string-name>
            <surname>Hiemstra</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Moens</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Mothe</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Perego</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Potthast</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Sebastiani</surname>
          </string-name>
          (Eds.),
          <source>Advances in Information Retrieval - 43rd European Conference on IR Research</source>
          , ECIR
          <year>2021</year>
          ,
          <string-name>
            <surname>Virtual</surname>
            <given-names>Event</given-names>
          </string-name>
          ,
          <year>March</year>
          28 - April 1,
          <year>2021</year>
          , Proceedings,
          <string-name>
            <surname>Part</surname>
            <given-names>I</given-names>
          </string-name>
          , volume
          <volume>12656</volume>
          of Lecture Notes in Computer Science, Springer,
          <year>2021</year>
          , pp.
          <fpage>32</fpage>
          -
          <lpage>47</lpage>
          . URL: https://doi.org/10.1007/978-3-
          <fpage>030</fpage>
          -72113-
          <issue>8</issue>
          _3. doi:
          <volume>10</volume>
          .1007/978-3-
          <fpage>030</fpage>
          -72113-8\_3.
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>D.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Qu</surname>
          </string-name>
          ,
          <article-title>Participatory cultural mapping based on collective behavior data in location-based social networks</article-title>
          ,
          <source>ACM TIST 7</source>
          (
          <year>2016</year>
          )
          <volume>30</volume>
          :
          <fpage>1</fpage>
          -
          <lpage>30</lpage>
          :
          <fpage>23</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>J.</given-names>
            <surname>Konecný</surname>
          </string-name>
          , H. B.
          <string-name>
            <surname>McMahan</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Ramage</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Richtárik</surname>
          </string-name>
          ,
          <article-title>Federated optimization: Distributed machine learning for on-device intelligence</article-title>
          ,
          <source>CoRR abs/1610</source>
          .02527 (
          <year>2016</year>
          ). arXiv:
          <volume>1610</volume>
          .
          <fpage>02527</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>K.</given-names>
            <surname>Bonawitz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Eichner</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Grieskamp</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Huba</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Ingerman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Ivanov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Kiddon</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Konecný</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Mazzocchi</surname>
          </string-name>
          , H. B.
          <string-name>
            <surname>McMahan</surname>
            ,
            <given-names>T. V.</given-names>
          </string-name>
          <string-name>
            <surname>Overveldt</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Petrou</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Ramage</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Roselander</surname>
          </string-name>
          ,
          <article-title>Towards federated learning at scale: System design</article-title>
          , CoRR abs/
          <year>1902</year>
          .01046 (
          <year>2019</year>
          ). arXiv:
          <year>1902</year>
          .01046.
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>P.</given-names>
            <surname>Kairouz</surname>
          </string-name>
          , H. B.
          <string-name>
            <surname>McMahan</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Avent</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Bellet</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Bennis</surname>
            ,
            <given-names>A. N.</given-names>
          </string-name>
          <string-name>
            <surname>Bhagoji</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          <string-name>
            <surname>Bonawitz</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          <string-name>
            <surname>Charles</surname>
            , G. Cormode,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Cummings</surname>
          </string-name>
          , et al.,
          <article-title>Advances and open problems in federated learning</article-title>
          , arXiv preprint arXiv:
          <year>1912</year>
          .
          <volume>04977</volume>
          (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Koren</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. M.</given-names>
            <surname>Bell</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Volinsky</surname>
          </string-name>
          ,
          <article-title>Matrix factorization techniques for recommender systems</article-title>
          ,
          <source>IEEE Computer 42</source>
          (
          <year>2009</year>
          )
          <fpage>30</fpage>
          -
          <lpage>37</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <surname>D. kumar Bokde</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Girase</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Mukhopadhyay</surname>
          </string-name>
          ,
          <article-title>Role of matrix factorization model in collaborative filtering algorithm: A survey</article-title>
          ,
          <source>CoRR abs/1503</source>
          .07475 (
          <year>2015</year>
          ). arXiv:
          <volume>1503</volume>
          .
          <fpage>07475</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>S.</given-names>
            <surname>Rendle</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Freudenthaler</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Gantner</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Schmidt-Thieme</surname>
          </string-name>
          ,
          <article-title>BPR: bayesian personalized ranking from implicit feedback</article-title>
          , in: J. A.
          <string-name>
            <surname>Bilmes</surname>
          </string-name>
          , A. Y. Ng (Eds.),
          <source>UAI 2009, Proc. of the Twenty-Fifth Conf. on Uncertainty in Artificial Intelligence</source>
          , Montreal, QC, Canada, June 18-21,
          <year>2009</year>
          , AUAI Press,
          <year>2009</year>
          , pp.
          <fpage>452</fpage>
          -
          <lpage>461</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>A.</given-names>
            <surname>Gunawardana</surname>
          </string-name>
          , G. Shani,
          <article-title>Evaluating recommender systems</article-title>
          , in: F.
          <string-name>
            <surname>Ricci</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <string-name>
            <surname>Rokach</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          Shapira (Eds.),
          <source>Recommender Systems Handbook</source>
          , Springer,
          <year>2015</year>
          , pp.
          <fpage>265</fpage>
          -
          <lpage>308</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>V. W.</given-names>
            <surname>Anelli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T. D.</given-names>
            <surname>Noia</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E. D.</given-names>
            <surname>Sciascio</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Ragone</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Trotta</surname>
          </string-name>
          ,
          <article-title>Local popularity and time in top-n recommendation</article-title>
          ,
          <source>in: European Conf. on Information Retrieval</source>
          , volume
          <volume>11437</volume>
          , Springer,
          <year>2019</year>
          , pp.
          <fpage>861</fpage>
          -
          <lpage>868</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>D.</given-names>
            <surname>Liang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. G.</given-names>
            <surname>Krishnan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. D.</given-names>
            <surname>Hofman</surname>
          </string-name>
          , T. Jebara,
          <article-title>Variational autoencoders for collaborative filtering</article-title>
          ,
          <source>in: Proceedings of the 2018 World Wide Web Conference</source>
          ,
          <year>2018</year>
          , pp.
          <fpage>689</fpage>
          -
          <lpage>698</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>