<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Exploiting Social Tags in Matrix Factorization Models for Cross-domain Collaborative Filtering</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Ignacio Fernández-Tobías</string-name>
          <email>ignacio.fernandezt@uam.es</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Iván Cantador Escuela Politécnica Superior</string-name>
          <email>ivan.cantador@uam.es</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Universidad Autónoma de Madrid 28049</institution>
          <addr-line>Madrid</addr-line>
          ,
          <country country="ES">Spain</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2014</year>
      </pub-date>
      <fpage>33</fpage>
      <lpage>40</lpage>
      <abstract>
        <p>Cross-domain recommender systems aim to generate or enhance personalized recommendations in a target domain by exploiting knowledge (mainly user preferences) from other source domains. Due to the heterogeneity of item characteristics across domains, content-based recommendation methods are difficult to apply, and collaborative filtering has become the most popular approach to cross-domain recommendation. Nonetheless, recent work has shown that the accuracy of cross-domain collaborative filtering based on matrix factorization can be improved by means of content information; in particular, social tags shared between domains. In this paper, we review state of the art approaches in this direction, and present an alternative recommendation model based on a novel extension of the SVD++ algorithm. Our approach introduces a new set of latent variables, and enriches both user and item profiles with independent sets of tag factors, better capturing the effects of tags on ratings. Evaluating the proposed model in the movies and books domains, we show that it can generate more accurate recommendations than existing approaches, even in cold-start situations.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Recommender systems</kwd>
        <kwd>collaborative filtering</kwd>
        <kwd>cross-domain recommendation</kwd>
        <kwd>social tagging</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>
        1. INTRODUCTION
Recommender systems [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] have been successfully used in
numerous domains and applications to identify potentially
relevant items for users according to their preferences (tastes,
interests and goals). Examples include suggested movies and TV
programs in Netflix1, music albums in Last.fm2, and books in
Barnes&amp;Noble3.
1 Netflix online movies &amp; TV shows provider, http://www.netflix.com
2 Last.fm music discovery service, http://www.lastfm.com
3 Barnes&amp;Noble retail bookseller, http://www.barnesandnoble.com
Even though the majority of recommender systems focus on a
single domain or type of item, there are cases in which providing
the user with cross-domain recommendations could be beneficial.
For instance, large e-commerce sites like Amazon4 and eBay5
collect user feedback for items from multiple domains, and in
social networks users often share their tastes and interests on a
variety of topics. In these cases, rather than exploiting user
preference data from each domain independently, recommender
systems could exploit more exhaustive, multi-domain user models
that allow generating item recommendations spanning several
domains. Furthermore, exploiting additional knowledge from
related, auxiliary domains could help improve the quality of item
recommendations in a target domain, e.g. addressing the cold-start
and sparsity problems [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ].
      </p>
      <p>
        These benefits rely on the assumption that there are similarities or
relations between user preferences and/or item attributes from
different domains. When such correspondences exist, one way to
exploit them is by aggregating knowledge from the involved
domain data sources, for example by merging user preferences
into a unified model [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], and by combining single-domain
recommendations [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. An alternative way consists of transferring
knowledge from a source domain to a target domain, for example
by sharing implicit latent features that relate source and target
domains [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ][
        <xref ref-type="bibr" rid="ref17">17</xref>
        ], and by exploiting implicit rating patterns from
source domains in the target domain [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ][
        <xref ref-type="bibr" rid="ref14">14</xref>
        ].
      </p>
      <p>In either of the above cases, most of the existing approaches to
cross-domain recommendation are based on collaborative
filtering, since it merely needs rating data, and does not require
information about the users’ and items’ characteristics, which are
usually highly heterogeneous among domains.</p>
      <p>
        However, inter-domain links established through content-based
features and relations may have several advantages, such as a
better interpretability of the cross-domain user models and
recommendations, and the establishment of more reliable methods
to support the knowledge transfer between domains. In particular,
social tags assigned to different types of items –such as movies,
music albums, and books–, may act as a common vocabulary
between domains [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ][
        <xref ref-type="bibr" rid="ref17">17</xref>
        ]. Hence, as domain independent
contentbased features, tags can be used to overcome the information
heterogeneity across domains, and are suitable for building the
above mentioned inter-domain links.
      </p>
      <p>
        In this paper, we review state of the art cross-domain
recommendation approaches that utilize social tags to exploit
knowledge from an auxiliary source domain for enhancing
collaborative filtering rating predictions in a target domain.
4 Amazon e-commerce website, http://www.amazon.com
5 eBay consumer-to-consumer website, http://www.ebay.com
Specifically, we focus on several extensions of the matrix
factorization technique proposed in [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ], which incorporates latent
factors related to the users’ social tags. By jointly learning tag
factors in both the source and target domains, hidden correlations
between ratings and tags in the source domain can be used in the
target domain. Hence, for instance, a movie recommender system
may estimate a higher rating for a particular movie tagged as
interesting or amazing if these tags are usually assigned to books
positively rated. Also, books tagged as romantic or suspenseful
may be recommended to a user if it is found that such tags
correlate with high movie ratings.
      </p>
      <p>
        Enrich et al. [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] presented several recommendation models that
exploit different sets of social tags when computing rating
predictions, namely tags assigned by the active user to the item for
which the rating is estimated, and all the tags assigned by the
community to the target item. Despite their good performance,
these models do have difficulties in cold-start situations where no
tagging information is available for the target user/item.
In this paper, we propose a method that expands the users’ and
items’ profiles to overcome these limitations. More specifically,
we propose to incorporate additional parameters to the above
models, separating user and item latent tag factors in order to
capture the contributions of each to the ratings more accurately.
Furthermore, by modeling user and item tags independently we
are able to compute rating predictions even when a user has not
assigned any tag to an item, or for items that have not been tagged
yet. For such purpose, we adapt the gSVD++ algorithm [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] –
designed to integrate content metadata into the matrix
factorization process– for modeling social tags in the
crossdomain recommendation scenario.
      </p>
      <p>Through a series of experiments in the movies and books
domains, we show that the proposed approach outperforms the
state of the art methods, and validate the main contribution of this
work: A model that separately captures user and item tagging
information, and effectively transfers auxiliary knowledge to the
target domain in order to provide cross-domain recommendations.
The reminder of the paper is structured as follows. In section 2 we
review state of the art approaches to the cross-domain
recommendation problem, focusing on algorithms based on matrix
factorization, and on algorithms that make use of social tags to
relate the domains of interest. In section 3 we provide a brief
overview of matrix factorization methods for single-domain
recommendation, and in section 4 we describe their extensions for
the cross-domain recommendation case. In section 5 we present
and discuss the conducted experimental work and obtained
results. Finally, in section 6 we summarize some conclusions and
future research lines.</p>
    </sec>
    <sec id="sec-2">
      <title>2. RELATED WORK</title>
      <p>
        Cross-domain recommender systems aim to generate or enhance
personalized recommendations in a target domain by exploiting
knowledge (mainly user preferences) from other source domains
[
        <xref ref-type="bibr" rid="ref7">7</xref>
        ][
        <xref ref-type="bibr" rid="ref19">19</xref>
        ]. This problem has been addressed from various perspectives
in several research areas. It has been faced by means of user
preference aggregation and mediation strategies for the
crosssystem personalization problem in user modeling [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ][
        <xref ref-type="bibr" rid="ref3">3</xref>
        ][
        <xref ref-type="bibr" rid="ref16">16</xref>
        ], as a
potential solution to mitigate the cold-start and sparsity problems in
recommender systems [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ][
        <xref ref-type="bibr" rid="ref17">17</xref>
        ][
        <xref ref-type="bibr" rid="ref18">18</xref>
        ], and as a practical application of
knowledge transfer techniques in machine learning [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ][
        <xref ref-type="bibr" rid="ref14">14</xref>
        ][
        <xref ref-type="bibr" rid="ref15">15</xref>
        ].
We can distinguish between two main types of cross-domain
approaches: Those that aggregate knowledge from various source
domains to perform recommendations in a target domain, and
those that link or transfer knowledge between domains to support
recommendations in the target domain.
      </p>
      <p>
        The knowledge aggregation methods merge user preferences (e.g.
ratings, social tags, and semantic concepts) [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], mediate user
modeling data exploited by various recommender systems (e.g. user
similarities and user neighborhoods) [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ][
        <xref ref-type="bibr" rid="ref16">16</xref>
        ], and combine
singledomain recommendations (e.g. rating estimations and rating
probability distributions) [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. The knowledge linkage and transfer
methods relate domains by common information (e.g. item
attributes, association rules, semantic networks, and inter-domain
correlations) [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ][
        <xref ref-type="bibr" rid="ref18">18</xref>
        ], share implicit latent features that relate source
and target domains [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ][
        <xref ref-type="bibr" rid="ref17">17</xref>
        ], and exploit explicit or implicit rating
patterns from source domains in the target domain [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ][
        <xref ref-type="bibr" rid="ref14">14</xref>
        ].
Cross-domain recommendation models based on latent factors are a
popular choice among knowledge linkage and transfer methods,
since they allow automatically discovering and exploiting implicit
domain relations within the data from different domains. For
instance, Zhang et al. [
        <xref ref-type="bibr" rid="ref20">20</xref>
        ] proposed an adaptation of the matrix
factorization model to include a probability distribution that
captures inter-domain correlations, and Cao et al. [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] presented a
method that learns similarities between item latent factors in
different domains as parameters in a Bayesian framework. Aiming
to exploit heterogeneous forms of user feedback, Pan et al. [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ]
proposed an adaptive model in which the latent features learned in
the source domain are transferred to the target domain in order to
regularize the matrix factorization there. Instead of the more
common two-way decomposition of the rating matrix, Li et al. [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ]
used a nonnegative matrix tri-factorization to extract rating patterns
–the so-called codebook– in the source domain. Then, rather than
transferring user and item latent factors, the rating patterns are
shared in the target domain and used to predict the missing ratings.
Despite the ability of matrix factorization models to discover
latent implicit relations, there are some methods that use tags as
explicit information to bridge the domains. Shi et al. [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ] argued
that explicit relations established through common social tags are
more effective for such purpose, and used them to compute
useruser and item-item cross-domain similarities. In this case, rating
matrices from the source and target domains are jointly factorized,
but user and item latent factors are restricted so that they are
consistent with the tag-based similarities.
      </p>
      <p>
        Instead of focusing on sharing user or item latent factors, Enrich et
al. [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] studied the influence of social tags on rating prediction.
More specifically, the authors presented a number of models based
on the well-known SVD++ algorithm [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ], to incorporate the effect
of tag assignments into rating estimations. The underlying
hypothesis is that information about how users annotate items in
the source domain can be exploited to improve rating prediction in
a different target domain, as long as a set of common tags between
the domains exists. In all the proposed models, tag factors are
added into the latent item vectors, and are then combined with user
latent features to compute rating estimations. The difference
between these models is the set of tags considered for rating
prediction. Two of the proposed models use the tags assigned by
the user to a target item, and the other model takes the tags of the
whole community into account. We note that the first two models
require the active user to tag, but not rate the item in the target
domain. In all the models, the transfer of knowledge is performed
through the shared tag factors in a collective way, since these
factors are learned jointly for the source and the target domains.
The results reported in the movies and books domains confirmed
that shared knowledge can be effectively exploited to outperform
single-domain rating predictions.
      </p>
      <p>
        The model we propose in this paper follows the same line as
Enrich et al. [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ], in the sense that tags are directly integrated as
latent factors into the rating prediction process, as opposed to
Shi’s and colleagues’ approach [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ], which estimates the ratings
using only user and item factors. The main difference of our
model with the approaches presented in [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] is the way in which
the rating matrix is factorized. Rather than using a single set of tag
factors to extend the item’s factorization component, we introduce
additional latent variables in the user component to separately
capture the effect of tags utilized by the user and the tags assigned
to the item. For this purpose, we adapt the gSVD++ algorithm
[
        <xref ref-type="bibr" rid="ref10">10</xref>
        ], which extends SVD++ by introducing a set of latent factors
to take item metadata into account for rating prediction. In this
model, both user and item factors are respectively enhanced with
implicit feedback and content information, which allows
improving the accuracy of rating predictions.
      </p>
    </sec>
    <sec id="sec-3">
      <title>3. OVERVIEW OF MATRIX</title>
    </sec>
    <sec id="sec-4">
      <title>FACTORIZATION METHODS</title>
      <p>Since the proposed cross-domain recommendation model is built
upon a matrix factorization collaborative filtering method, in this
section we provide a brief overview of the well-known standard
rating matrix factorization technique, and the SVD++ and
gSVD++ algorithms, which extend the former by incorporating
implicit user feedback and item metadata, respectively.</p>
    </sec>
    <sec id="sec-5">
      <title>3.1 MF: Standard rating matrix factorization</title>
      <p>
        Matrix factorization (MF) methods [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ][
        <xref ref-type="bibr" rid="ref12">12</xref>
        ] are a popular approach
to latent factor models in collaborative filtering. In these methods,
the rating matrix is decomposed as the product of low-rank
matrices of user and item latent features. In its most basic form, a
factor vector  ∈ ℝ is assigned to each user , and a factor
vector  ∈ ℝ to each item , so that ratings are estimated as:
where the term  is a baseline estimate that captures the
deviation of user and item ratings from the average, and is defined
as:
̂ =  + 
 =  +  + 
The parameter  corresponds to the global average rating in the
training set, and  and  are respectively the deviations in the
ratings of user  and item  from the average. The baseline
estimates can be explicitly defined or learned from the data. In the
latter case, the parameters of the model are found by solving the
following regularized least squares problem:
min
∗,∗,∗
      </p>
      <p>
          −  −  −  − 
,∈ℛ
+  +  + ‖‖ + ‖‖

In this formula, the parameter  controls the amount of
regularization to prevent high model variance and overfitting. The
minimization can be performed by using gradient descent over the
set ℛ of observed ratings [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]. This method is popularly called
SVD, but it is worth noticing that it is not completely equivalent
to the singular value decomposition technique, since the rating
matrix is usually very sparse and most of its entries are actually
not observed.
      </p>
      <p>For simplicity purposes, in the following we omit the baseline
estimates. They, nonetheless, can be easily considered by adding
the  term into the rating estimation formulas.</p>
    </sec>
    <sec id="sec-6">
      <title>3.2 SVD++: Adding implicit user feedback to the rating matrix factorization method</title>
      <p>
        The main motivation behind the SVD++ algorithm, proposed by
Koren [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ][
        <xref ref-type="bibr" rid="ref13">13</xref>
        ], is to exploit implicit additional user feedback for
rating prediction, since it is arguably to use a more available and
abundant source of user preferences.
      </p>
      <p>In this model, user preferences are represented as a combination
of explicit and implicit feedback, searching for a better
understanding of the user by looking at what items she rated,
purchased or watched. For this purpose, additional latent factors
are combined with the user’s factors as follows:

̂ =   + ||

∈

In the previous formula,  ∈ ℝ,  ∈ ℝ,  ∈ ℝ represent
user, item, and implicit feedback factors, respectively.  is the
set of items for which the user  provided implicit preference,
and  is the number of latent features.</p>
      <p>Similarly to the SVD algorithm, the parameters of the model can
be estimated by minimizing the regularized squared error loss
over the observed training data:
min
∗,∗,∗</p>
      <p>
,∈ℛ</p>
      <p>
 −   + ||
+  ‖‖ + ‖‖ +

  
∈

∈


(4)
(5)
(1)
(2)
(3)</p>
      <p>Again, the minimization problem can be efficiently solved using
stochastic gradient descent.</p>
    </sec>
    <sec id="sec-7">
      <title>3.3 gSVD++: Adding item metadata to the rating matrix factorization method</title>
      <p>
        The gSVD++ algorithm [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] further extends SVD++ considering
information about the items’ attributes in addition to the users’
implicit feedback.
      </p>
      <p>The model introduces a new set of latent variables  ∈ ℝ for
metadata that complement the item factors. This idea combined
with the SVD++ algorithm leads to the following formula for
computing rating predictions:</p>
      <p>
        
̂ =  + ||

   + ||
∈

∈
 (6)
The set  contains the attributes related to item , e.g. comedy
and romance in the case of movie genres. The parameter  is set
to 1 when the set  ≠ ∅, and 0 otherwise. We note that in the
previous formula, both user and item factors are enriched with
new uncoupled latent variables that separately capture information
about the users and items, leading to a symmetric model with four
types of parameters. Again, parameter learning can be performed
by minimizing the associated squared error function with gradient
descent:


∗,m∗,in∗,∗ ,∈ℛ  −  + || ∈   + || ∈


(7)
+  ‖‖ + ‖‖ +   +  
∈ ∈
The use of additional latent factors for item metadata is reported
to improve prediction accuracy over SVD++ in [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ]. In this paper,
we adapt this model to separately learn user and item tag factors,
aiming to support the transfer of knowledge between domains.
      </p>
    </sec>
    <sec id="sec-8">
      <title>4. TAG-BASED MODELS FOR CROSS</title>
    </sec>
    <sec id="sec-9">
      <title>DOMAIN COLLABORATIVE FILTERING</title>
      <p>
        In this section, we first describe the tag-based cross-domain
collaborative filtering models presented in [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ], which are an
adaptation of the SVD++ algorithm, and next introduce our
proposed model, which is built upon the gSVD++ algorithm.
      </p>
    </sec>
    <sec id="sec-10">
      <title>4.1 Adaptation of SVD++ for Tag-based</title>
    </sec>
    <sec id="sec-11">
      <title>Cross-domain Collaborative Filtering</title>
      <p>
        The main hypothesis behind the models proposed in [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] is that the
effect of social tags on ratings can be shared between domains to
improve the rating predictions in the target domain. In that work,
three different adaptations of the SVD++ algorithm were explored
that utilize tags as implicit user feedback to enhance the item
factors, as opposed to user factors like in the original model.
The first of the algorithms proposed by Enrich et al. is the
UserItemTags model, which only exploits the tags  that the
active user  assigned to the target item :
̂ =   +
      </p>
      <p>1
||

∈

We note here that if the user has not tagged the item, i.e.,  =
∅, then the model corresponds to the standard matrix factorization
technique. Also, even though the tag factors  are only combined
with the item factors , the user and item factorization
components are not completely uncoupled, since the set  still
depends on the user .</p>
      <p>
        An improvement over the model was also presented in [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ], based
on the observation that not all the tags are equally relevant (i.e.
discriminative) to predict the ratings. The proposed alternative is
to filter the tags in the set  that are not relevant according to
certain criterion. In that work, the Wilcoxon rank-sum test is
performed for each tag to decide if the mean rating significantly
changes in the presence/absence of the tag in the dataset. In this
model, rating predictions are computed in an analogous manner:
̂ =   +
      </p>
      <p>1
||</p>
      <p>
∈

Here, the set  ⊆  only contains those tags for which
the p-value of the abovementioned test is  &lt; 0.05. This method
was called as UserItemRelTags.</p>
      <p>As noted by the authors, the previous methods are useful when the
user has tagged but not rated an item. However, these methods do
not greatly improve over the standard matrix factorization
technique in the cold-start situations where new users or items are
considered. Aiming to address this limitation, a last approach was
proposed, the ItemRelTags model:
̂ =   +</p>
      <p>1
||</p>
      <p>
∈

(10)
Now, the set  contains all the relevant tags assigned by the
whole community to the item , with possible repetitions. Tags
that appear more often contribute with more factors to the
(8)
(9)
prediction, and  is the number of times tag  was applied to
item . In this case, the normalization factor is || =
∑∈ .</p>
      <p>We note that the set  does not depend on the user, and that
the user and item components of the factorization are fully
uncoupled. This has the advantage that tag factors can also be
exploited in the rating predictions for new users for whom tagging
information is not available yet, improving over the standard
matrix factorization method. The ItemRelTags model, however,
does not take into account the possibility that the user has tagged
different items other than the one for which the rating is being
estimated. In such cases, it may be beneficial to enrich the user’s
profile by considering other tags the user has chosen in the past as
evidence of her preferences. In the next subsection, we propose a
model that aims to exploit this information to generate more
accurate recommendations.</p>
      <p>Similarly to the SVD++ algorithm, all of the above models can be
trained by minimizing the associated loss function with stochastic
gradient descent.</p>
    </sec>
    <sec id="sec-12">
      <title>4.2 Adaptation of gSVD++ for Tag-based</title>
    </sec>
    <sec id="sec-13">
      <title>Cross-domain Collaborative Filtering</title>
      <p>Although the previous recommendation models can successfully
transfer tagging information between domains, they suffer from
some limitations. The UserItemTags and UserItemRelTags models
cannot do better than the standard matrix factorization if the user
has not tagged the item for which the rating is being estimated,
while the ItemRelTags model does not fully exploits the user’s
preferences expressed in the tags assigned to other items.
In this paper, we propose to adapt the gSVD++ algorithm by
introducing an additional set of latent variables  ∈ ℝ that
enrich the user’s factors and better capture the effect of her tags in
the rating estimation. Specifically, we distinguish between two
different sets of tags for users and items, and factorize the rating
matrix into fully uncoupled user and item components as follows:
̂ =  +
1
</p>
      <p>1
   +
|| ∈</p>
      <p> 
|| ∈
(11)
The set  contains all the tags assigned by user  to any item.
Respectively,  is the set of tags assigned by any user to item ,
and plays the role of item metadata  in the gSVD++
algorithm. As in the ItemRelTags model, there may be repeated
tags in each of the above tag sets, which we account for by
considering the number of times a tag appears in  or  ,

respectively. In (11),  is the number of items on which the user
 applied tag , and  is the number of users that applied tag  to
item . As previously, tag factors are normalized by || =
∑∈  and || = ∑∈ , so that factors  and  do not
dominate over the rating factors  and  for users and items with
a large number of tags.</p>
      <p>In the proposed model, which we call as TagGSVD++, a user’s
profile is enhanced with the tags she used, since we hypothesize
that her interests are better captured, and that transferring this
information between domains can be beneficial for estimating
ratings in the target domain. Likewise, item profiles are extended
with the tags that were applied to them, as in the ItemRelTags
model.</p>
      <p>The parameters of TagGSVD++ can be learned from the observed
training data by solving the following unconstrained minimization
problem:
∗,m∗,in∗,∗ ,∈ℛ , , ∈, ∈</p>
      <p>1 1 1
= ∗,m∗,in∗,∗ ,∈ℛ 2  −  + || ∈   + || ∈  (12)
+ 2 ‖‖ + ‖‖ + ∈‖‖ + ∈‖‖
The factor 1⁄2 simplifies the following derivations with no effect
on the solution. As in the previous models, a minimum can be
found by stochastic gradient descent. For completeness, in the
following we list the update rules of TagGSVD++ taking the
derivatives of the error function in (12) with respect to the
parameters:
</p>
      <p>
 + 
∀ ∈ 





= −  +
= −  +
∈
where the error term  is  − ̂. In the training phase, we
loop over the observed ratings simultaneously updating the
parameters according to the following rules:

 + 
∀ ∈ 
 ←  −   −   + || ∑∈ 
The learning rate  determines to what extent the parameters are
updated in each iteration. A small learning rate can make the
learning slow, whereas with a large learning rate the algorithm
may fail to converge. The choice of both the learning rate and the
regularization parameter  is discussed later in section 5.3.</p>
    </sec>
    <sec id="sec-14">
      <title>5. EXPERIMENTS</title>
      <p>We have evaluated the proposed TagGSVD++ algorithm (section
4.2) in a cross-domain collaborative filtering setting, by
empirically comparing it with the single-domain matrix
factorization methods (section 3) and the state-of-the-art
crossdomain recommendation approaches described in section 4.1.</p>
    </sec>
    <sec id="sec-15">
      <title>5.1 Dataset</title>
      <p>
        We have attempted to reproduce the cross-domain dataset used in
[
        <xref ref-type="bibr" rid="ref6">6</xref>
        ], aiming to compare our approach with those presented in that
paper. For the sake of completeness, we also describe the data
collection process here.
      </p>
      <p>In order to simulate the cross-domain collaborative filtering
setting, we have downloaded two publicly available datasets for
the movies and books domains. The MovieLens 10M dataset6
(ML) contains over 10 million ratings and 100,000 tag
assignments by 71,567 users to 10,681 movies. The LibraryThing
dataset7 (LT) contains over 700,000 ratings and 2 million tag
6 MovieLens datasets, http://grouplens.org/datasets/movielens
7 LibraryThing dataset, http://www.macle.nl/tud/LT
assignments by 7,279 users on 37,232 books. Ratings in both of
the datasets are expressed on a 1-5 scale, with interval steps of
0.5.</p>
      <p>Since we were interested in analyzing the effect of tags on rating
prediction, we only kept ratings in MovieLens on movies for
which at least one tag was applied, leaving a total of 24,564
ratings. Also following the setup done by Enrich et al., we
considered the same amount of ratings in LibraryThing, and took
the first 24,564 ratings. We note, however, that the original
dataset contained duplicate rows and inconsistencies, i.e., some
user-item pairs had more than one rating. Hence, we preprocessed
the dataset removing such repetitions and keeping only the
repeated ratings that appeared first in the dataset’s file. We also
converted the tags to lower case in both datasets. Table 1 shows
the characteristics of the final datasets.</p>
    </sec>
    <sec id="sec-16">
      <title>5.2 Evaluation methodology</title>
      <p>As mentioned above, we have compared the performance of the
proposed model against the single-domain matrix factorization
baselines from section 3, and the state-of-the-art tag-based
algorithms described in section 4.1. All these methods are
summarized next:
MF The standard matrix factorization method trained by
stochastic gradient descent over the observed ratings of both
movies and books domains.</p>
      <p>SVD++ An adaptation of MF to take implicit data into account. In
our experiments, the set  contains all the items rated by user .
gSVD++ An extension of SVD++ to include item metadata into
the factorization process. In our experiments, we have considered
as set of item attributes  the tags  assigned to item  by any
user. Note that, as tags are content features for both movies and
books, this method is suitable for cross-domain recommendation,
since knowledge can be transferred through the metadata (tag)
factors. This differs from the proposed TagGSVD++ in that users
are modeled as in SVD++ by considering rated items as implicit
feedback instead of their tags. Also, normalization of the implicit
data factors on the user component involves a square root; see
equations (6) and (11).</p>
      <p>UserItemTags A method that expands an item ’s profile with
latent factors of tags that the target user assigned to . Its
parameters are learned by simultaneously factorizing the rating
matrices of both source and target domains.</p>
      <p>UserItemRelTags A variation of the previous method that only
takes relevant tags into account, as determined by a Wilcoxon
rank-sum test.</p>
      <p>ItemRelTags Instead of tags assigned by the user, this method
exploits all relevant tags applied by the whole user community,
and is thus able to compute rating predictions even if the user has
not tagged the target item.</p>
      <p>
        We evaluated all these recommendation methods in two settings,
using MovieLens as source domain and LibraryThing as target
domain, and vice-versa. In both cases, we evaluated the methods
through 10-fold cross-validation, i.e., we shuffled the target
ratings and split them into 10 non-overlapping folds. In each fold,
we left out one part, 10% of the ratings, as test set to estimate the
performance of the methods. The rest 90% of the ratings were
used as a training set to learn the models, and a validation set to
find the optimal values of the models’ parameters. Specifically,
we randomly chose 80% of these remaining ratings, and combined
them with the source domain ratings to build the models. The final
20% left was used for the validation set to select the best number
of factors , learning rate , and regularization . Figure 1 depicts
the split of the data into training, validation, and test sets.
As in [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ], we also wanted to investigate how the number of
available ratings in the target domain affects the quality of the
recommendations. For such purpose, we further split the training
data from the target domain into 10 portions to simulate different
rating sparsity levels. First, in order to evaluate the performance
of the methods in cold-start situations, we used only 10% of the
target training ratings, i.e., 0.1  0.8  0.9  24,564 = 1,768 ratings
(see Table 1). Then, we incrementally added additional 10% of
the ratings to analyze the behavior of the methods with an
increasingly larger amount of observed rating data. In each
sparsity level, the full set of source domain ratings was also used
to build the models.
      </p>
      <p>Since all the methods are designed for the rating prediction task,
we measured their performance as the accuracy of the estimated
ratings. Specifically, we computed the Mean Absolute Error
(MAE) of each model in the different settings described above:
 =
1</p>
      <p>
|ℛ| ,∈ℛ
| − ̂|
where ℛ contains the ratings in the test set we left out for
evaluation.</p>
    </sec>
    <sec id="sec-17">
      <title>5.3 Results</title>
      <p>As previously mentioned, we reserved 20% of the target domain
training data in each fold for validating the models and finding the
best model parameters, in order not to overestimate the
performance of the methods.</p>
      <p>For hyperparameter optimization, with each method and sparsity
level in the target domain, we performed a grid (stepsize) search
on the validation set for the values of the learning rate , the
amount of regularization , and the number of latent features .
To get an idea of the typical values found for the parameters,
Table 2 shows the average best values for each method.
From the table, we observe that there is not a large difference in
the optimal number of factors and learning rates between
configurations. In contrast, we note that the amount of
regularization needed for the proposed TagGSVD++ method is
relatively large, e.g. compare  = 0.036 of TagGSVD++ with
 = 0.009 of MF. This may be due to the additional set of latent
variables for tags that our model uses; more complex models are
able to account for greater variance in the data and tend to overfit
more easily, thus requiring more regularization. In order to
analyze how the available information in the target domain affects
the stability of the model, Figure 2 shows the optimal value for
the regularization parameter for different sparsity levels.
We note that the gSVD++, upon which our model is defined, also
introduces additional latent variables and yet requires a lower
regularization. We argue that the differences between gSVD++
and TagGSVD++ regularizations are caused by the 
and sets, see equations (6) with  =  and (11). In Table 1
we see that, on average, the number of tags applied by a user is
much larger than the number of rated items. This results in more
variables actually taking part in the rating predictions, and hence
in a more complex model that requires more regularization to
prevent overfitting.</p>
      <p>Once we found the best parameters for each method and sparsity
level, we ran the models separately in the test set of each fold. The
final performance was estimated as the average MAE over the 10
folds. Figure 3a shows the results obtained using LibraryThing as
source domain and MovieLens as target domain. All the
differences with respect to the TagGSVD++ algorithm are
statistically significant as determined with a Wilcoxon signed rank
test at the 95% confidence level. It can be seen that the proposed
TagGSVD++ method is able to consistently outperform the rest of
the methods for all sparsity levels in the target domain, also in the
cold-start setting when only 10%-20% of the ratings are available.
We also note that cross-domain methods always achieve better
accuracy than single-domain MF, although SVD++ effectively
exploits implicit feedback and remains competitive until the 50%
sparsity level. Then, as the sparsity decreases, cross-domain
models provide greater improvements. This indicates that even if
plenty of target domain rating data is available, it is still beneficial
to transfer knowledge from the source domain.</p>
      <p>
        The results using MovieLens as source domain and LibraryThing
as target domain are shown in Figure 3b. As before, the difference
in MAE between TagGSVD++ and the rest of the methods is
statistically significant, according to the Wilcoxon signed rank
test with 95% confidence level. Again, TagGSVD++ is the best
performing method for all rating sparsity levels, followed by the
cross-domain methods. We now observe that the values of MAE
are in general larger than in the previous case, which seems to
indicate that the transfer of knowledge is not as effective in this
setting. This observation is in accordance with the results reported
in [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ], where the authors argue that this may be caused by
differences in the ratio of overlapping tags between the domains.
      </p>
      <p>Only 13.81% of the tags in MovieLens are shared in LibraryThing
(see Table 1), and thus less latent tag factors learned in the source
domain can be used in the target to compute rating predictions.
Figure 4 shows the average rating prediction error for users with
different amounts of observed ratings and tag assignments, using
LibraryThing as source domain and MovieLens as target domain.
We see that our model achieves the best improvements in
coldstart situations, where few ratings and tag assignments are
available in the target domain. We also note that the performance
degrades for users with more than 20 ratings (respectively, 100
tag assignments), when enough target domain data is available.
Nonetheless, in these cases, TagGSVD++ is still able to exploit
the learned tag factors to compute more accurate predictions.</p>
    </sec>
    <sec id="sec-18">
      <title>6. CONCLUSIONS AND FUTURE WORK</title>
      <p>Cross-domain recommendation has potential benefits over
traditional recommender systems that focus on single domains,
such as alleviating rating sparsity in a target domain by exploiting
data from a related source domain, improving the quality of
recommendations in cold-start situations by inferring new user
preferences from other domains, and by personalizing
crossselling strategies to provide customers with suggestions of items
of different types.</p>
      <p>
        Despite these advantages, cross-domain recommendation is a
fairly new topic with plenty of research opportunities to explore.
One of the major difficulties that arises is how to link or relate the
different domains to support the transfer of knowledge. Due to the
common heterogeneity of item attributes across domains,
collaborative filtering techniques have become more popular than
content-based methods. However, recent work [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ][
        <xref ref-type="bibr" rid="ref17">17</xref>
        ] has
concluded that more reliable and meaningful relations can be
established between the domains by exploiting certain content
information, such as social tags.
      </p>
      <p>In this paper, we have adapted a novel extension of the
wellknown SVD++ algorithm to separately model the effect of user
and item tags in the observed ratings. By introducing a new set of
latent variables that represent tags in the user profile, our
TagGSVD++ method is able to transfer knowledge from a source
domain more effectively, providing accurate rating predictions in
the target domain, even in cold-start situations. From our
experiments in the movies and books domains, we conclude that
exploiting additional tag factors, and decoupling user and item
components in the factorization process improves the transfer of
knowledge and the accuracy of the recommendations.
In the future, we plan to further investigate the effect of tags in the
quality of recommendations. In particular, we want to study how
the recommendation performance depends on the number of
shared tags between domains. Increasing the overlap by grouping
tags with similar semantics but expressed differently in the
domains could favor the transfer of knowledge.</p>
      <p>
        In our experiments we altered the amount of observed rating data
in the target domain, but it would also be interesting to evaluate
the methods varying the number of available ratings in the source
domain. Moreover, we will perform a more exhaustive evaluation
with other datasets including more cross-domain recommendation
methods from the state of the art, such as [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ].
      </p>
    </sec>
    <sec id="sec-19">
      <title>7. ACKNOWLEDGEMENTS</title>
      <p>This work was supported by the Spanish Government
(TIN201347090-C3-2).</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <surname>Abel</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Helder</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Houben</surname>
          </string-name>
          , G.-J.,
          <string-name>
            <surname>Henze</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Krause</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <year>2013</year>
          .
          <article-title>Cross-system User Modeling and Personalization on the Social Web</article-title>
          .
          <source>User Modeling and User-Adapted Interaction</source>
          <volume>23</volume>
          (
          <issue>2-3</issue>
          ), pp.
          <fpage>169</fpage>
          -
          <lpage>209</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <surname>Adomavicius</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tuzhilin</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <year>2005</year>
          .
          <article-title>Toward the Next Generation of Recommender Systems: A Survey of the Stateof-the-art and Possible Extensions</article-title>
          .
          <source>IEEE Transactions on Knowledge and Data Engineering</source>
          <volume>17</volume>
          , pp.
          <fpage>734</fpage>
          -
          <lpage>749</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <surname>Berkovsky</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kuflik</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ricci</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <year>2008</year>
          .
          <article-title>Mediation of User Models for Enhanced Personalization in Recommender Systems</article-title>
          .
          <source>User Modeling and User-Adapted Interaction</source>
          <volume>18</volume>
          (
          <issue>3</issue>
          ), pp.
          <fpage>245</fpage>
          -
          <lpage>286</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <surname>Cao</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Liu</surname>
            ,
            <given-names>N. N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yang</surname>
            ,
            <given-names>Q.</given-names>
          </string-name>
          <year>2010</year>
          .
          <article-title>Transfer Learning for Collective Link Prediction in Multiple Heterogeneous Domains</article-title>
          .
          <source>In Proceedings of the 27th International Conference on Machine Learning</source>
          , pp.
          <fpage>159</fpage>
          -
          <lpage>166</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <surname>Cremonesi</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tripodi</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Turrin</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <year>2011</year>
          .
          <article-title>Cross-domain Recommender Systems</article-title>
          .
          <source>In Proceedings of the 11th IEEE International Conference on Data Mining Workshops</source>
          , pp.
          <fpage>496</fpage>
          -
          <lpage>503</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <surname>Enrich</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Braunhofer</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ricci</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <year>2013</year>
          .
          <article-title>Cold-Start Management with Cross-Domain Collaborative Filtering and Tags</article-title>
          .
          <source>In Proceedings of the 14th International Conference on E-Commerce and Web Technologies</source>
          , pp.
          <fpage>101</fpage>
          -
          <lpage>112</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <surname>Fernández-Tobías</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cantador</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kaminskas</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ricci</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <year>2012</year>
          .
          <article-title>Cross-domain Recommender Systems: A Survey of the State of the Art</article-title>
          .
          <source>In Proceedings of the 2nd Spanish Conference on Information Retrieval</source>
          , pp.
          <fpage>187</fpage>
          -
          <lpage>198</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <surname>Funk</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <year>2006</year>
          .
          <article-title>Netflix Update: Try This At Home</article-title>
          . http://sifter.org/~simon/journal/20061211.html
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <surname>Gao</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Luo</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chen</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Li</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gallinari</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Guo</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <year>2013</year>
          .
          <article-title>Cross-Domain Recommendation via Cluster-Level Latent Factor Model</article-title>
          .
          <source>In Proceedings of the 17th and 24th European Conference on Machine Learning and Knowledge Discovery in Databases</source>
          , pp.
          <fpage>161</fpage>
          -
          <lpage>176</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>Garcia</given-names>
            <surname>Manzato</surname>
          </string-name>
          ,
          <string-name>
            <surname>M.</surname>
          </string-name>
          <year>2013</year>
          . gSVD++
          <article-title>: Supporting Implicit Feedback on Recommender Systems with Metadata Awareness</article-title>
          .
          <source>In Proceedings of the 28th Annual ACM Symposium on Applied Computing</source>
          , pp.
          <fpage>908</fpage>
          -
          <lpage>913</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <surname>Koren</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          <year>2008</year>
          .
          <article-title>Factorization Meets the Neighborhood: A Multifaceted Collaborative Filtering Model</article-title>
          .
          <source>In Proceedings of the 14th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining</source>
          , pp.
          <fpage>426</fpage>
          -
          <lpage>434</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <surname>Koren</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bell</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Volinsky</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <year>2009</year>
          .
          <article-title>Matrix Factorization Techniques for Recommender Systems</article-title>
          . IEEE Computer
          <volume>42</volume>
          (
          <issue>8</issue>
          ), pp.
          <fpage>30</fpage>
          -
          <lpage>37</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <surname>Koren</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bell</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <year>2011</year>
          .
          <article-title>Advances in Collaborative Filtering</article-title>
          .
          <source>Recommender Systems Handbook</source>
          , pp.
          <fpage>145</fpage>
          -
          <lpage>186</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <surname>Li</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yang</surname>
            ,
            <given-names>Q.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Xue</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          <year>2009</year>
          .
          <article-title>Can Movies and Books Collaborate? Cross-domain Collaborative Filtering for Sparsity Reduction</article-title>
          .
          <source>In Proceedings of the 21st International Joint Conference on Artificial Intelligence</source>
          , pp.
          <fpage>2052</fpage>
          -
          <lpage>2057</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <surname>Pan</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Xiang</surname>
            ,
            <given-names>E.W.</given-names>
          </string-name>
          , Liu,
          <string-name>
            <given-names>N.N.</given-names>
            ,
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <surname>Q.</surname>
          </string-name>
          <year>2010</year>
          .
          <article-title>Transfer Learning in Collaborative Filtering for Sparsity Reduction</article-title>
          .
          <source>In Proceedings of the 24th AAAI Conference on Artificial Intelligence</source>
          , pp.
          <fpage>210</fpage>
          -
          <lpage>235</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <surname>Shapira</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rokach</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Freilikhman</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <year>2013</year>
          .
          <article-title>Facebook Single and Cross Domain Data for Recommendation Systems</article-title>
          .
          <source>User Modeling and User-Adapted Interaction</source>
          <volume>23</volume>
          (
          <issue>2-3</issue>
          ), pp.
          <fpage>211</fpage>
          -
          <lpage>247</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <surname>Shi</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Larson</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hanjalic</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <year>2011</year>
          .
          <article-title>Tags as Bridges between Domains: Improving Recommendation with Taginduced Cross-domain Collaborative Filtering</article-title>
          .
          <source>In Proceedings of the 19th International Conference on User Modeling, Adaption, and Personalization</source>
          , pp.
          <fpage>305</fpage>
          -
          <lpage>316</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <surname>Tiroshi</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Berkovsky</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kaafar</surname>
            ,
            <given-names>M. A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chen</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kuflik</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          <year>2013</year>
          .
          <article-title>Cross Social Networks Interests Predictions Based on Graph Features</article-title>
          .
          <source>In Proceedings of the 7th ACM Conference on Recommender Systems</source>
          , pp.
          <fpage>319</fpage>
          -
          <lpage>322</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <surname>Winoto</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tang</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          <year>2008</year>
          .
          <article-title>If You Like the Devil Wears Prada the Book, Will You also Enjoy the Devil Wears Prada the Movie? A Study of Cross-Domain Recommendations</article-title>
          .
          <source>New Generation Computing 26</source>
          , pp.
          <fpage>209</fpage>
          -
          <lpage>225</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <surname>Zhang</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cao</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yeung</surname>
          </string-name>
          , D.-Y.
          <year>2010</year>
          .
          <article-title>Multi-Domain Collaborative Filtering</article-title>
          .
          <source>In Proceedings of the 26th Conference on Uncertainty in Artificial Intelligence</source>
          , pp.
          <fpage>725</fpage>
          -
          <lpage>732</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>