<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>Y. Yao); chanw@sz.tsinghua.edu.cn
(W. K. V. Chan)</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>CDR-Adapter: Learning Adapters to Dig Out More Transferring Ability for Cross-Domain Recommendation Models</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Yanyu Chen</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Yao Yao</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Wai Kin Victor Chan</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Li Xiao</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Kai Zhang</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Liang Zhang</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Yun Ye</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Ant Group</institution>
          ,
          <addr-line>Shanghai, China, 200135</addr-line>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Shenzhen International Graduate School, Tsinghua University</institution>
          ,
          <addr-line>Shenzhen, Guangdong, China, 518055</addr-line>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Tsinghua-Berkeley Shenzhen Institute, Shenzhen International Graduate School, Tsinghua University</institution>
          ,
          <addr-line>Shenzhen, Guangdong, China, 518055</addr-line>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2023</year>
      </pub-date>
      <volume>000</volume>
      <fpage>0</fpage>
      <lpage>0003</lpage>
      <abstract>
        <p>Data sparsity and cold-start problems are persistent challenges in recommendation systems. Crossdomain recommendation (CDR) is a promising solution that utilizes knowledge from the source domain to improve the recommendation performance in the target domain. Previous CDR approaches have mainly followed the Embedding and Mapping (EMCDR) framework, which involves learning a mapping function to facilitate knowledge transfer. However, these approaches necessitate re-engineering and re-training the network structure to incorporate transferrable knowledge, which can be computationally expensive and may result in catastrophic forgetting of the original knowledge. In this paper, we present a scalable and eficient paradigm to address data sparsity and cold-start issues in CDR, named CDR-Adapter, by decoupling the original recommendation model from the mapping function, without requiring reengineering the network structure. Specifically, CDR-Adapter is a novel plug-and-play module that employs adapter modules to align feature representations, allowing for flexible knowledge transfer across diferent domains and eficient fine-tuning with minimal training costs. We conducted extensive experiments on the benchmark dataset, which demonstrated the efectiveness of our approach over several state-of-the-art CDR approaches.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Cross-domain recommendation</kwd>
        <kwd>Decoupling representation learning</kwd>
        <kwd>Cold-start problem</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Recommender systems have become increasingly prevalent in online platforms, serving as vital
components in providing personalized user experiences [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. However, a significant challenge in
constructing such systems is the provision of accurate recommendations to new users, who
possess little or no interaction history with the system, commonly referred to as cold-start users.
To overcome this challenge, cross-domain recommendation (CDR) models have emerged as a
promising solution that leverages knowledge and information from various domains to enhance
recommendation performance.
      </p>
      <p>CDR comprises two domains: the target domain and the source domain, with users in the
source domain classified into two groups: overlapping users and cold-start users. Overlapping
users have active records in both domains, while the remaining users are considered
coldstart users in the target domain. The primary objective of CDR is to boost recommendation
performance for cold-start users in the target domain.</p>
      <p>
        Earlier CDR models [
        <xref ref-type="bibr" rid="ref2 ref3">2, 3</xref>
        ] mainly focused on learning a cross-domain mapping function
that could transfer information and knowledge from the source domain to the target domain,
restricted to only the relevant information of the overlapping users, which usually led to
suboptimal recommendation results. Subsequent works [
        <xref ref-type="bibr" rid="ref4 ref5 ref6 ref7">4, 5, 6, 7</xref>
        ] have improved upon the earlier
models by enriching and expanding the transferrable information, such as user-item interaction,
thereby reducing the dependence on overlapping users. Despite these advancements, these CDR
models continue to have several limitations. Firstly, these models often require a large number
of overlapping users to transfer information, which leads to data sparsity issues and models
becoming biased toward overlapping users. Secondly, they generally require re-engineering
and re-training the network structure to incorporate transferrable information, leading to high
computational expenses and risking the catastrophic forgetting of original knowledge. Thirdly,
the mapping function learned by earlier models is typically inflexible, which limits the model’s
ability to transfer knowledge across various domains.
      </p>
      <p>
        Certain approaches attempt to disentangle domain-specific and cross-domain information.
For instance, Cao et al. [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] proposed two mutual-information-based disentanglement
regularizers that exclusively transfer domain-shared information to enhance model recommendation
performance. Additionally, Cao et al. [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] proposed two information bottleneck regularizers
to simultaneously model domain-specific and cross-domain information, deriving unbiased
representations. However, these CDR models necessitate adjusting and retraining the original
network to achieve a domain-shared latent space, where representations from diferent domains
are aligned to facilitate knowledge transfer. Training large eCommerce recommendation models
can be computationally expensive, and restructuring and retraining the network can alter
the intrinsic semantic space of the pre-trained model. Generally, this paradigm sufers from
ineficient training and catastrophic forgetting of the original knowledge of pre-trained models.
      </p>
      <p>
        To address these challenges, we propose a novel cross-domain recommendation framework,
CDR-Adapter, inspired by the adapter technique in natural language processing [
        <xref ref-type="bibr" rid="ref10 ref11">10, 11</xref>
        ]. Our
approach decouples recommendation models from the mapping function by learning an adapter
that aligns feature representations of recommendation models across the source and target
domains. This preserves the original model information while enabling flexible knowledge transfer.
Requiring much less overlapping user information, our approach mitigates the challenges of
data sparsity. Scalable and eficient, it allows for eficient fine-tuning with minimal training
cost. We evaluate our approach on several benchmark datasets, and the results demonstrate
that our method outperforms several state-of-the-art CDR approaches.
      </p>
      <p>Our main contributions are as follows:
• We propose a novel CDR framework that leverages adapter modules to align feature
representations, enabling flexible knowledge transfer across diferent domains and eficient
ifne-tuning with minimal training cost.
• We introduce a scalable and eficient solution to the cold-start problem in CDR by
decoupling recommendation models from the mapping function, without adjusting pre-trained
models or facing the problem of catastrophic forgetting.
• We conduct extensive experiments on several benchmark datasets and demonstrate the
efectiveness of our approach over several state-of-the-art CDR approaches.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Methodology</title>
      <sec id="sec-2-1">
        <title>2.1. Notations and Problem Formulation</title>
        <p>Notations. Without loss of generality, we consider a general CDR scenario where there exist
two domains:  and  . In this paper, we aim to design an efective and eficient method
to improve the recommendation performance in both domains simultaneously, so we do not
explicitly diferentiate between the source domain and the target domain. Each domain has its
corresponding user set  and item set . For simplicity, we further introduce binary matrix
 ∈ {0, 1}||×|| , whose elements indicate whether there is an interaction between a user and
an item.</p>
        <p>Problem Formulation. Given the observed interaction data  of both domains, we aim to
make recommendations for cold-start users in domain  , who are only observed in domain 
and do not have interaction records in domain  . Formally, given a cold-start user  ∈   , we
would like to recommend a list of items from  (vice versa, in the case of users from   and
items from  ).</p>
      </sec>
      <sec id="sec-2-2">
        <title>2.2. Overview of the CDR-Adapter Framework</title>
        <p>Pre-Trained Model</p>
        <p>Adapter</p>
        <p>Downstream Model</p>
        <p>Tuned</p>
        <p>Frozen</p>
        <p>In this paper, we propose to learn a small and simple CDR-Adapter to align the
representations generated by two pre-trained models, which can efectively decouple the pre-trained
representation model from the downstream recommendation model. The illustration of the
components of our method is presented in Figure 1, which is composed of the pre-trained
representation models, Adapters and downstream recommendation models.</p>
        <p>Generally, the common recommendation approaches can be divided into two modules: the
representation learning module and the downstream recommendation module. The
representation learning module is to extract the characteristics of the features and obtains the dense
representations in the latent space. The representations learned by large-scale pre-trained
representation models have robust generalization properties and can be applied to downstream
tasks. The downstream recommendation module is to obtain the recommendation lists for users,
which are adaptive to specific scenarios and usually hard to be transferred and generalized. We
start from the representation models, which provide initialized user/item representations (a.k.a.
latent variables) in each domain for the following components. Then, the adapter module along
with three regularizers, regarded as an auxiliary module, is proposed to align the representations
to extract the knowledge across the domain. Specifically, the dimensions of the input and output
are aligned, so there are no compatibility issues. Afterward, the alignment representations are
reconstructed as input for downstream models to get final recommendation lists for cold-start
users.</p>
        <p>Figure 2 illustrates the inference procedure of CDR-Adapter in practice. Above the dotted line
shows the inference process for cold-start users in domain  by transferring the knowledge from
domain . Note that the inference procedure for cold-start users in domain  is similar and
omitted for simplicity. We present an example to further facilitate understanding. Suppose the
task is to obtain a recommendation list for cold-start user A in domain  . We try to transfer the
knowledge of user A in domain  to improve the recommendation performance for cold-start
user A in domain  . The pre-trained representation model generates the representation of
user A in the latent space and feeds representations to the Adapter. The prior is to get the
alignment representation for user A in domain  , which can efectively transfer the
crossdomain knowledge. The decoder is to reconstruct the alignment representation in order to
ifne-tune the input of the downstream model without retraining the model, which fully uses
the knowledge in domain  to get the more efective representation and saves amounts of
computation cost. Finally, through the downstream recommendation model (e.g. Top-N
multihop neighborhood recommendation), the recommendation list for cold-start user A can be
obtained.</p>
      </sec>
      <sec id="sec-2-3">
        <title>2.3. Adapter Architecture</title>
        <p>In this paper, we design three tasks as regularizers for the adapter module to capture the
correlations across domains, aiming to learn unbiased representations with domain-shared
information. Specifically, the contrastive cross-domain regularizer aims to capture the users’
correlation across domains. The scale alignment regularizer aims to linearly align the scale of
users to map each other in the two domains. The reconstruction information regularizer aims
to minimize the loss of information after alignment and reconstruction procedures, which can
guarantee that the reconstructed representations can be directly used as input in the downstream
recommendation models without retraining the models.</p>
        <p>Contrastive Cross-Domain Regularizer. In order to better align the representations
from each domain, we design the contrastive cross-domain regularizer, which improves the
capability to make recommendations in both domains. The same overlapping users (′ , ′ )
are regarded as the positive pair. The diferent users (′ , ′ ) are regarded as the negative
pair. We refine the representations of users by measuring the mutual information between the
representations from domain  and domain  . Specifically, the distance between positive pairs
is minimized to make the representations aligned in cross-domain, while the distance between
negative pairs is maximized to distinguish diferent users. In this way, the user representations
are enforced to capture the domain-shared information from both domains, thus deriving
the general representations for CDR. Thus, the contrastive cross-domain regularizer can be
formulated as follows.</p>
        <p>⎡
ℒ1 = −  ⎣log</p>
        <p>exp [︀  (︀ ′ , ′ )︀ / ]︀
∑︀,=1 [︁ (︁ ′ , ′ )︁ /</p>
        <p>⎤
]︁ ⎦
where ′ is the output alignment representation of the prior layer. ′ and ′ are the
representation of the same overlapping user  in domain  and domain  , while ′ and ′
are the representation of diferent overlapping user  in domain  and domain  .  &gt; 0 is a
tunable temperature hyperparameter. (′ , ′ ) is the cosine similarity between vector ′
and ′ , e.g.,  (,  ) can be calculated as follows:
 (,  ) =</p>
        <p>⟨,  ⟩ .</p>
        <p>‖‖ · ‖  ‖</p>
        <p>Scale Alignment Regularizer. In order to align the scale of overlapping user
representations in each domain, we design a linear scale alignment regularizer to extract domain-shared
information to the greatest extent. Ideally, we hope ′ = ′ , since they are essentially the
same users. However, this task is hard to learn and requires high precision of the prior model.
So here we propose an approximation method, which is to train the linear transformation which
is essentially reciprocal. The formulation is as follows.</p>
        <p>1 (︀ ′ )︀
2 (︀ ′ )︀
=  1′ +  1
=  2′ +  2
Specifically, here we train the Multi-Layer Perceptron (MLP) without the activation layer to
obtain the parameters of  and  .</p>
        <p>Then, the scale alignment regularizer is formulated as follows.</p>
        <p>ℒ2 = ⃦⃦ 1 (︀ ′ )︀ − ′ ⃦⃦ 2 + ⃦⃦ 2 (︀ ′ )︀ − ′ ⃦⃦ 2</p>
        <p>Reconstruction Information Regularizer. To reconstruct the cold-start users’
representation for direct prediction in the downstream recommendation models without retraining the
downstream model, we propose the reconstructing information regularizer. In this part, we
hope the reconstructed representation through the decoder can maintain the cross-domain
knowledge and has similarity with the original representation, simultaneously. In this way, the
reconstructed representations can be directly used as input for the downstream model to obtain
the final recommendation lists. The reconstructing information regularizer is formulated as
follows.</p>
        <p>ℒ3 = ⃦⃦  − ˆ ⃦⃦ 2 + ⃦⃦  − ˆ ⃦⃦ 2</p>
        <p>Optimizing the Overall Model. Based on the above three regularizers, we can optimize
the overall model in an end-to-end framework. In summary, we build the prior and decoder to
transfer the overlapping users’ knowledge. The conclude final objective function is as follows.</p>
        <p>ℒ =  1ℒ1 +  2ℒ2 +  3ℒ3,
where  1,  2, and  3 are the hyper-parameter, which control the importance of each regularizer.</p>
      </sec>
      <sec id="sec-2-4">
        <title>2.4. The Properties of CDR-Adapter</title>
        <p>The proposed CDR-Adapter can be used to learn alignment representation with a relatively
small amount of data, which can perfectly address the challenges of data sparsity and
coldstart problems in the recommendation field. The CDR-Adapter, as the extra network to inject
condition information, also has the following properties.</p>
        <p>Plug-and-Play. The original network topology and transfer ability of the existing models
does not be changed by adding this CDR-Adapter. Besides, the CDR-Adapter can also be easily
composed to any pre-trained and downstream models. There is no need to retrain the pre-trained
models and downstream task models.</p>
        <p>Simple and Small. It can be easily inserted into any recommendation model with low
training costs and fully transfer cross-domain knowledge. It has a small number of parameters
and small storage space, which will not introduce much computation cost.</p>
        <p>Cascade Composable and Flexible. As to make recommendations in any two domains
from the total  domains, the traditional methods need to train ( − 1)/2 models, while our
method only needs training  − 1 adapters to make cascaded recommendations. For example,
we have three scenarios, A, B, and C. Instead of training three adapters as, A↔B, A↔C, B↔C, it
only needs to train two adapters in any two of scenarios, such as A↔B, B↔C. When it refers to
making recommendations under the A↔C scenario, we can cascade the two adapters to realize
A↔B↔C.</p>
        <p>Generalizable. Once trained, it can be used on custom models as long as they are fine-tuned
from the same cross-domain models. No retraining is required for this transfer.</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Experiments</title>
      <p>Extensive experiments are conducted to answer the following questions:
Q1: How does our CDR-Adapter perform compared with the competitive baselines?
Q2: How does the proportion of overlapping users impact the model performance?
Q3: Does CDR-Adapter indeed infer more accurate representations of the cold-start users in
the latent space?
Q4: Does our CDR-Adapter reach the desirable decoupling? Further, what impact does our
CDR-Adapter achieve?</p>
      <sec id="sec-3-1">
        <title>3.1. Experimental Settings</title>
        <p>
          Datasets. Following previous works[
          <xref ref-type="bibr" rid="ref7">7</xref>
          ], we adopt the same datasets, and the same preprocessing
settings to build our CDR scenarios. Specifically, we conduct experiments on the largest scale
of public datasets: Amazon. The most popular pair of domains are selected to evaluate our
CDR-Adapters for the bi-directional CDR scenarios. The detailed statistics of each domain are
listed in Table 1.
        </p>
        <p>Implementation Details. We filter out the items that have fewer than 10 interactions and
the users that have fewer than 5 interactions in each domain, making the users/items access to
learning efective representation from each domain. We randomly select 20% as cold-start users
for testing and validation (e.g. the 10% from the Book domain to recommend items in the Movie
domain and the residual 10% from the Movie domain to recommend items in the Book domain)
and the remaining users are used for training.</p>
        <p>Baselines. In order to verify the efectiveness of our method to cold-start users, we compare
our CDR-Adapter with the following state-of-the-art baselines, which can be divided into three
groups.</p>
        <p>
          Single-domain recommendation: The methods in this group consider all domains as a whole
single domain. We construct a unified matrix so that it includes all users and items as its rows
and columns, respectively. Then, we apply the following widely-used methods.
• CML[
          <xref ref-type="bibr" rid="ref12">12</xref>
          ] models the user and item representation by matrix learning, which calculates L2
distance and supposes that the distance between a user and interacted items is small while
the distance between a user and not interacted items is large. CML is a state-of-the-art
collaborative filtering method.
• BPR[
          <xref ref-type="bibr" rid="ref13">13</xref>
          ] models the latent vector by pairwise ranking loss, which optimizes the order of
the inner product of user and item latent vectors.
• NGCF[14] is a graph neural network method to learn user and item representations,
which uses GCN to learn high-order information between users and items.
        </p>
        <p>Single-directional cross-domain recommendation: Single-domain recommendation methods
fail to consider the diferences between two domains, which makes it hard to efectively transfer
knowledge. To better transfer useful knowledge, researchers propose single-direction CDR
approaches. We adopt several typical CDR models as baselines. Note that, all of these following
methods transfer information from the source to the target domain in one direction, we run
two times to achieve bi-directional CDR.</p>
        <p>
          • EMCDR[
          <xref ref-type="bibr" rid="ref2">2</xref>
          ] first learns user and item representations, and then adopts a network to
bridge the representations from the source domain to the target domain.
• SSCDR[
          <xref ref-type="bibr" rid="ref4">4</xref>
          ] is a self-supervised bridge-based method that gets the final item list by
multihop neighborhood inference.
• TMCDR[
          <xref ref-type="bibr" rid="ref5">5</xref>
          ] is the expansion of EMCDR, which designs a meta-learning framework for
        </p>
        <p>
          CDR to cold-start users.
• CLCDR[
          <xref ref-type="bibr" rid="ref7">7</xref>
          ] is a contrastive learning-based CDR model, which simultaneously transfers
knowledge about overlapping users and user-item interactions to optimize the user and
item representations.
        </p>
        <p>Bi-directional cross-domain recommendation: Since our method can realize bi-directional CDR,
We also compare our method with the following bi-directional CDR methods.
• DAN[15] captures high-order relationships to learn user preferences by utilizing the
user-item interaction graph end-to-end.
• DTCDR[16] designs a kind of multi-task learning to combine the representation of users
across the domains and improve the recommendation performance on both richer and
sparser domains simultaneously.
• SA-VAE[17] is the state-of-the-art bi-directional CDR method, which is a variational
method that utilizes the VAE framework to generate the latent matrix for each domain,
and then trains the mapping function for prediction.</p>
        <p>
          Evaluation Metrics. Following the previous works[
          <xref ref-type="bibr" rid="ref4 ref5">4, 5</xref>
          ], we use the leave-one-out evaluation
method to verify the efectiveness of our CDR-Adapter. For instance, given a ground truth
interaction (,  ) in domain  , we first randomly select 999 items from the item set 
as negative samples. Then, we calculate 1000 records (1 positive and 999 negative samples)
by the learned representation ˆ from domain  and  from domain  . Next, we rank
the recommendation list and use evaluation metrics: Hit Rate(HR), Normalized Discounted
Cumulative Gain(NDCG), and Mean Reciprocal Rank(MRR) to show the performance of top-N
recommendations.
        </p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Overall Performance (Q1)</title>
        <p>The overall performance is listed in Table 2, which reports the mean result under HR, NDCG,
and MRR over ten runs with outliers removed. From an overall point of view, our CDR-Adapter
method obtains statistically significant improvements compared with the several baselines. We
analyze the results from several following perspectives.
* indicates that the improvements are statistically significant for p &lt; 0.05 judged with the runner-up
result in each case by paired t-test.</p>
        <p>Comparison with Single-Domain Models. (1) First, we found that, the graph-based model
NGCF consistently outperform CML and BPR in term of all evaluation metrics, which indicates
that transferring the multi-hop neighborhood knowledge is efective for learning better user
and item representations. (2) Second, the performance of our CDR-Adapter with diferent
pre-trained models demonstrates the significant importance of user and item representations
generated by pre-trained models. The final performance of the CDR-Adapter model is positively
correlated with the domain-specific representation generated by the pre-trained model.</p>
        <p>(3) Third, these single-domain models perform mostly worse than the cross-domain models,
due to these methods ignoring the diference between diferent domains just combining them
together in a simple way and making recommendations, which is hard to learn the transferable
knowledge for cold-start users. So it is necessary to dig out transferring ability for CDR to
cold-start users.</p>
        <p>Comparison with Single-Directional Cross-Domain Recommendation Models. (1)
First, in general, the cross-domain methods are superior to corresponding single-domain
methods, which demonstrates that adopting diferent transferring components for CDR is better
than using one single neural network to model the mixed matrix.</p>
        <p>(2) Second, the improvement of the EMCDR model is limited and even worse than some
singledomain models, which indicates that a simple function may be not efective to learn the complex
mapping relation of cross-domain representations. (3) Third, since the single-directional CDR
models could only improve the recommendation performance in the target domain, it should be
run two times to achieve bi-directional CDR, which requires high computing costs and time
consumption. Besides, the transfer might be negative transferring in some cases. (4) Forth,
since EMCDR-based models mainly transfer the overlapping users’ information, user-item
interactions, and even user-user social relationships, the generative representations would be
biased toward overlapping users. Compared with all EMCDR-based baselines, our CDR-Adapter
achieves statistically significant improvements with all evaluation metrics, which demonstrates
that learning the mapping function on the biased representations can be hard to obtain the
optimal results. In contrast, our CDR-Adapter digs out the transferring ability of domains and
utilizes three kinds of regularizers to encourage the representations to focus on aligning the
domain-specific representations. In this way, the unbiased cold-start user representation on
each domain can be directly obtained in the target domain.</p>
        <p>Comparison with Bi-Directional Cross-Domain Recommendation Models. (1) The
recommendation results of DTCDR and SA-VAE in two domains improve, due to their dual
objective optimization. (2) Since those bi-directional CDR models jointly optimize the objective
by overlapping users, the existing models are still not capable of efectively capturing the
domain-shared information in the small number of overlapping users case. While our
CDRAdapter can achieve better performance by aligning the representation of both overlapping
users and domain-specific users.</p>
      </sec>
      <sec id="sec-3-3">
        <title>3.3. The Impact of Overlapping Users (Q2)</title>
        <p>To study the robustness of our CDR-Adapter method regarding the proportion of overlapping
users, we conduct several experiments with a certain proportion  ∈ {5%, 20%, 50%, 100%}
of overlapping users for training. Table 3 and Figure 3 show the recommendation performances
of SA-VAE and CDR-Adapter on the cross-domain scenario (e.g., the target Movie domain with
the source Book domain).</p>
        <p>From Table 3, we have the following findings. (1) With the proportion of overlapping
users increasing, the performance of SA-VAE steadily improves, which relies on transferring</p>
        <p>The robustness performance (%). The  denotes the proportion of overlapping users in the training</p>
        <p>NDCG@10

5%
20%
50%
100%
overlapping users to enhance the correlation across the domains. While CDR-Adapter is not
sensitive to the proportion of overlapping users. (2) Compare with the strongest baseline,
CDR-Adapter performs well even in the 5% overlapping users, which shows strong robustness.
The results reveal that our CDR-Adapter is capable of efectively mapping even in the small
number of overlapping users in the training process, due to its useful alignment function, which
successfully overcomes the limitation of the existing methods. From Figure 3, we find that
CDR-Adapter is robust enough, especially in the data sparsity domain (e.g., Movie). In contrast,
the improvement is not obvious (e.g., Book). In addition, in the case of a few overlapping users
(e.g., 5%), our CDR-Adapter method achieves nearly the same performance as the SA-VAE in
the case of 100% overlapping users.</p>
      </sec>
      <sec id="sec-3-4">
        <title>3.4. The Analysis of Latent Space Inference (Q3)</title>
        <p>To further investigate why our CDR-Adapter outperforms the state-of-the-art method, we
conduct both qualitative and quantitative analyses on the target domain latent space.</p>
        <p>As for the qualitative analysis, we make the target domain latent space visualization by
adopting t-distributed Stochastic Neighbor Embedding[18] (t-SNE) to analyze the transferring
quality compared with ground truth. As Figure 4 shows, the first row is the user representation
in the target domain latent space. The inferred user representations by EMCDR, CLCDR,
and CDR-Adapter are shown with blue dots and the ground truths are shown with pink dots.
Because the EMCDR method infers user representation just by transferring the overlapping
user information, the inferred representations are not very close to ground truths. The CLCDR
(a)
(e)
(b)
(f)
(c)
(g)
method transfers the information of overlapping users and user-item interactions and designs
two contrastive-based domain-specific and domain-shared loss functions, which can improve
the representation quality, shown in Figure 4(b). Our method CDR-Adapter aligns the
domainspecific representation instead of mapping and changing the latent space, which can infer the
cold-start user representation close to the ground truths.</p>
        <p>The second row in Figure 4 is the gathering situation. The black dots and translucent circles
denote the overlapping users and their surrounding areas, respectively. The user representation
obtained by EMCDR and CLCDR inference is biased toward overlapping users. We can find
that there are more users clustered around overlapping users, because these two methods
improve the cold-start user recommendation efect mainly by transferring overlapping users’
information. While our CDR-Adapter can infer the less biased cold-start user representation in
the target domain latent space.</p>
        <p>For the quantitative analysis, we measure the actual average distance between the inferred
latent representation and their ground truths, shown in Table 4. Our CDR-Adapter performs
best with the smallest distance.</p>
      </sec>
      <sec id="sec-3-5">
        <title>3.5. The Analysis of Disentanglement (Q4)</title>
        <p>Table 5 demonstrates that our CDR-Adapter method achieves the desirable disentanglement to
learn domain-specific and cross-domain representations for users. Specifically, we calculate the
average KL divergence to measure the mutual information of domain-specific user representation
and cross-domain user representation after the training process of both domains is finished
(higher KL divergence means lower mutual information). According to Table 5, it is obvious
that our CDR-Adapter method KL divergence is much higher than the EMCDR method, which
verifies our CDR-Adapter achieves the outstanding disentanglement between domain-specific
and cross-domain representations.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Conclusion</title>
      <p>In summary, we proposed a novel and scalable cross-domain recommendation paradigm that
addresses the limitations of existing approaches by introducing an adapter plugin that decouples
the original representation model from the mapping function. Our approach can align the
feature representations of the recommendation models in both domains and enable eficient
ifnetuning with minimal training cost. It can be easily plugged in between the pre-train
representation module and the downstream recommendation module of any kind of
crossdomain recommendation model. We believe that our framework provides a more scalable and
eficient solution to the cold-start problem in the cross-domain recommendation and ofers a
promising direction for future research.</p>
    </sec>
    <sec id="sec-5">
      <title>Acknowledgments</title>
      <p>This research was funded by the Ant Group through CCF-Ant Research Fund
(CCFAFSG RF20220216), the Science and Technology Innovation Commission of Shenzhen
(JCYJ20210324135011030), Guangdong Pearl River Plan (2019QN01X890), and National Natural
Science Foundation of China (Grant No. 71971127).
[14] X. Wang, X. He, M. Wang, F. Feng, T.-S. Chua, Neural graph collaborative filtering, in:
Proceedings of the 42nd international ACM SIGIR conference on Research and development
in Information Retrieval, 2019, pp. 165–174.
[15] B. Wang, C. Zhang, H. Zhang, X. Lyu, Z. Tang, Dual autoencoder network with swap
reconstruction for cold-start recommendation, in: Proceedings of the 29th ACM International
Conference on Information &amp; Knowledge Management, 2020, pp. 2249–2252.
[16] F. Zhu, C. Chen, Y. Wang, G. Liu, X. Zheng, Dtcdr: A framework for dual-target
crossdomain recommendation, in: Proceedings of the 28th ACM International Conference on
Information and Knowledge Management, 2019, pp. 1533–1542.
[17] A. Salah, T. B. Tran, H. Lauw, Towards source-aligned variational models for cross-domain
recommendation, in: Proceedings of the 15th ACM Conference on Recommender Systems,
2021, pp. 176–186.
[18] L. Van der Maaten, G. Hinton, Visualizing data using t-sne., Journal of machine learning
research 9 (2008).</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Yao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>He</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Sheng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Xiao</surname>
          </string-name>
          , H. Cao,
          <article-title>i-razor: A diferentiable neural input razor for feature selection and dimension search in dnn-based recommender systems</article-title>
          ,
          <source>IEEE Transactions on Knowledge and Data Engineering</source>
          (
          <year>2023</year>
          )
          <fpage>1</fpage>
          -
          <lpage>14</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>T.</given-names>
            <surname>Man</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Shen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Jin</surname>
          </string-name>
          , X. Cheng,
          <article-title>Cross-domain recommendation: An embedding and mapping approach</article-title>
          .,
          <source>in: International Joint Conference on Artificial Intelligence</source>
          , volume
          <volume>17</volume>
          ,
          <year>2017</year>
          , pp.
          <fpage>2464</fpage>
          -
          <lpage>2470</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>W.</given-names>
            <surname>Fu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Peng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <article-title>Deeply fusing reviews and contents for cold start users in cross-domain recommendation systems</article-title>
          ,
          <source>in: Proceedings of the AAAI Conference on Artificial Intelligence</source>
          ,
          <year>2019</year>
          , pp.
          <fpage>94</fpage>
          -
          <lpage>101</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>S.</given-names>
            <surname>Kang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Hwang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Yu</surname>
          </string-name>
          ,
          <article-title>Semi-supervised learning for cross-domain recommendation to cold-start users</article-title>
          ,
          <source>in: Proceedings of the 28th ACM International Conference on Information and Knowledge Management</source>
          ,
          <year>2019</year>
          , pp.
          <fpage>1563</fpage>
          -
          <lpage>1572</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Ge</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Zhuang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Xie</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Xi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Lin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>He</surname>
          </string-name>
          ,
          <article-title>Transfer-meta framework for cross-domain recommendation to cold-start users</article-title>
          ,
          <source>in: Proceedings of the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval</source>
          ,
          <year>2021</year>
          , pp.
          <fpage>1813</fpage>
          -
          <lpage>1817</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>C.</given-names>
            <surname>Zhao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Xiao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Deng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Sun</surname>
          </string-name>
          , Catn:
          <article-title>Cross-domain recommendation for cold-start users via aspect transfer network</article-title>
          ,
          <source>in: Proceedings of the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval</source>
          ,
          <year>2020</year>
          , pp.
          <fpage>229</fpage>
          -
          <lpage>238</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Yao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W. K. V.</given-names>
            <surname>Chan</surname>
          </string-name>
          ,
          <article-title>Clcdr: Contrastive learning for cross-domain recommendation to cold-start users</article-title>
          ,
          <source>in: Neural Information Processing: 29th International Conference, ICONIP</source>
          <year>2022</year>
          ,
          <string-name>
            <given-names>Virtual</given-names>
            <surname>Event</surname>
          </string-name>
          ,
          <source>November 22-26</source>
          ,
          <year>2022</year>
          , Proceedings,
          <string-name>
            <surname>Part</surname>
            <given-names>II</given-names>
          </string-name>
          , Springer,
          <year>2023</year>
          , pp.
          <fpage>331</fpage>
          -
          <lpage>342</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>J.</given-names>
            <surname>Cao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Lin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Cong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Ya</surname>
          </string-name>
          , T. Liu,
          <string-name>
            <given-names>B.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <article-title>Disencdr: learning disentangled representations for cross-domain recommendation</article-title>
          ,
          <source>in: Proceedings of the 45th International ACM SIGIR Conference on Research and Development in Information Retrieval</source>
          ,
          <year>2022</year>
          , pp.
          <fpage>267</fpage>
          -
          <lpage>277</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>J.</given-names>
            <surname>Cao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Sheng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Cong</surname>
          </string-name>
          , T. Liu,
          <string-name>
            <given-names>B.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <article-title>Cross-domain recommendation to cold-start users via variational information bottleneck</article-title>
          ,
          <source>in: 2022 IEEE 38th International Conference on Data Engineering (ICDE)</source>
          , IEEE,
          <year>2022</year>
          , pp.
          <fpage>2209</fpage>
          -
          <lpage>2223</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>R.</given-names>
            <surname>He</surname>
          </string-name>
          , L. Liu,
          <string-name>
            <given-names>H.</given-names>
            <surname>Ye</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Tan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Ding</surname>
          </string-name>
          , L. Cheng, J.
          <string-name>
            <surname>Low</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <string-name>
            <surname>Bing</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <string-name>
            <surname>Si</surname>
          </string-name>
          ,
          <article-title>On the efectiveness of adapter-based tuning for pretrained language model adaptation</article-title>
          ,
          <source>in: Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing</source>
          (Volume
          <volume>1</volume>
          :
          <string-name>
            <surname>Long</surname>
            <given-names>Papers)</given-names>
          </string-name>
          ,
          <year>2021</year>
          , pp.
          <fpage>2208</fpage>
          -
          <lpage>2222</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>N.</given-names>
            <surname>Ding</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Qin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Wei</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Su</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Hu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <surname>C.-M. Chan</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          <string-name>
            <surname>Chen</surname>
          </string-name>
          , et al.,
          <article-title>Parameter-eficient fine-tuning of large-scale pre-trained language models</article-title>
          ,
          <source>Nature Machine Intelligence</source>
          (
          <year>2023</year>
          )
          <fpage>1</fpage>
          -
          <lpage>16</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <surname>C.-K. Hsieh</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <string-name>
            <surname>Yang</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          <string-name>
            <surname>Cui</surname>
            , T.-
            <given-names>Y.</given-names>
          </string-name>
          <string-name>
            <surname>Lin</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Belongie</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Estrin</surname>
          </string-name>
          ,
          <article-title>Collaborative metric learning</article-title>
          ,
          <source>in: Proceedings of the 26th international conference on world wide web</source>
          ,
          <year>2017</year>
          , pp.
          <fpage>193</fpage>
          -
          <lpage>201</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>S.</given-names>
            <surname>Rendle</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Freudenthaler</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Gantner</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Schmidt-Thieme</surname>
          </string-name>
          ,
          <article-title>Bpr: Bayesian personalized ranking from implicit feedback</article-title>
          ,
          <source>arXiv preprint arXiv:1205.2618</source>
          (
          <year>2012</year>
          ).
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>