=Paper= {{Paper |id=Vol-2431/paper8 |storemode=property |title=Latent Modeling of Unexpectedness for Recommendations |pdfUrl=https://ceur-ws.org/Vol-2431/paper8.pdf |volume=Vol-2431 |authors=Pan Li,Alexander Tuzhilin |dblpUrl=https://dblp.org/rec/conf/recsys/LiT19 }} ==Latent Modeling of Unexpectedness for Recommendations== https://ceur-ws.org/Vol-2431/paper8.pdf
                                                               Latent Modeling of Unexpectedness
                                                               for Recommendations
                                              Pan Li                                                    Alexander Tuzhilin
                                              New York University                                       New York University
                                              New York, NY, USA                                         New York, NY, USA
                                              pli2@stern.nyu.edu                                        atuzhili@stern.nyu.edu

Figure 1: Visualization of Latent Convex
Hull. Unexpectedness is defined as the dis-   ABSTRACT
tance between new item and latent convex      Unexpectedness constitutes an important factor for recommender system to improve user satisfaction
hull generated by all consumed items.         and avoid filter bubble issues. Previous methods model unexpectedness in the feature space, making
                                              them difficult to capture the latent, complex and heterogeneous interactions between users and
                                              items. In this paper, we propose to model unexpectedness in the latent space and utilize a latent
                                              convex hull structure to provide unexpected recommendations, as illustrated in Figure 1. Extensive
                                              experiments on two real-world datasets demonstrate effectiveness of latent unexpectedness over
                                              explicit unexpectedness and show that the proposed model significantly outperforms baseline models
                                              in terms of unexpectedness measures while achieving the same level of accuracy.

                                              KEYWORDS
                                              Unexpected Recommendation, Latent Embeddings, Heterogeneous Information Network

                                              INTRODUCTION AND RELATED WORK
                                              Recommender systems have been playing an important role in assisting users in filtering for the best
                                              content and shaping their consumption behaviors. To solve the problem of filter bubbles and boredom
Figure 2: Visualization of Heterogeneous      [11, 12], researchers introduce the measure of unexpectedness [10, 14] beyond accuracy, the goal of
Information Network                           which is to provide novel and not previously seen recommendations. It is positively correlated with
                                              user satisfaction and helps to achieve strong recommendation performance [3–5, 8].
                                                However, one limitation with unexpectedness is that it is defined in the feature space of previously
                                              consumed items, which relies completely on explicit purchase information, and may not work well
                                              ACM RecSys 2019 Late-breaking Results, 16th-20th September 2019, Copenhagen, Denmark
                                              Copyright ©2019 for this paper by its authors. Use permitted under Creative Commons License Attribution 4.0 International
                                              (CC BY 4.0).
Latent Modeling of Unexpectedness for Recommendations          ACM RecSys 2019 Late-breaking Results, 16th-20th September 2019, Copenhagen, Denmark

                                                        in the case when purchase actions are sparse or noisy. Besides, it does not include specific user and
                                                        item features, neither does it address the latent and complex relationships between users and items.
                                                        Furthermore, the distance metric between discrete items is hard to define in the feature space, while
      Evaluation Metrics                                in the latent space we have well-defined euclidean distances. Thus, we cannot model the unexpected
                                                        relations efficiently in the feature space.
RMSE & MAE                                                 Therefore, though researchers have successfully developed algorithms to improve unexpectedness
Root Mean Square Error and Mean Ab-                     and other novelty metrics, they always come at a price of losing accuracy performance, as pointed out
solute Error measures the differences                   in [19]. This severely limits the practical use of the unexpected recommendation, especially in business
between estimated ratings and actual                    applications where the goal is to increase commercial sales and enhance user satisfaction. Thus, it is
ratings.                                                important to design a novel unexpected recommender to increase novelty of recommendations while
Precision@N & Recall@N                                  keeping the same level of accuracy performance.
Precision is the fraction of the rec-                      In this paper, we propose to define unexpectedness as the distance metric in the latent space (latent
ommended items that are of high                         feature and attribute embeddings) rather than feature space (explicit users and items) by utilizing
utility to the user. Recall is the fraction             Heterogeneous Information Network Embeddings (HINE) [15, 16] to capture the latent, complex
of the high utility items that are                      and heterogeneous relations between them. We take the natural closure of convex hull in the latent
recommended to the user.                                space, and calculate unexpectedness as the euclidean distance from new item embedding to the latent
Unexpectedness                                          convex hull of expected items of the user. The proposed unexpectedness measure is subsequently
Unexpectedness is calculated through                    combined with estimated ratings for providing recommendations.
equation (3) following our proposed                        We make the following contributions in this paper. We propose to model unexpectedness in the
definition.                                             latent space by defining the expected set for each user as a latent convex hull generated by item
Serendipity                                             embeddings. We also conduct extensive experiments on two real-world datasets and show that our
Serendipity [8] is computed by                          model significantly improves novelty measures without losing any accuracy metrics.
                   RS &P M &U se f ul
Serendipity =             RS          where
                                                        METHOD
RS stands for the recommended items
using the target model, PM stands                       In prior literature [4], unexpectedness is defined as the distance of recommended item from the
for the recommendation items using                      closure set of expected items, the latter including items that either previously consumed by the user or
a primitive prediction algorithm and                    closely related to the consumptions. However, this definition does not include the feature information
USEFUL stands for the items whose                       for users and items, neither does it consider the latent and heterogeneous interactions between them.
utility is above certain threshold.                     Thus, it might not serve as a perfect definition, for we cannot manually codify all explicit associations
Diversity                                               and relations between users and items. Also, it is hard to mathematically formalize the expected set
Diversity is computed as the                            in the feature space, where the distribution of users and items is discrete and unstructured.
averageÍ intra-list       distance [17]:                   On the other hand, heterogeneous information network [16] (HIN) along with latent embedding
D(R) = i ∈R j,i ∈R d(i, j)
               Í                                        method [7] provides us with an efficient tool to model users, items and their associated features
Coverage                                                simultaneously by linking the associated features with corresponding entities in the network and
Coverage is computed as the percent-                    subsequently map them into the latent space. Thus in the paper, we propose to combine those two
age of distinctive recommended items                    techniques for latent modeling of unexpectedness.
over all distinctive items in the dataset.
Latent Modeling of Unexpectedness for Recommendations           ACM RecSys 2019 Late-breaking Results, 16th-20th September 2019, Copenhagen, Denmark

                                                          We denote the heterogeneous network as G = (V , E,T ), in which each node v and each link e
      Baseline Models
                                                        are assigned with specific type Tv and Te . To effectively learn node representations, we enable the
                                                        skip-gram mechanism to maximize the probability of each context node c t within the neighbors of v,
SPR [9]
                                                        denoted as Nt (v), where we add the subscript t (t ∈ Tv ) to limit the node to a specific type:
Serendipitous Personalized Ranking ex-
tends traditional personalized ranking
                                                                                                Õ Õ Õ
                                                                                      arд max θ                    loдP(c t |v; θ )                       (1)
methods by considering item popular-                                                                 v ∈V t ∈Tv c t ∈N t (v)
ity in AUC optimization, which makes
                                                        To calculate P(c t |v; θ ), we propose to use heterogeneous random walk to generate meta-paths of
the ranking sensitive to the popularity                                                               R1        R2
of negative examples.                                   multiple types of nodes in the form of V1 −−→ V2 −−→ V3 · · · Vn wherein R = R 1 ◦ R 2 ◦ · · · Rn defines the
Auralist [18]                                           composite relations between the start and the end of the heterogeneous random walk. The transition
Auralist is a personalized recommenda-                  probability within each random walk between two nodes is defined as follows:
tion system that balances between the                                                                 ( C(T ,T )
                                                                                                         |N t +1 (Vt )| , (Vt , Vt +1 ) ∈ E
                                                                                                           Vt Vt +1
desired goals of accuracy and novelty                                                 p(Vt +1 |Vt ) =                                                              (2)
simultaneously.                                                                                        0,                 (Vt , Vt +1 ) < E
HOM-LIN [4]
                                                        where C(TVt ,TVt +1 ) stands for the transition coefficient between the type of node Vt and the type
HOM-LIN is the state-of-the-art unex-
                                                        of node Vt +1 . |Nt +1 (Vt )| stands for the number of nodes of type Vt +1 in the neighborhood of Vt . We
pected algorithm that utilizes the com-
                                                        apply heterogeneous random walk iteratively to each node and generate the collection of meta-path
bination of estimated ratings and unex-
                                                        sequences for the skip-gram mechanism.
pectedness for recommendation.
                                                           Note that in previous definition [4], the idea of ‘’closure set” plays an important role. As a natural
DPP [6]
                                                        closure in the latent space, latent convex hull of each user is the smallest convex set that contains all
The Determinantal Point Process uti-
                                                        the item embeddings that the user has visited. Under this setting, we assume that the expected set
lizes a fast greedy MAP inference ap-
                                                        maintains its convexity in the growing process. For example, if a person enjoy rap music and rock &
proach to generate relevant and diverse
                                                        roll, she/he shall also have certain expectations on rap-rock music, an innovation combination of those
recommendations.
                                                        two genres. Subsequently, we define the unexpectedness as the distance between the embedding
Random
                                                        of new item and the latent convex hull generated from the embeddings of all visited items
Random is the baseline model where
                                                        of the user:
we randomly recommend items to
users without considering any informa-                                                              U nexpu,i = d(i; LC(Ni ))                                  (3)
tion about the ratings, unexpectedness                  where Ni = (i 1 , i 2 , · · · , i n ) contains the embeddings of items that link to the user in the heterogeneous
and utility.                                            information network. Specifically, the distance function is well defined as the minimal distance from
                                                        the given point to all the boundaries of the latent closure. We visualize this definition in Figure 1.
                                                           Once we set up the definition of unexpectedness, we perform the unexpected recommendation
                                                        using hybrid-utility based collaborative filtering methods that linearly combine estimated ratings
                                                        (representing accuracy) with unexpectedness (representing novelty)
                                                                                    U tilityu,i = (1 − α) ∗ EstRatinдu,i + α ∗ U nexpu,i                              (4)
Latent Modeling of Unexpectedness for Recommendations                    ACM RecSys 2019 Late-breaking Results, 16th-20th September 2019, Copenhagen, Denmark

                                                                 The key idea lies in that, instead of recommending the similar items that the users are very familiar with
      Dataset     Yelp        TA                                 as the classical recommenders do, we recommend unexpected and useful items to the users that they
      # Review    5,996,996 878,561                              might have not thought about, but indeed fit well to their satisfactions. These two adversarial forces of
      # Business 188,593      576,689                            accuracy and novelty work together to obtain the optimal solution, thus improving recommendation
      # User      1,518,169 3,945                                performance and user satisfaction. In real practice, the unexpected coefficient α is optimized through
      # Sparsity 0.002%       0.039%                             Bayesian optimization.
Table 1: Descriptive Statistics for the two
Datasets.                                                        EXPERIMENTS AND RESULTS
                                                                 We implement our model on two real-world datasets: the Yelp Challenge Dataset Round 12 [2] and
                                                                 the TripAdvisor Dataset [1]. The descriptive statistics of these two datasets are listed in Table 1. To
 Data        Model         Unexp       Ser      Div      Cov
              FM           0.0326    0.0978    0.0135   0.5369   address the cold-start issue, we filter out users and items that appear less than 5 times in the dataset.
           FM+Unexp       0.1122*   0.4793*   0.3798*   0.5904   We measure the recommendation results using accuracy and novelty metrics listed in page 2.
           CoCluster       0.0338    0.4595    0.3106   0.5811
        CoCluster+Unexp   0.1400*   0.4660*   0.3269*   0.5904      We compare the recommendation performance with and without considering the proposed unex-
             SVD           0.0457    0.1352    0.0479   0.5221   pectedness by 5-fold cross-validation experiments using five popular collaborative filtering algorithms
          SVD+Unexp       0.0620*   0.1469*   0.0524*   0.5406
             NMF           0.0333    0.4954    0.3268   0.5867   including KNN, SVD, CoClustering, NMF and FM [13]. We also report results for baseline unexpected
 Yelp     NMF+Unexp       0.1390*   0.5469*   0.3430*   0.5904   recommendation models. As shown in Table 2 and 3, when considering unexpectedness in the rec-
             KNN           0.0448    0.0977    0.0129   0.5369
          KNN+Unexp       0.0610*   0.1165*   0.0259*   0.5406   ommendation process, we obtain significant amount of improvements in terms of Unexpectedness,
              SPR          0.0668    0.3720    0.2532   0.5697   Serendipity and Diversity measures, without sacrificing any accuracy performance in RMSE, MAE,
            Auralist       0.0663    0.3637    0.2047   0.5457
           HOM-LIN         0.0751    0.4329    0.3011   0.5365   Precision and Recall. Therefore, the proposed definition of unexpectedness enables us to provide
             DPP           0.0670    0.4488    0.2488   0.5904   unexpected and useful recommendations at the same time. It is crucial for successful deployments of
            Random         0.1733    0.4848    0.3763   0.5457
              FM           0.0222    0.3979    0.0017   0.1798   unexpected recommendation models in industrial applications. In addition, the proposed approach
           FM+Unexp       0.0643*   0.4631*   0.0493*   0.1798   outperforms all previous unexpected recommendation models introduced in page 3, which validates
           CoCluster       0.0234    0.3973    0.0015   0.1855
        CoCluster+Unexp   0.0652*   0.4619*   0.0471*   0.1807   the superiority of modeling unexpectedness in the latent space over feature space. Finally, we point out
             SVD           0.0231    0.3967    0.0006   0.1798   that the proposed model achieves greater improvements on Yelp dataset, which contains richer feature
          SVD+Unexp       0.0644*   0.4621*   0.0499*   0.1798
             NMF           0.0227    0.3979    0.0010   0.1798   information compared to the TripAdvisor dataset. This observation is in line with our assumption
  TA      NMF+Unexp       0.0644*   0.4627*   0.0499*   0.1798   that feature information matters in the definition of unexpectedness.
             KNN           0.0233    0.3979    0.0019   0.1798
          KNN+Unexp       0.0643*   0.4631*   0.0492*   0.1798
              SPR          0.0474    0.3593    0.0375   0.1834   CONCLUSION
            Auralist       0.0473    0.3462    0.0355   0.1834
           HOM-LIN         0.0572    0.3729    0.0411   0.1807   In this paper, we propose to model unexpectedness and user expectations in the latent space, which
             DPP           0.0464    0.3245    0.0311   0.1807   makes it possible to capture the latent, complex and heterogeneous relations between users and items,
            Random         0.0833    0.4468    0.0650   0.1834
                                                                 thus significantly improving the performance and practicability of unexpected recommendations.
Table 2: Comparison of recommenda-
                                                                 We empirically demonstrate that the proposed approach consistently and significantly outperforms
tion performance in novelty measures, ”*”
stands for 95% statistical significance                          baseline models in terms of unexpected measures without sacrificing the performance of accuracy.
                                                                    As the future work, we plan to conduct live experiments with real business environment in order to
                                                                 further evaluate the effectiveness of unexpected recommendations and analyze both qualitative and
                                                                 quantitative aspects in an online retail setting, especially with the utilization of A/B test.
Latent Modeling of Unexpectedness for Recommendations                  ACM RecSys 2019 Late-breaking Results, 16th-20th September 2019, Copenhagen, Denmark

                                                              REFERENCES
                                                               [1] 2019. TripAdvisor Dataset. http://www.cs.cmu.edu/~jiweil/html/hotel-review.html.
                                                               [2] 2019. Yelp Challenge Dataset. https://www.yelp.com/dataset/challenge.
                                                               [3] Panagiotis Adamopoulos. 2014. On discovering non-obvious recommendations: Using unexpectedness and neighborhood
                                                                   selection methods in collaborative filtering systems. In Proceedings of the 7th WSDM.
 Data        Model        RMSE      MAE     Pre@5    Rec@5     [4] Panagiotis Adamopoulos and Alexander Tuzhilin. 2015. On unexpectedness in recommender systems: Or how to better
              FM          0.9197   0.6815   0.7699   0.6123
                                                                   expect the unexpected. ACM Transactions on Intelligent Systems and Technology (TIST) (2015).
           FM+Unexp       0.9178   0.6820   0.7700   0.6123
           CoCluster      0.9499   0.7140   0.7255   0.5913    [5] Li Chen, Yonghua Yang, Ningxia Wang, Keping Yang, and Quan Yuan. 2019. How Serendipity Improves User Satisfaction
        CoCluster+Unexp   0.9504   0.7138   0.7196   0.5864        with Recommendations? A Large-Scale User Evaluation. In The World Wide Web Conference. ACM, 240–250.
             SVD          0.9132   0.7071   0.7692   0.5944    [6] Laming Chen, Guoxin Zhang, and Eric Zhou. 2018. Fast greedy MAP inference for Determinantal Point Process to improve
          SVD+Unexp       0.9134   0.7076   0.7701   0.5975
                                                                   recommendation diversity. In Advances in Neural Information Processing Systems. 5622–5633.
             NMF          0.9533   0.7181   0.7197   0.5318
 Yelp     NMF+Unexp       0.9522   0.7154   0.7222   0.5833    [7] Yuxiao Dong, Nitesh V Chawla, and Ananthram Swami. 2017. metapath2vec: Scalable representation learning for
             KNN          0.9123   0.7048   0.7687   0.6085        heterogeneous networks. In Proceedings of the 23rd ACM SIGKDD. ACM.
          KNN+Unexp       0.9128   0.7051   0.7659   0.6073    [8] Mouzhi Ge, Carla Delgado-Battenfeld, and Dietmar Jannach. 2010. Beyond accuracy: evaluating recommender systems
              SPR         1.0351   0.7729   0.7692   0.6188
                                                                   by coverage and serendipity. In Proceedings of the fourth ACM conference on Recommender systems. ACM, 257–260.
            Auralist      1.0377   0.7799   0.7678   0.6000
           HOM-LIN        0.9609   0.7447   0.7621   0.6150    [9] Qiuxia Lu, Tianqi Chen, Weinan Zhang, Diyi Yang, and Yong Yu. 2012. Serendipitous personalized ranking for top-n
             DPP          1.0288   0.7702   0.7598   0.6012        recommendation. In 2012 WI-IAT. IEEE.
            Random        1.4959   1.2456   0.4250   0.3333   [10] Tomoko Murakami, Koichiro Mori, and Ryohei Orihara. 2007. Metrics for evaluating the serendipity of recommendation
              FM          1.1105   0.8340   0.6768   0.9590        lists. In Annual conference of the Japanese society for artificial intelligence. Springer, 40–46.
           FM+Unexp       1.1275   0.8445   0.7040   0.9656
           CoCluster      1.0178   0.7643   0.6845   0.9732   [11] Tien T Nguyen, Pik-Mai Hui, F Maxwell Harper, Loren Terveen, and Joseph A Konstan. 2014. Exploring the filter bubble:
        CoCluster+Unexp   1.0285   0.7541   0.6865   0.9703        the effect of using recommender systems on content diversity. In Proceedings of the 23rd WWW. ACM.
             SVD          0.9868   0.7533   0.7210   0.9465   [12] Eli Pariser. 2011. The filter bubble: How the new personalized web is changing what we read and how we think. Penguin.
          SVD+Unexp       0.9937   0.7517   0.7085   0.9594   [13] Francesco Ricci, Lior Rokach, and Bracha Shapira. 2015. Recommender systems: introduction and challenges. In
             NMF          1.0241   0.7709   0.6850   0.9681
 TA       NMF+Unexp       1.0262   0.7533   0.6881   0.9775        Recommender systems handbook. Springer, 1–34.
             KNN          0.9940   0.7531   0.6969   0.9689   [14] Guy Shani and Asela Gunawardana. 2011. Evaluating recommendation systems. In Recommender systems handbook.
          KNN+Unexp       1.0001   0.7483   0.6907   0.9763   [15] Chuan Shi, Binbin Hu, Xin Zhao, and Philip Yu. 2018. Heterogeneous Information Network Embedding for Recommenda-
              SPR         1.0328   0.8008   0.6395   0.9325        tion. IEEE Transactions on Knowledge and Data Engineering (2018).
            Auralist      1.0318   0.7997   0.6460   0.9390
           HOM-LIN        1.0298   0.7902   0.6420   0.9418   [16] Yizhou Sun and Jiawei Han. 2013. Mining heterogeneous information networks: a structural analysis approach. Acm
             DPP          1.0304   0.8158   0.6264   0.9303        Sigkdd Explorations Newsletter 14, 2 (2013), 20–28.
            Random        1.6857   1.3100   0.3238   0.2500   [17] Mi Zhang and Neil Hurley. 2008. Avoiding monotony: improving the diversity of recommendation lists. In Proceedings of
Table 3: Comparison of recommendation                              the 2008 ACM conference on Recommender systems. ACM, 123–130.
performance in accuracy measures, ”*”                         [18] Yuan Cao Zhang, Diarmuid Ó Séaghdha, Daniele Quercia, and Tamas Jambor. 2012. Auralist: introducing serendipity into
stands for 95% statistical significance                            music recommendation. In Proceedings of the fifth ACM WSDM. ACM.
                                                              [19] Zainab Zolaktaf, Reza Babanezhad, and Rachel Pottinger. 2018. A Generic Top-N Recommendation Framework For
                                                                   Trading-off Accuracy, Novelty, and Coverage. In 2018 IEEE 34th ICDE. IEEE.