=Paper= {{Paper |id=Vol-2068/wii5 |storemode=property |title=Improved Listwise Collaborative Filtering with High-Rating-Based Similarity and Temporary Ratings |pdfUrl=https://ceur-ws.org/Vol-2068/wii5.pdf |volume=Vol-2068 |authors=Yoshiki Tsuchiya,Hajime Nobuhara |dblpUrl=https://dblp.org/rec/conf/iui/TsuchiyaN18 }} ==Improved Listwise Collaborative Filtering with High-Rating-Based Similarity and Temporary Ratings== https://ceur-ws.org/Vol-2068/wii5.pdf
        Improved Listwise Collaborative Filtering with High-
          Rating-Based Similarity and Temporary Ratings
                                Yoshiki Tsuchiya                            Hajime Nobuhara
                              University of Tsukuba                        University of Tsukuba
                                  Tsukuba, Japan                              Tsukuba, Japan
                          tsuchiya@cmu.iit.tsukuba.ac.jp                 nobuhara@iit.tsukuba.ac.jp
ABSTRACT                                                                [7,10] and listwise ranking-oriented [2]. Pairwise ranking-
In this paper, we make two proposals to speed up listwise               oriented CF predicts the order of pairs of items but requires
collaborative filtering and improve its accuracy. The first is          large computation time.
to speed up computation by only using a subset of the rating
information (the high ratings). The second is to improve
accuracy using temporary ratings that estimate the rating
scores that neighboring users are not rating. Experiments
using MovieLens datasets (1M and 10M) demonstrate that
these proposals effectively reduce computation time about
1/50 and improve accuracy 2.22% compared with ListCF, a
well-known listwise collaborative filtering algorithm.
Author Keywords
Recommender        system;     Ranking-oriented       collaborative
filtering;
                                                                                Figure 1. Collaborative filtering classification
ACM Classification Keywords
• Information systems, Collaborative filtering                          In contrast, listwise ranking-oriented CF predicts the order
INTRODUCTION                                                            of the complete list of items. Although this produces better
In recent years, due to the development of the Web,                     accuracy than a typical pairwise CF algorithm, calculating
recommender systems have become increasingly important                  the required similarities is time-consuming and there is
in various situations; many researchers are now focusing on             room to improve the ranking accuracy. In this paper, we
recommendation technologies and systems [2,6,7,9,10].                   propose an efficient listwise ranking-oriented CF algorithm
Collaborative filtering (CF) is a widely used                           that is both faster and has higher ranking accuracy.
recommendation algorithm that is based on the similarity                The proposed method implements two improvements. First,
between users or items, as calculated from a user and rating            when calculating the similarity between users, it only
matrix. Various CF algorithms have been proposed, and                   considers the highest-rated items, greatly speeding up the
they can be divided into two types: rating-oriented [6,9] and           calculation. Second, it introduces temporary ratings when
ranking-oriented [2,7,10], as shown in Fig. 1. Rating-                  making ranking predictions. Experimental comparisons
oriented CF algorithms, such as item-based CF [9], predict              using MovieLens 1M (6,040 users, 3,952 movies,
the ratings of items that have not been evaluated by users              1,000,209 ratings) and 10M (71,567 users, 10,681 movies,
and make recommendations by calculating the similarity                  10,000,054 ratings) confirm that the proposed method
between users or items. On the contrary, ranking-oriented               reduces about 1/50 computation time of similarity and
CF uses user similarity to predict the item ranking and                 improves 2.22% ranking accuracy than a conventional CF
recommends items based on this. We will focus on this                   algorithm.
method due to its performance. Ranking-oriented CF can be
further divided into two types: pairwise ranking-oriented               RELATED WORK
                                                                        Overview of ListCF
© 2018. Copyright for the individual papers remains with the authors.
Copying permitted for private and academic purposes.
                                                                        In this section, we give an overview of ListCF [2], a well-
WII'18, March 11, Tokyo, Japan.                                         known ranking-oriented listwise CF algorithm. ListCF
                                                                        proceeds in two phases, first calculating the similarities
                                                                        between users and then predicting ranks for the target user’s
                                                                        unrated items. The first phase is based on a probability
                                                                        distribution of item permutations, calculated by combining
                                                                        the Plackett–Luce [8] and top-k probability [1] models and
                                                                        finding each user’s neighboring users.
Similarity calculation
In ListCF, the similarity of a pair of users 𝑢 and 𝑣 is
calculated based on a probability distribution of item
permutations for each user, calculated using the Plackett–
Luce model [8], which is a representative permutation
probability model. The flow of similarity calculation is
shown on the left of Fig. 2 and Fig. 3. Let the set 𝐼 =
{𝑖1 , 𝑖2 , ⋯ , 𝑖𝑛 } of items 𝜋 𝑖 = (𝜋1𝑖 , 𝜋2𝑖 , ⋯ , 𝜋𝑛𝑖 ) (∈ 𝐼 𝑛 ) be an
ordered list where 𝜋𝑗𝑖 ∈ 𝐼 and 𝜋𝑗𝑖 ≠ 𝜋𝑘𝑖 if 𝑗 ≠ 𝑘 , and let
Ω𝐼 (⊂ 𝐼 𝑛 ) be the set of all possible permutations of 𝐼. Given
the item ratings (𝑟𝜋𝑖 , 𝑟𝜋𝑖 , ⋯ , 𝑟𝜋𝑛𝑖 ) (e.g., on a real interval
                          1  2
in the case of MovieLens [1,5]), where 𝑟𝜋𝑖 is the rating                      Figure 3. Procedure for similarity calculation of conventional
                                                      𝑗                                                  ListCF.
score of 𝜋𝑗𝑖 , the probability of 𝜋 𝑖 , 𝑃(𝜋 𝑖 ), is defined using an
                                                                               by users 𝑢 and 𝑣, and 𝑃𝑢 and 𝑃𝑣 (∈ [0,1]) as the probability
increasing and strictly positive function 𝛷(∙) ≥ 0 as                                             𝐼
follows:                                                                      distributions over 𝑔𝑘𝑢,𝑣 (⊂ Ω𝐼𝑢,𝑣 ) calculated by Eq. (3) using
                                                                              the users’ rating scores. The similarity score is now
                       𝑛      𝛷 (𝑟𝜋𝑖 )                                        obtained from the Kullback–Leibler (KL) divergence [5]
                 𝑖)                   𝑗
           𝑃(𝜋        =∏                     (∈ [0,1]),                 (1)   calculated from 𝑃𝑢 and 𝑃𝑣 . Given a pair of users 𝑢 and 𝑣,
                           𝑛
                      𝑗=1 ∑𝑘=𝑗 𝛷 (𝑟𝜋𝑘
                                    𝑖)                                        the KL divergence of 𝑃𝑢 and 𝑃𝑣 is defined as
where the function is defined as 𝛷(𝑟) = 𝑒 𝑟 . However, this                                                              𝑃𝑢 (𝑔)
                                                                                   𝐷𝐾𝐿 (𝑃𝑢 ∥ 𝑃𝑣 ) = ∑ 𝑃𝑢 (𝑔) 𝑙𝑜𝑔2 (             ).       (4)
requires us to consider 𝑛! different permutations of the 𝑛                                                               𝑃𝑣 (𝑔)
                                                                                                       𝐼
items, which would take a long time to compute. To speed                                            𝑔∈𝑔𝑘𝑢,𝑣
up the computation, the top-k probability model 𝑔𝑘 [1] is
                                                                              The KL divergence is asymmetric, but ListCF defines the
introduced as follows:
                                                                              similarity 𝑠(𝑢, 𝑣)(∈ (−∞, 1 ]) of users 𝑢 and 𝑣 to be
                        𝑔𝑘 (𝑖1 , 𝑖2 , ⋯ , 𝑖𝑘 ) =                              symmetric as follows:
                                                              𝑛!                                 1
    {𝜋 𝑙 |𝜋 𝑙 ∈ Ω𝐼 , 𝜋𝑗𝑙 = 𝑖𝑗 , 𝑗 = 1, ⋯ , 𝑘, 𝑙 = 1, ⋯ ,            }               𝑠(𝑢, 𝑣) = 1 − [𝐷𝐾𝐿 (𝑃𝑢 ∥ 𝑃𝑣 ) + 𝐷𝐾𝐿 (𝑃𝑣 ∥ 𝑃𝑢 )].     (5)
                                                           (𝑛 − 𝑘)!                              2
                               (⊂ Ω𝐼 ),                                 (2)   If the set 𝐼𝑢,𝑣 only includes a few items, the similarity will
                                                                              be high, so this is relaxed by multiplying by the similarity
and the probability of the top-k permutation is calculated as
                                                                              function by min{𝐼𝑢,𝑣 ⁄𝑐 , 1}, where 𝑐 is a threshold. Each
                                  𝑘
                                          𝛷 (𝑟𝜋𝑗 )                            user’s neighboring users can then be found from the
      𝑃(𝑔𝑘 (𝑖1 , 𝑖2 , ⋯ 𝑖𝑘 )) = ∏                    (∈ [0,1]),               similarities calculated using Eqs. (3)–(5).
                                    ∑𝑛 𝛷(𝑟𝜋𝑙 )
                                 𝑗=1 𝑙=𝑗
                                                                              Ranking prediction
                      ∀𝑗 = 1, ⋯ , 𝑘 ∶ 𝜋𝑗 = 𝑖𝑗 .                         (3)   The flow of ranking prediction is shown on the left of Fig. 4.
                                                                              Let U be the set of users, 𝑁𝑢 (⊂ 𝑈) be the set of the user 𝑢’s
Let 𝑔𝑘𝐼 (⊂ Ω𝐼 ) be the set of top-k permutations of 𝐼, and let                neighbors and 𝑇𝑢 (⊂ 𝐼 ∖ 𝐼𝑢 ) be the set of items whose ranks
the probabilities of these permutations form the probability                  are to be predicted. Let 𝑃̂𝑢 (∈ [0,1]) be the probability
distribution. Then, define 𝐼𝑢,𝑣 (⊂ 𝐼) as the set of items rated                                                            𝑇
                                                                              distribution of the top-k permutations 𝑔𝑘𝑢 (⊂ Ω𝑇𝑢 ) of 𝑇𝑢 ,
                                                                              written as
                                                                                                                 𝜑𝑢,𝑔
                                                                                                  𝑃̂𝑢 (𝑔) =                ,             (6)
                                                                                                            ∑𝑔′ ∈𝑔𝑇𝑢 𝜑𝑢,𝑔′
                                                                                                                 𝑘

                                                                                                 𝑇
                                                                              where {𝜑𝑢,𝑔 |∀𝑔 ∈ 𝑔𝑘𝑢 } are unknown variables assigned to
                                                                              the top-k permutations. In ListCF, the cross entropy is used
                                                                              as a loss function for prediction. Consider a target user 𝑢,
                                                                              for whom you want to rank a set of items 𝑇𝑢 , and a
                                                                              neighboring user 𝑣 ∈ 𝑁𝑢 . Let the set of items rated by 𝑣 be
                                                                                                               𝑇
                                                                              𝐼𝑣 , with 𝑇𝑢,𝑣 = 𝑇𝑢 ∩ 𝐼𝑣 , and 𝑔𝑘𝑢,𝑣 (⊂ Ω𝑇𝑢,𝑣 ) be the set of
  Figure 2. Similarity calculation: conventional ListCF (left),               top-k permutations of 𝑇𝑢,𝑣 . The cross entropy is calculated
                   proposed method (right).
                                                                              Figure 5. Procedure of similarity calculation of proposed
                                                                                                      method.
                                                                           propose a method of calculating user similarity by focusing
  Figure 4. Ranking prediction: conventional ListCF (left),                only on high ratings. The flow of similarity calculation is
                 proposed method (right).                                  shown on the right of Fig. 2 and Fig.5. Let 𝐻𝑢 (⊂ 𝐼𝑢 ) be the
                                                    𝑇                      set of items that were rated highly by the target user 𝑢, and
using probability distributions 𝑃̂ ′𝑢 and 𝑃′𝑣 over 𝑔𝑘𝑢,𝑣 as                𝐻𝑢,𝑣 (⊂ 𝐼𝑢,𝑣 ) be the set of items that both 𝑢 and 𝑣 rated
follows:                                                                   highly. Given the user and rating matrix, we define the
                                                                           similarity between 𝑢 and 𝑣 as follows:
            𝐸(𝑃̂ ′𝑢 , 𝑃′𝑣 ) = − ∑ 𝑃′ 𝑣 (𝑔) log 2 𝑃̂ ′𝑢 (𝑔) .         (7)
                                    𝑇
                                 𝑔∈𝑔𝑘 𝑢,𝑣                                                                 |𝐻𝑢,𝑣 |
                                                                                    𝑠𝑖𝑚𝑖𝑙𝑎𝑟𝑖𝑡𝑦(𝑢, 𝑣) =            (∈ [0,1]),         (12)
                                                                                                           |𝐻𝑢 |
ListCF makes predictions by minimizing the following
cross entropy weighted sum:                                                where 𝑠𝑖𝑚𝑖𝑙𝑎𝑟𝑖𝑡𝑦(𝑢, 𝑣) is asymmetric. We define
                                                                           𝑠𝑖𝑚𝑖𝑙𝑎𝑟𝑖𝑡𝑦(𝑢, 𝑣) = 0 if 𝑢 = 𝑣 not to add him/herself to
                arg min ∑ 𝑠(𝑢, 𝑣) ∙ 𝐸(𝑃̂ ′𝑢 , 𝑃′ 𝑣 ),                (8)   neighborhood users when predicting each user’s ranking. In
                     𝜑𝑢
                          𝑣∈𝑁𝑢                                             conventional ListCF, based on the top-k probability model,
                                   𝑇                                       the time complexity is 𝑛!⁄(𝑛 − 𝑘)! , which is smallest
                    s. t ∀𝑔 ∈ 𝑔𝑘𝑢 ∶ 𝜑𝑢,𝑔 ≥ 0.                              (i.e., 𝑛 ) when 𝑘 = 1 . As the proposed method’s time
The objective function 𝐹(𝜑𝑢 ) in Eq. (8) is then transformed               complexity is 𝑛, it is never worse than that of conventional
as follows, using Eqs. (5) and (7):                                        ListCF, and often better. Changing how the high ratings are
                                                                           determined changes the ranking accuracy, as explained in
                      𝐹(𝜑𝑢 ) = ∑ 𝑠(𝑢, 𝑣) ∙                                 the experimental section.
                                   𝑣∈𝑁𝑢                                    Improving ranking accuracy using temporary ratings
                                                                           We also propose to improve the ListCF’s ranking accuracy
                                                                           by introducing temporary ratings. In ListCF, the cross
   ∑ 𝑃′ 𝑣 (𝑔) [log 2 ( ∑                 𝜑𝑢,𝑔′ ) − log 2 (𝜑𝑢,𝑔 )]    (9)   entropy in Eq. (7) is calculated from the probability
    𝑇                            𝑇                                                            𝑇
 𝑔∈𝑔𝑘 𝑢,𝑣                   𝑔′ ∈𝑔𝑘 𝑢,𝑣                                     distribution of 𝑔𝑘𝑢,𝑣 , the set of permutations of 𝑇𝑢,𝑣 .
                                                                           However, optimizing the objective function may fail if 𝑇𝑢,𝑣
Equation (9) is optimized by the gradient descent method.
Partially differentiating F with respect to 𝜑𝑢,𝑔 , we obtain               has too few elements, e.g., if it only includes one element
                                                                           and that item received a low rating from the neighboring
                                  𝜕𝐹                                       user 𝑣 . In this case, despite the item’s low rating,
                                       =                                   optimization generates a high item rank, which is unhelpful
                                 𝜕𝜑𝑢,𝑔
                                                                           for user 𝑢. Fig. 6 shows how the probability distribution for
                 𝑠(𝑢, 𝑣)           ∑𝑣∈𝑁𝑢 𝑠(𝑢, 𝑣)𝑃′ 𝑣 (𝑔)                   target user 𝑢’s unrated items before optimization (upper)
   ∑                             −                       .          (10)
         ln 2 ∙ ∑     𝑇𝑢,𝑣 𝜑𝑢,𝑔′
                  𝑔′∈𝑔𝑘
                                       ln 2 ∙ 𝜑𝑢,𝑔                         and neighboring user 𝑣’s distribution (middle) give a graph
  𝑣∈𝑁𝑢
                                                                           like the one on the lower. Item 1 was given a low rating by
For a given learning rate 𝜂, 𝜑𝑢,𝑔 is updated as follows:                   𝑣 but, since no other items were rated, it is guaranteed to
                                                                           top the rankings. As a result, 𝑢’s probability distribution is
                                             𝜕𝐹
                    𝜑𝑢,𝑔 ← 𝜑𝑢,𝑔 − 𝜂               .                 (11)   updated to place item 1 at a higher position. The proposed
                                            𝜕𝜑𝑢,𝑔                          method therefore makes ranking predictions after giving
PROPOSED METHOD                                                            temporary estimated ratings to items that were not rated by
                                                                           the neighboring user 𝑣. The flow of ranking prediction is
Rapid similarity calculation using only high ratings                       shown on the right of Fig 4 and Fig.7. Let 𝑟𝑢,𝑖 be the rating
Conventional ListCF is slow because it has to calculate                    given by user 𝑢 to item 𝑖, and let unrated items have a score
similarities using all rating scores. To speed up computation,             of zero.
we now
                                                                      𝑠(𝑢, 𝑣) in Eq. (8) with 𝑠𝑖𝑚𝑖𝑙𝑎𝑟𝑖𝑡𝑦(𝑢, 𝑣) (Eq. (12)), ranking
                                                                      predictions can then be made in the same way as for ListCF
                                                                      by using Eqs. (8)–(11). Calculating temporary ratings
                                                                      allows the neighbors’ probability distributions in situations
                                                                      such as Fig. 6 to be more reasonable. As a result, since the
                                                                      parameters are less frequently updated in an undesirable
                                                                      way, this should improve the ranking accuracy.
                                                                      EXPERIMENTAL EVALUATION
                                                                      Overview of the experiment
                                                                      To confirm the effectiveness of the two proposed changes,
                                                                      we experimentally compared the methods in terms of
                                                                      computation time and ranking accuracy using the
                                                                      MovieLens 1M and 10M datasets, details of which are
                                                                      shown in Table 1. We selected 10 sets of ratings from each
                                                                      user’s rating information to create the test dataset and used
                                                                      the rest as training data. In this experiment, since it is
                                                                      shown that the ranking accuracy does not improve greatly
                                                                      by increasing the value of k [2], we compared the proposed
                                                                      method with conventional ListCF based on the top-1
  Figure 6. Examples of unsuitable parameter updates that             probability model, as in [3,4]. Also, the threshold 𝑐 of the
 occur when users have only one low-rated item in common.
      The horizontal axis represents the item number.
                                                                      similarity function is 300. The normalized discounted
                                                                      cumulative gain (NDCG) metric was used to assess ranking
                                                                      prediction accuracy. The NDCG value, considering the top
                                                                      𝑛 predicted rankings for user 𝑢, is defined as follows:
                                                                                                      𝑛      𝑝
                                                                                                      2𝑟𝑢 − 1
                                                                                    𝑁𝐷𝐶𝐺𝑢 @𝑛 = 𝑍𝑢 ∑               ,            (14)
                                                                                                    log 2 (1 + 𝑝)
                                                                                                     𝑝=1

                                                                      where 𝑍𝑢 is a normalization term that ensures the NDCG
                                                                                                             𝑝
                                                                      value of the correct ranking is 1 and 𝑟𝑢 is the rating of the
                                                                      pth-ranked item for user 𝑢. For a set of users 𝑈, the overall
                                                                      NDCG@n score is calculated as follows:
                                                                                                     |𝑈|
                                                                                               1
                                                                                     𝑁𝐷𝐶𝐺@𝑛 =     ∑ 𝑁𝐷𝐶𝐺𝑢 @𝑛 .                 (15)
                                                                                              |𝑈|
                                                                                                    𝑢=1

                                                                      Comparison of similarity computation time
    Figure 7. Procedure for calculating temporary ratings.
                                                                      In this experiment, we measured the similarity computation
For a given neighborhood user 𝑣 ∈ 𝑁𝑢 and set 𝑇𝑢 =                     time. We compared the time taken to calculate the user
{𝑡1, 𝑡2 , ⋯ , 𝑡𝑝 } (⊂ 𝐼 ∖ 𝐼𝑢 ) of items not rated by 𝑢 , 𝑣 ’s         similarities using both the proposed and ListCF methods,
temporary ratings are defined as follows:                             and then compared the ranking accuracy of ListCF
                                                                      predictions made using the calculated similarities. The
                   𝑡𝑒𝑚𝑝𝑜𝑟𝑎𝑟𝑦 𝑟𝑎𝑡𝑖𝑛𝑔(𝑣, 𝑡𝑗 ) =                         criterion used to determine high ratings was changed from 1
                                                                      to 5 in increments of 1, and similarities were obtained for
         𝑟𝑣,𝑡𝑗          (if 𝑟𝑣,𝑡𝑗 is nonzero)                         each criterion. A criterion of 1 means all ratings are used,
      ∑𝑣′ ∈𝑁𝑣 𝑟𝑣′ ,𝑡𝑗                                          (13)   while a criterion of 5 means that only items rated as 5 are
                        (if 𝑟𝑣,𝑡𝑗 is zero) , (𝑗 = 1, ⋯ , 𝑝),          used.
         |𝑁 ′ 𝑣,𝑡𝑗 |
  {                                                                                         MovieLens 1M         MovieLens 10M
where 𝑁′𝑣,𝑡𝑗 is the set of 𝑣’s neighbors 𝑣′ ∈ 𝑁𝑣 who have
                                                                            Users                6,040                71,567
rated item 𝑡𝑗 . If none of 𝑣’s neighbors have rated 𝑡𝑗 , the
temporary rating will still be zero. In that case, we can                   Items                3,952                10,681
obtain a nonzero temporary rating by calculating the                       Ratings             1,000,209           10,000,054
temporary ratings 𝑡𝑒𝑚𝑝𝑜𝑟𝑎𝑟𝑦 𝑟𝑎𝑡𝑖𝑛𝑔(𝑣′, 𝑡𝑗 ) for each
neighbor 𝑣′ of 𝑣. By calculating the cross entropy of Eq. (7)                   Table 1. Detailed Explanation of Dataset
using these temporary ratings (Eq. (13)) and replacing
  12                                        80                                 80                                           80
  10
   8
   6                                        75                                 75                                           75
   4
   2
   0                                        70                                 70                                           70


                (a)                                            (b)                                 (c)                                   (d)

   Figure 8. (a) is the comparison of calculation times of similarity and (b)-(d) are comparison of ranking accuracy by using
 MovieLens 1M. All horizontal axes represent criteria of high rating and each vertical axis represents minutes (a), NDCG@1 (b),
                                          NDCG@3 (c) and NDCG@5 (d) respectively.

  70                                        80                                 80                                           80
  60
  50
  40
                                            75                                 75                                           75
  30
  20
  10
   0                                        70                                 70                                           70


                (a)                                            (b)                                 (c)                                   (d)

    Figure 9. (a) is the comparison of calculation times of similarity and (b)-(d) are comparison of ranking accuracy by using
  MovieLens 10M. All horizontal axes represent criteria of high rating and each vertical axis represents hours (a), NDCG@1 (b),
                                           NDCG@3 (c) and NDCG@5 (d) respectively.
Figs. 8(a) and 9(a) show the computation time results for                      Comparison of ranking accuracy
MovieLens 1M and 10M, respectively, while Figs. 8(b)-(d)                       Next, we conducted experiments to examine the effect of
and 9(b)-(d) show the corresponding ranking accuracy                           using temporary ratings on ranking prediction accuracy. We
results. The horizontal axes show the high rating criterion in                 compared the accuracy of ranking predictions made using
all graphs, while the vertical axes show calculation time in                   both the proposed and ListCF methods using the similarities
Figs. 8(a) and 9(a), NDCG@1 in Figs. 8(b) and 9(b),                            calculated in the previous subsection. We used an initial
NDCG@3 in Figs. 8(c) and 9(c) and NDCG@5 in Figs.                              𝜑𝑢,𝑔 value of 10 for both datasets with learning rates of
8(d) and 9(d) respectively. As can be seen from Figs. 8(a)                     0.025 and 0.01 for MovieLens 1M and 10M, respectively.
and 9(a), similarity computation is considerably more rapid                    The gradient descent method was repeated 50 times. The
for the proposed method than for conventional ListCF. In                       computation time results are shown in Fig. 10, while
addition, Figs. 8(b)-(d) and 9(b)-(d) show that when scores                    ranking accuracy results are shown in Figs. 11 and 12.
of 4 and 5 are considered to be high ratings, the rankings
are as accurate as those of the conventional method.

                                ListCF prediction    proposed prediction                  ListCF prediction   proposed prediction

                                160                                                       60
                                                                                          50
                                120
                                                                                          40
                                                                                minutes
                      seconds




                                 80                                                       30
                                                                                          20
                                 40
                                                                                          10
                                  0                                                        0
                                      ListCF     1   2         3     4     5                   ListCF    1    2         3        4   5

                                                         (a)                                                      (b)

                                Figure 10. Ranking prediction times for MovieLens 1M (a) and MovieLens 10M (b).
      ListCF prediction       proposed prediction        ListCF prediction       proposed prediction           ListCF prediction       proposed prediction

 85                                                 85                                                    85


 80                                                 80                                                    80


 75                                                 75                                                    75


 70                                                 70                                                    70
       ListCF   1         2       3     4      5          ListCF   1         2          3      4     5          ListCF   1         2       3     4      5

                          (a)                                                (b)                                                   (c)

  Figure 11. Ranking accuracy for MovieLens 1M. Each vertical axis represents NDCG@1 (a), NDCG@3 (b) and NDCG@5 (c).

      ListCF prediction       proposed prediction        ListCF prediction       proposed prediction           ListCF prediction       proposed prediction

 85                                                 85                                                    85


 80                                                 80                                                    80


 75                                                 75                                                    75


 70                                                 70                                                    70
       ListCF   1         2       3      4     5          ListCF   1         2          3      4     5          ListCF   1         2       3     4      5

                          (a)                                                (b)                                                   (c)
  Figure 12. Ranking accuracy for MovieLens 10M. Each vertical axis represents NDCG@1 (a), NDCG@3 (b) and NDCG@5 (c).
Fig. 10 shows that the computation times for the proposed                          REFERENCES
ranking prediction method were shorter in all cases. In                            1.       Zhe Cao, Tao Qin, Tie-Yan Liu, Ming-Feng Tsai, and
addition, Figs. 11 and 12 show that its NDCG@1,                                             Hang Li. 2007. Learning to rank: from pairwise
NDCG@3 and NDCG@5 values were better for all cases.                                         approach to listwise approach. In Proceedings of the
Since MovieLens 1M has less data than MovieLens 10M,                                        24th international conference on Machine learning
the number of temporary ratings increases, and the                                          (ICML '07), 129-136.
difference between the proposed method and the                                              http://dx.doi.org/10.1145/1273496.1273513
conventional ListCF becomes larger than MovieLens 10.
                                                                                   2.       Shanshan Huang, Shuaiqiang Wang, Tie-Yan Liu, Jun
CONCLUSION
                                                                                            Ma, Zhumin Chen, and Jari Veijalainen. 2015. Listwise
In this paper, we have made two proposals. The first is to
                                                                                            Collaborative Filtering. In Proceedings of the 38th
calculate user similarity scores using only high ratings,
                                                                                            International ACM SIGIR Conference on Research and
instead of all ratings, to speed up computation. The second
                                                                                            Development in Information Retrieval (SIGIR '15),
is to introduce temporary estimated ratings for items that
                                                                                            343-352.
have not been rated by neighboring users to improve
                                                                                            http://dx.doi.org/10.1145/2766462.2767693
ranking accuracy.
                                                                                   3.       Kalervo Järvelin and Jaana Kekäläinen. 2000. IR
In experiments using the MovieLens 1M and 10M datasets,
                                                                                            evaluation methods for retrieving highly relevant
we have compared the computation time and accuracy of
                                                                                            documents. In Proceedings of the 23rd annual
both proposals with those of conventional ListCF. The
                                                                                            international ACM SIGIR conference on Research and
results demonstrated that the first proposal provided a
                                                                                            development in information retrieval (SIGIR '00), 41-
considerable reduction in computation time, compared with
                                                                                            48.
conventional ListCF, while maintaining equal or greater
                                                                                            http://dx.doi.org/10.1145/345508.345545
ranking accuracy. In addition, the second proposal
shortened the prediction time in most cases while always                           4.       Kalervo Järvelin and Jaana Kekäläinen. 2002.
improving the ranking accuracy.                                                             Cumulated gain-based evaluation of IR techniques.
                                                                                            ACM Trans. Inf. Syst. 20, 4: 422-446.
In the future, we plan on improving the way neighboring
users are selected and search for a better objective function.                     5.       S.Kullback. 1997. Information Theory and Statistics.
                                                                                            Courier Corporation.
6.   Linden, G., Smith, B., & York, J. 2003. Amazon. com
     recommendations: Item-to-item collaborative filtering.
     IEEE Internet computing, 7, 1: 76-80.
7.   Nathan N. Liu and Qiang Yang. 2008. EigenRank: a
     ranking-oriented approach to collaborative filtering. In
     Proceedings of the 31st annual international ACM
     SIGIR conference on Research and development in
     information      retrieval   (SIGIR     '08),    83-90.
     http://dx.doi.org/10.1145/1390334.1390351
8.   J. I. Marden. 1996. Analyzing and modeling rank data.
     CRC Press.
9.   Badrul Sarwar, George Karypis, Joseph Konstan, and
     John Riedl. 2001. Item-based collaborative filtering
     recommendation algorithms. In Proceedings of the
     10th international conference on World Wide Web
     (WWW                     '01),            285-295.
     http://dx.doi.org/10.1145/371920.372071
10. Shuaiqiang Wang, Jiankai Sun, Byron J. Gao, and Jun
    Ma. 2014. VSRank: A Novel Framework for Ranking-
    Based Collaborative Filtering. ACM Trans. Intell. Syst.
    Technol. 5, 3: 24.