<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>div2vec: Diversity-Emphasized Node Embedding</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Jisu Jeong</string-name>
          <email>jisu.jeong@navercorp.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Young-Jin Park</string-name>
          <email>young.j.park@navercorp.com</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Jeong-Min Yun</string-name>
          <email>matthew@watcha.com</email>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Zimin Park</string-name>
          <email>holden@watcha.com</email>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Hongi Keam</string-name>
          <email>paul@watcha.com</email>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Junki Cho</string-name>
          <email>leo@watcha.com</email>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Clova AI Research, NAVER Corp.</institution>
          ,
          <addr-line>Seongnam</addr-line>
          ,
          <country country="KR">South Korea</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Naver R&amp;D Center</institution>
          ,
          <addr-line>NAVER Corp., Seoul</addr-line>
          ,
          <country country="KR">South Korea</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>WATCHA Inc.</institution>
          ,
          <addr-line>Seoul</addr-line>
          ,
          <country country="KR">South Korea</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Recently, the interest of graph representation learning has been rapidly increasing in recommender systems. However, most existing studies have focused on improving accuracy, but in real-world systems, the recommendation diversity should be considered as well to improve user experiences. In this paper, we propose the diversityemphasized node embedding div2vec, which is a random walk-based unsupervised learning method like DeepWalk and node2vec. When generating random walks, DeepWalk and node2vec sample nodes of higher degree more and nodes of lower degree less. On the other hand, div2vec samples nodes with the probability inversely proportional to its degree so that every node can evenly belong to the collection of random walks. This strategy improves the diversity of recommendation models. Ofline experiments on the MovieLens dataset showed that our new method improves the recommendation performance in terms of both accuracy and diversity. Moreover, we evaluated the proposed model on two real-world services, WATCHA and LINE Wallet Coupon, and observed the div2vec improves the recommendation quality by diversifying the system.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>CCS CONCEPTS</title>
      <p>• Computing methodologies → Learning latent
representations; Machine learning algorithms; Knowledge representation
and reasoning; Neural networks.</p>
    </sec>
    <sec id="sec-2">
      <title>INTRODUCTION</title>
      <p>
        Most recommender system studies have focused on finding users’
immediate demands; they try to build models that maximize the
click-through rate (CTR). The learned system suggests high-ranked
items that users are likely to click in a myopic sense [
        <xref ref-type="bibr" rid="ref30 ref32 ref6 ref9">6, 9, 30, 32</xref>
        ].
Such recommendation strategies have successfully altered simple
popularity-based or handmade lists, thus being widely adopted on
many online platforms including Spotify [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ], Netflix [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ], and so
on.
      </p>
      <p>However, the previous approaches have a potentially severe
drawback, a lack of diversity. For example, consider a user just
watched Iron Man. Since a majority of people tend to watch other
Marvel Cinematic Universe (MCU) films like Iron Man 2, Thor, and
Marvel’s The Avengers after watching Iron Man, the system would
recommend such MCU films based on historical log data. While
the approach may lead to CTR maximization, 1) can we say that
users are satisfied with these apparent results? Or, 2) would a wider
variety of recommendations achieve better user experience?</p>
      <p>
        Recently, a method that addresses the first question is presented
on Spotify [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. This work categorized those who listen to very
similar songs and diferent sets of entities as specialists and generalists,
respectively. This work observed that generalists are much more
satisfied than specialists based on long-term user metrics (i.e., the
conversion to subscriptions and retention on the platform). Thus,
even if some users are satisfied with the evident recommendations
(clicked or played), this satisfaction does not imply the users
continue to use the platform.
      </p>
      <p>
        To answer the second question, we propose the diversity-emphasized
node embedding div2vec. Recently, the number of studies on graph
structure [
        <xref ref-type="bibr" rid="ref10 ref17 ref24 ref26 ref28 ref8">8, 10, 17, 24, 26, 28</xref>
        ] is increasing. Unfortunately, most of
those studies have merely focused on the accuracy. DeepWalk and
node2vec are the first and the most famous node embedding
methods [
        <xref ref-type="bibr" rid="ref24 ref8">8, 24</xref>
        ]. DeepWalk, node2vec, and our method div2vec generates
random walks first and then use the Skip-gram model [
        <xref ref-type="bibr" rid="ref20">20</xref>
        ] to
compute embedding vectors of all nodes. When generating random
walks, their methods choose nodes of high degree more. It makes
sense because if a node had many neighbors in the past, it will have
many neighbors in the future, too. However, it may be an obstacle
for personalizing. Using our new method, all nodes are evenly
distributed in the collection of random walks. Roughly speaking, the
key idea is to choose a node with weight 1 where  is the degree
of the node. Also, we propose a variant of div2vec, which we call
rooted div2vec, obtained by changing the weight 1 to √1 in order to

balance accuracy and diversity. To the best of our knowledge, our
approach is the first node embedding method focusing on diversity.
Details are in Section 3.
      </p>
      <p>We evaluate our new methods on a benchmark and two
realword datasets. As the benchmark, we verify the ofline metrics on
the MovieLens dataset. As a result, div2vec got higher scores in the
diversity metrics, such as coverage, entropy-diversity, and average
intra-list similarity, and lower scores in the accuracy metrics, such
as AUC score, than DeepWalk and node2vec. Furthermore, its
variant rooted div2vec had the highest AUC score and also the diversity
scores of rooted div2vec are the best or the second-best.</p>
      <p>We figure out that increasing diversity actually improves online
performance. We test on two diferent live services, WATCHA and
LINE Wallet Coupon. Screenshots of the services are in Figure 1.
WATCHA is one of the famous OTT streaming services in South
Korea. Like Netflix, users can watch movies and TV series using
WATCHA. LINE is the most popular messenger in Japan, and LINE
Wallet Coupon service provides various coupons, such as, pizza,
coffee, shampoo, etc. In the above two diferent kinds of recommender
systems, we used our diversity-emphasized node embedding and
succeeded to enhance online performances. It is the biggest
contribution of our work to prove that users in real world prefer a diverse
and personalized recommendation.</p>
      <p>The structure of the paper is as follows. In Section 2, we review
random walk-based node embedding methods and the study on
diversity problems. The proposed method will be described in
Section 3. Section 4 and Section 5 reports the results of our experiments
on ofline tests and online tests, respectively. Section 6 concludes
our research.</p>
    </sec>
    <sec id="sec-3">
      <title>RELATED WORK</title>
    </sec>
    <sec id="sec-4">
      <title>Random walk-based node embeddings</title>
      <p>
        The famous word2vec method transforms words into embedding
vectors such that similar words have similar embeddings. It uses
a language model, called Skip-gram [
        <xref ref-type="bibr" rid="ref20">20</xref>
        ], that maximizes the
cooccurrence probability among the words near the target word.
      </p>
      <p>
        Inspired by word2vec, Perozzi et al. [
        <xref ref-type="bibr" rid="ref24">24</xref>
        ] introduced DeepWalk that
transforms nodes in a graph into embedding vectors. A walk is a
sequence of nodes in a graph such that two consecutive nodes are
adjacent. A random walk is a walk such that the next node in the
walk is chosen randomly from the neighbors of the current node.
DeepWalk first samples a collection of random walks from the
input graph where each node in random walks are chosen uniformly
at random. Once a collection of random walks is generated, we
treat nodes and random walks as words and sentences, respectively.
Then by applying word2vec method, we can obtain an embedding
vector of each node.
      </p>
      <p>
        node2vec [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] is a generalization of DeepWalk. When nodes in
random walks are chosen, node2vec uses two parameters  and .
Suppose we have an incomplete random walk 1, 2, . . . ,  and we
will choose one node in the neighborhood  ( ) of  to be +1.
Here, for  in  ( ), we set the weight  (,  ) as follows:
1
 
 (,  ) = 1
 1
 
if  = −1,
if  is adjacent to −1,
otherwise.
      </p>
      <p>Note that if a graph is bipartite, the second case does not appear.
node2vec chooses +1 at random with the weight  (,  ).</p>
      <p>The most advantage of graph representation learning or graph
neural networks is that these models can access both local and
higher-order neighborhood information. However, as the number
of edges is usually too many, they may be ineficient. The random
walk-based method solves this problem. Instead of considering all
nodes and all edges, it only considers the nodes in the collection
of random walks. Therefore, the way to generate random walks is
important and it afects performance.
2.2</p>
    </sec>
    <sec id="sec-5">
      <title>Diversity problems</title>
      <p>
        The word “filter bubble” refers to a phenomenon in which the
recommender system blocks providing various information and
iflters only information similar to the user’s taste. In [
        <xref ref-type="bibr" rid="ref21 ref22 ref3">3, 21, 22</xref>
        ],
they show the existence of the filter bubble in their recommender
system. Some research [
        <xref ref-type="bibr" rid="ref1 ref27">1, 27</xref>
        ] claim that diversity is one of the
essential components in the recommender system.
      </p>
      <p>
        Some studies are proving that diversity increases the user’s
satisfaction. Spotify, one of the best music streaming services, observed
that diverse consumption behaviors are highly associated with
long-term metrics like conversion and retention [
        <xref ref-type="bibr" rid="ref13 ref2">2, 13</xref>
        ]. Also, Liu
et al. [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ] improve the user’s preference by using neural graph
ifltering which learns diverse fashion collocation.
      </p>
      <p>
        One may think that if a recommender system gains diversity,
then it looses the accuracy. However, the following research
succeeds in improving both. Adomavicius and Kwon [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] applied a
ranking technique to original collaborative filtering in order to
increase diversity without decreasing the accuracy. Zheng et al. [
        <xref ref-type="bibr" rid="ref31">31</xref>
        ]
proposed a Deep Q-Learning based reinforcement learning
framework for news recommendation. Their model improves both
clickthrough rate and intra-list similarity.
In the framework of DeepWalk and node2vec, the model first
generates a collection of random walks, and then runs the famous
word2vec algorithm to obtain embedding vectors. In their way,
(a) DeepWalk, movieId
(b) DeepWalk, userId
(c) rooted div2vec, movieId
(d) rooted div2vec, userId
(e) div2vec, movieId
(f) div2vec, userId
nodes of high degree should be contained more than nodes of low
degree in the collection of random walks because, roughly speaking,
if a node  has  neighbors, then there are  chances that  can
belongs to the collection of random walks. Figure 2a and Figure 2b
represent this phenomenon. The -axis are the nodes sorted by the
degree and the blue line means the degree of nodes. So, the blue line
is always increasing and it means that nodes of higher degrees are
on the right side in each figure. Orange bars mean the frequencies
of nodes in the collection of random walks. It is easy to observe
that nodes of high degree appear extremely more than that of low
degree.
      </p>
      <p>As the collection of random walks are the training set of the
skimgram model, DeepWalk and node2vec might be trained with a bias
to nodes of high degree. It may not be a trouble for solving problems
focused on the accuracy. For example, in link prediction,
highdegree nodes in the original graph might have a higher probability
of being linked with other nodes than low-degree nodes. In terms
of movies, if the movie is popular, then many people will like this
movie. However, it must be a problem when we want to focus
on personalization and diversity. Unpopular movies might not be
trained enough so that they are not well-represented. So, even if
a person actually prefers some unpopular movie to some popular
movie, the recommender system tends to recommend the popular
movie for safe.</p>
      <p>Motivated by Figure 3, which shows the diference between
reality and equity, we decided to consider the degree of the next
candidate nodes inversely. The main idea is ‘Low degree, choose
more.’. We propose a simple but creative method, which will be
formally described in the next subsection, which gives Figure 2e
and Figure 2f. Compare to Figure 2a and Figure 2b, the nodes in
Now, we introduce the diversity-emphasized node embedding method.
Similarly to DeepWalk and node2vec, we first produce a lot of
random walks and train skip-gram model. We apply an easy but bright
idea to generate random walks so that our model can capture the
diversity of the nodes in their embedding vectors.</p>
      <p>Suppose a node  is the last node in an incomplete random walk
and we are going to choose the next node among the neighbors
of .</p>
      <p>• DeepWalk picks the next node in  () at random with the
same probability.
• In node2vec, if  is the node added to the random walk just
before , then there are three types of probability depend
on whether a node  ∈  () is adjacent with  or not, or
 =  .
• Our method will choose the next node according to the
degree of neighbors.</p>
      <p>Formally, our method chooses the next node  ∈  () with the
probability</p>
      <p>(deg())
Í ∈ ()  (deg( ))
for some function  . For example, when  () = 1/, if  has two
neighbors  and  whose degree is 10 and 90 respectively, then  is
chosen with probability (1/10)/(1/10 + 1/90) = 0.9 and  is chosen
with probability 0.1. That is, since the degree of  is smaller than
the degree of , the probability that  is chosen is larger than the
probability that  is chosen. In Section 4, we set  to the inverse
of the identity function  ( ) = 1/ and the inverse of the square
root function  ( ) = 1/√ . We call this method div2vec when
 ( ) = 1/ and rooted div2vec when  ( ) = 1/√ .</p>
      <p>Intuitively, DeepWalk chooses the next node without
considering the past or the future nodes, node2vec chooses the next node
according to the past node, and div2vec chooses the next node
with respect to the future node. Note that it is possible to
combine node2vec and div2vec by first dividing into three types and
then consider the degree of neighbors. Since there are too many
hyperparameters to control, we remain it to the next work.</p>
      <p>Figure 2 is the result for generating random walks with several
methods. The detail for the dataset is in Subsection 4.1. If we use
DeepWalk, then Figure 2a and Figure 2b show that high-degree
nodes appears extremely much more than low-degree nodes. The
problem is that, if the result is too skew, then the skip-gram model
might not train some part of data well. For example, a popular
movie will appear many times in the collection of random walks
and then the model should overfit to the popular movie. On the
other hands, an unpopular movie will appears only few times in
the collection of random walks and then the model should underfit
to the unpopular movie.</p>
      <p>This problem is solvable by using our method. Using div2vec,
we can have the nodes evenly in the collection of random walks.
Figure 2e and Figure 2f show that our method solves this problem.
The nodes are chosen equally regardless of the degree of nodes.
Normally, popular movies are consumed more than unpopular movies.
So div2vec may decrease the accuracy. Our experiments prove that
even if we emphasize the diversity, the accuracy decrease very little.
Furthermore, we suggest the variant rooted div2vec. Figure 2c and
Figure 2d can be treated as the combination of DeepWalk and
div2vec. In Subsection 4.4, our experiments show that compare to
DeepWalk and node2vec, rooted div2vec records better scores for
every metric.
4
4.1</p>
    </sec>
    <sec id="sec-6">
      <title>OFFLINE EXPERIMENTS</title>
    </sec>
    <sec id="sec-7">
      <title>DataSets</title>
      <p>
        We used the famous MovieLens dataset [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] for an ofline test. We
used MovieLens-25M because it is the newest data and we only
used the recent 5 years in the dataset. Rating data is made on
1This figure is from http://www.brainkart.com/article/Equality_34271/.
10 steps, but we need binary data, which means watched/not or
satisfied/unsatisfied, in order to train a model and compute AUC
score. We set more than 4 stars to be positive and less than 3 stars to
be negative. To prevent noises, we remove both the movies having
less than 10 records and the users having less than 10 or more than
1000 records. At last, there are 2,009,593 records with 16,002 users
and 5,298 movies. For the test set, 20% of the data are used.
      </p>
      <p>
        To compute intra-list similarity, which will be described in
Subsection 4.3, we use ‘Tag Genome’ [
        <xref ref-type="bibr" rid="ref29">29</xref>
        ] from MovieLens-25M. It
contains data in ‘movieId, tagId, relevance’ format for every pair
of movies and tags. Relevance values are real numbers between
0 and 1. So, it can be treated as a dense matrix and one row that
represents one movie means a vector containing tag information.
4.2
      </p>
    </sec>
    <sec id="sec-8">
      <title>Experiment settings</title>
      <p>
        In movie recommender systems, a model recommends a list of
movies to each user. In other words, a model needs to find out which
movies will be connected with an individual user. It means that
our task is a link prediction. However, the methods we discussed
so far are only compute node embeddings. That is, we have an
embedding vector for movies and users but not for their interactions.
Grover and Leskovec [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] introduced four operators to obtain edge
embeddings from node embeddings as follows. Let  and  be two
nodes and  () and  ( ) be their embedding vectors.
(1) Average:  ()+ ()
      </p>
      <p>2
(2) Hadamard:  () ∗  ( ) (element-wise product)
(3) Weighted-L1: | () −  ( ) |
(4) Weighted-L2: | () −  ( ) |2</p>
      <p>For each edge, we obtain 64-dim vector from the graph induced
by positive edges and 64-dim vector from the graph induced by
negative edges. And then we concatenate the positive edge
embedding vector and the negative edge embedding vector to represent
the edge embedding vector.</p>
      <p>To avoid disrupting the performance of a prediction model, we
use simple deep neural network with one hidden layer of size 128.
4.3</p>
    </sec>
    <sec id="sec-9">
      <title>Evaluation metrics</title>
      <p>For each embedding and each operator, we compute four metrics,
one for accuracy and the others for diversity. The larger score means
the better performance.
(a) The results with the operator Weighted-L1.
(b) The results with the operator Weighted-L2.</p>
      <p>(ℎ      )
FPR =</p>
      <p>(ℎ      ) + (ℎ     )
AUC score is in range 0 to 1. As close to 1, it gives better evaluation
and is close to perfect prediction. AUC score is useful evaluation
metric because of scale invariant and classification threshold
invariant to compare multiple prediction model.</p>
      <p>COVERAGE Coverage is how many items appear in the
recommended result. Formally, we can define as
Ø
coverage() =</p>
      <p>
        , ()

where  is a model, , is a set of top- recommended items
for a user  by . Many papers [
        <xref ref-type="bibr" rid="ref1 ref12 ref16 ref7">1, 7, 12, 16</xref>
        ] discuss the
importance of the coverage. If the coverage of the model is large, then
the model recommends a broad range of items, and it implicitly
ENTROPY-DIVERSITY() = − Õ  () ln  () .
  × | |  × | |
Note that we can say that if ENTROPY-DIVERSITY(1) &lt;
ENTROPYDIVERSITY(2), then 2 recommends more diverse items than 1.
Here is an example. For an item set  = {1, 2, . . . , 9}
and a user set  = { 1,  2,  3}, suppose a model 1 gives
1,3 () = {1, 2, 3} for every user , and a model
2 gives 2,3 ( 1) = {1, 2, 3}, 2,3 ( 2) =
{4, 5, 6}, 2,3 ( 3) = {7, 8, 9}. Then
ENTROPY-DIVERSITY(1) = −(3/9) ln(3/9) × 3 + 0 × 6 = ln 3 and
ENTROPY-DIVERSITY(2) = −(1/9) ln(1/9) × 9 = ln 9.
      </p>
      <p>AVERAGE INTRA-LIST SIMILARITY From the
recommendation model, every user will receive a list of items. Intra-List
Similarity (ILS) measures how dissimilar or similar items in the list are.</p>
      <p>
        || (|| − 1)/2
where  is the recommended item list and sim(,  ) is the similarity
measure between the tag-genome vectors of  and  , which are
given from the MovieLens dataset [
        <xref ref-type="bibr" rid="ref11 ref29">11, 29</xref>
        ]. We set sim(,  ) = 1 −
 · where  and   are corresponding tag-genome vectors
| | | | · | | | |
of  and  . By definition, if the value is small, then items in the
list are similar. Otherwise, they are dissimilar. Note that Bradly
and Smyth [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ], and Meymandpour and Davis [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ] use the same
definition in terms of ‘diversity’. For every user , we compute
  (, ()) and their average, which we call Average Intra-List
Similarity in order to measure how diverse a model is.
      </p>
      <p>
        • five methods: DeepWalk [
        <xref ref-type="bibr" rid="ref24">24</xref>
        ], node2vec [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] with diferent
hyperparameters, div2vec and its variant rooted div2vec
• four operators: Weighted-L1, Weighted-L2, Hadamard,
Average
• four metrics: AUC score, coverage, entropy-diversity, and
average intra-list similarity
• three sizes of recommendation lists: 1, 10, 50
AUC means AUC score. CO(), ED(), ILS() means coverage,
entropydiversity, average intra-list similarity of top  recommended items,
respectively. n2v-(p,q) is node2vec with hyperparameter , .
      </p>
      <p>Our proposed methods div2vec and rooted div2vec record the
highest scores on all metrics in Table 1a and Table 1b In Table 1c
and Table 1d, the average intra-list similarity is not the best but the
second with tiny gaps. Overall, it is easy to see that div2vec and
rooted div2vec got better scores than DeepWalk and node2vec in
diversity metrics. Furthermore, rooted div2vec got the best scores
in the accuracy metric. Thus, we can conclude that our proposed
methods increase the diversity of recommender systems.
5
5.1</p>
    </sec>
    <sec id="sec-10">
      <title>LIVE EXPERIMENTS</title>
    </sec>
    <sec id="sec-11">
      <title>Video Recommendation</title>
      <p>In the previous experiments, we verified that our methods, div2vec and
rooted div2vec, clearly increase the diversity of recommended
results. The remaining job is to prove that div2vec actually increases
user satisfaction in real-world recommender systems. To show this,
we conduct an A/B test in the commercial video streaming service,
WATCHA, and measure and compare various user activity statistics
that are related to user satisfaction.</p>
      <p>We collected four months watch-complete logs starting from
January 2020, here watch-complete means user completing a video.
We filter-out users who do not have watch-complete logs last few
days, also filter-out extreme case users (too many or too few logs);
results in 21,620 users. Two methods, node2vec and div2vec, were
trained with these filtered logs. For node2vec, we set the parameters
 =  = 1.</p>
      <p>An A/B test had been conducted two weeks, where 21,620 users
were randomly and evenly partitioned into two groups and each
group received either node2vec or div2vec recommending results. In
more detail, WATCHA has list-wise recommendation home whose
list consists of several videos, and our list inserted into the fifth
row. To make the list, we sorted all available videos by the final
scores and pick top  of them ( varies with devices), and also
apply random shufling of top 3 videos to avoid always the same
recommendation.</p>
      <p>We compare clicks and plays of node2vec and div2vec list by the
week. (The first two columns in Table 2) In the first week, div2vec list
received more than 60% more actions than node2vec list in both
clicks and plays; 39.69% more clicks and 28.52% more plays at the
second week. As we can see that div2vec beats node2vec with clicks
and plays by a significant margin.</p>
      <p>
        Someone may argue that the above improvement does not
improve actual user satisfaction; if users who received the node2vec list
are satisfied with other recommended list. To see this actually
happens, we compare total watch time of each group during the test.
(The last column in Table 2) In the first week, div2vec achieved 3.58%
more watch time than node2vec, and 4.19% in the second week. Let
me note that in watch time comparison, even 1% improvement
is hard to achieve [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ], thus our improvement is quite impressive
results.
5.2
      </p>
    </sec>
    <sec id="sec-12">
      <title>Coupon Recommendation</title>
      <p>
        To further demonstrate the efectiveness of using the div2vec
embedding in other real-world recommender systems, we run an A/B
test in the LINE Wallet Coupon service and evaluate the online
performance for two weeks in the spring of 2020. The system consists
of over six million users and over five hundred unique items. We
constructed the user-item bipartite graph by defining each user and
item as an individual node and connecting nodes that are interacted
each other. Using the graph, we obtained div2vec embedding
vectors for each nodes. In this experiment, we compared the number
of unique users clicked (Click UU) and click-through rate (CTR)
of the neural-network based recommendation model2 using the
precomputed div2vec embedding vectors as additional features to
the model that did not. As side information, the paper utilized
gender, age, mobile OS type, and interest information for users, while
brand, discount information, text, and image features for items. The
online experiment results show that the overall Click UU and CTR
have been improved by 6.55% and 2.90%, respectively. The relative
2The details in model architecture for the LINE Wallet Coupon recommender system
are presented in [
        <xref ref-type="bibr" rid="ref23 ref25">23, 25</xref>
        ].
1.15
R1.10
T
C
performance for two weeks is illustrated in Figure 4 by date. By
applying the div2vec feature, a larger number of users get interested
in the recommended coupon list and the ratio that the user reacts
to the exposed item increases, significantly. Considering that the
online tests were conducted for a relatively long period, we
conclude that the diversified recommendation based on the proposed
method has led to positive consequences in user experience rather
than to attract curiosity from users temporarily.
6
      </p>
    </sec>
    <sec id="sec-13">
      <title>CONCLUSION</title>
      <p>We have introduced the diversity-emphasized node embedding
div2vec. Several experiments showed the importance of our method.
Compared to DeepWalk and node2vec, the recommendation model
based on div2vec increased the diversity metrics like coverage,
entropy-diversity, average intra-list similarity in the MovieLens
dataset. The main contribution of this paper is that we verified that
users satisfy with the recommendation model using div2vec in two
diferent live services.</p>
      <p>We remark that as div2vec is an unsupervised learning method
like word2vec, it can be easily combined with other studies and
services, and it is possible to improve their performance. Also, by
changing the function  , the distribution of nodes in the collection
of random walks can be adjusted to each domain.</p>
    </sec>
    <sec id="sec-14">
      <title>ACKNOWLEDGMENTS</title>
      <p>Special thanks to those who lent their insight and technical support
for this work, including Jaehun Kim, Taehyun Lee, Kyung-Min Kim,
and Jung-Woo Ha.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>Gediminas</given-names>
            <surname>Adomavicius and YoungOk Kwon</surname>
          </string-name>
          .
          <year>2012</year>
          .
          <article-title>Improving Aggregate Recommendation Diversity Using Ranking-Based Techniques</article-title>
          .
          <source>IEEE Trans. on Knowl. and Data Eng</source>
          .
          <volume>24</volume>
          ,
          <issue>5</issue>
          (May
          <year>2012</year>
          ),
          <fpage>896</fpage>
          -
          <lpage>911</lpage>
          . https://doi.org/10.1109/TKDE.
          <year>2011</year>
          .15
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>Ashton</given-names>
            <surname>Anderson</surname>
          </string-name>
          , Lucas Maystre,
          <string-name>
            <surname>Ian Anderson</surname>
            ,
            <given-names>Rishabh</given-names>
          </string-name>
          <string-name>
            <surname>Mehrotra</surname>
            , and
            <given-names>Mounia</given-names>
          </string-name>
          <string-name>
            <surname>Lalmas</surname>
          </string-name>
          .
          <year>2020</year>
          .
          <article-title>Algorithmic efects on the diversity of consumption on spotify</article-title>
          .
          <source>In Proceedings of The Web Conference</source>
          <year>2020</year>
          .
          <fpage>2155</fpage>
          -
          <lpage>2165</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>Eytan</given-names>
            <surname>Bakshy</surname>
          </string-name>
          , Solomon Messing, and
          <string-name>
            <surname>Lada</surname>
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Adamic</surname>
          </string-name>
          .
          <year>2015</year>
          .
          <article-title>Exposure to ideologically diverse news and opinion on Facebook</article-title>
          .
          <source>Science</source>
          <volume>348</volume>
          ,
          <issue>6239</issue>
          (
          <year>2015</year>
          ),
          <fpage>1130</fpage>
          -
          <lpage>1132</lpage>
          . https://doi.org/10.1126/science.aaa1160 arXiv:https://science.sciencemag.org/content/348/6239/1130.full.pdf
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>Keith</given-names>
            <surname>Bradley</surname>
          </string-name>
          and
          <string-name>
            <given-names>Barry</given-names>
            <surname>Smyth</surname>
          </string-name>
          .
          <year>2001</year>
          . Improving Recommendation Diversity.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>Minmin</given-names>
            <surname>Chen</surname>
          </string-name>
          , Alex Beutel, Paul Covington, Sagar Jain, Francois Belletti, and Ed H Chi.
          <year>2019</year>
          .
          <article-title>Top-k of-policy correction for a REINFORCE recommender system</article-title>
          .
          <source>In Proceedings of the Twelfth ACM International Conference on Web Search and Data Mining</source>
          .
          <fpage>456</fpage>
          -
          <lpage>464</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <surname>Heng-Tze</surname>
            <given-names>Cheng</given-names>
          </string-name>
          , Levent Koc, Jeremiah Harmsen, Tal Shaked, Tushar Chandra, Hrishi Aradhye,
          <string-name>
            <surname>Glen Anderson</surname>
            , Greg Corrado, Wei Chai, Mustafa Ispir, Rohan Anil, Zakaria Haque, Lichan Hong, Vihan Jain, Xiaobing Liu, and
            <given-names>Hemal</given-names>
          </string-name>
          <string-name>
            <surname>Shah</surname>
          </string-name>
          .
          <year>2016</year>
          .
          <article-title>Wide Deep Learning for Recommender Systems</article-title>
          .
          <source>In Proceedings of the 1st Workshop on Deep Learning for Recommender Systems</source>
          (Boston, MA, USA) (
          <year>DLRS 2016</year>
          ).
          <article-title>Association for Computing Machinery</article-title>
          , New York, NY, USA,
          <fpage>7</fpage>
          -
          <lpage>10</lpage>
          . https://doi.org/10.1145/2988450.2988454
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>Mouzhi</given-names>
            <surname>Ge</surname>
          </string-name>
          , Carla Delgado-Battenfeld, and
          <string-name>
            <given-names>Dietmar</given-names>
            <surname>Jannach</surname>
          </string-name>
          .
          <year>2010</year>
          .
          <article-title>Beyond Accuracy: Evaluating Recommender Systems by Coverage and Serendipity</article-title>
          .
          <source>In Proceedings of the Fourth ACM Conference on Recommender Systems (Barcelona</source>
          , Spain) (
          <source>RecSys '10)</source>
          .
          <article-title>Association for Computing Machinery</article-title>
          , New York, NY, USA,
          <fpage>257</fpage>
          -
          <lpage>260</lpage>
          . https://doi.org/10.1145/1864708.1864761
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>Aditya</given-names>
            <surname>Grover</surname>
          </string-name>
          and
          <string-name>
            <given-names>Jure</given-names>
            <surname>Leskovec</surname>
          </string-name>
          .
          <year>2016</year>
          .
          <article-title>Node2vec: Scalable Feature Learning for Networks</article-title>
          .
          <source>In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining</source>
          (San Francisco, California, USA) (
          <source>KDD '16)</source>
          .
          <article-title>Association for Computing Machinery</article-title>
          , New York, NY, USA,
          <fpage>855</fpage>
          -
          <lpage>864</lpage>
          . https: //doi.org/10.1145/2939672.2939754
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>Huifeng</given-names>
            <surname>Guo</surname>
          </string-name>
          , Ruiming Tang, Yunming Ye,
          <string-name>
            <given-names>Zhenguo</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>and Xiuqiang</given-names>
            <surname>He</surname>
          </string-name>
          .
          <year>2017</year>
          .
          <article-title>DeepFM: A Factorization-Machine Based Neural Network for CTR Prediction</article-title>
          .
          <source>In Proceedings of the 26th International Joint Conference on Artificial Intelligence (Melbourne</source>
          , Australia) (
          <source>IJCAI'17)</source>
          . AAAI Press,
          <fpage>1725</fpage>
          -
          <lpage>1731</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <surname>William</surname>
            <given-names>L.</given-names>
          </string-name>
          <string-name>
            <surname>Hamilton</surname>
            , Rex Ying, and
            <given-names>Jure</given-names>
          </string-name>
          <string-name>
            <surname>Leskovec</surname>
          </string-name>
          .
          <year>2017</year>
          .
          <article-title>Inductive Representation Learning on Large Graphs</article-title>
          .
          <source>In Proceedings of the 31st International Conference on Neural Information Processing Systems</source>
          (Long Beach, California, USA) (
          <source>NIPS'17)</source>
          . Curran Associates Inc.,
          <string-name>
            <surname>Red</surname>
            <given-names>Hook</given-names>
          </string-name>
          ,
          <string-name>
            <surname>NY</surname>
          </string-name>
          , USA,
          <fpage>1025</fpage>
          -
          <lpage>1035</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>F.</given-names>
            <surname>Maxwell</surname>
          </string-name>
          Harper and
          <string-name>
            <given-names>Joseph A.</given-names>
            <surname>Konstan</surname>
          </string-name>
          .
          <year>2015</year>
          .
          <article-title>The MovieLens Datasets: History and Context</article-title>
          .
          <source>ACM Trans. Interact. Intell. Syst. 5</source>
          ,
          <issue>4</issue>
          ,
          <string-name>
            <surname>Article 19</surname>
          </string-name>
          (
          <issue>Dec</issue>
          .
          <year>2015</year>
          ),
          <volume>19</volume>
          pages. https://doi.org/10.1145/2827872
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <surname>Jonathan</surname>
            <given-names>L.</given-names>
          </string-name>
          <string-name>
            <surname>Herlocker</surname>
          </string-name>
          , Joseph A.
          <string-name>
            <surname>Konstan</surname>
          </string-name>
          , Loren G. Terveen, and John T. Riedl.
          <year>2004</year>
          .
          <article-title>Evaluating Collaborative Filtering Recommender Systems</article-title>
          .
          <source>ACM Trans. Inf. Syst</source>
          .
          <volume>22</volume>
          ,
          <issue>1</issue>
          (Jan.
          <year>2004</year>
          ),
          <fpage>5</fpage>
          -
          <lpage>53</lpage>
          . https://doi.org/10.1145/963770.963772
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>David</given-names>
            <surname>Holtz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Ben</given-names>
            <surname>Carterette</surname>
          </string-name>
          , Praveen Chandar, Zahra Nazari, Henriette Cramer, and
          <string-name>
            <given-names>Sinan</given-names>
            <surname>Aral</surname>
          </string-name>
          .
          <year>2020</year>
          .
          <article-title>The Engagement-Diversity Connection: Evidence from a Field Experiment on Spotify</article-title>
          .
          <source>In Proceedings of the 21st ACM Conference on Economics and Computation (Virtual Event, Hungary) (EC '20)</source>
          .
          <article-title>Association for Computing Machinery</article-title>
          , New York, NY, USA,
          <fpage>75</fpage>
          -
          <lpage>76</lpage>
          . https://doi.org/10.1145/ 3391403.3399532
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <surname>Xiao</surname>
            hua Liu, Yongbin Sun, Ziwei Liu, and
            <given-names>Dahua</given-names>
          </string-name>
          <string-name>
            <surname>Lin</surname>
          </string-name>
          .
          <year>2020</year>
          .
          <article-title>Learning Diverse Fashion Collocation by Neural Graph Filtering</article-title>
          . ArXiv abs/
          <year>2003</year>
          .04888 (
          <year>2020</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <surname>Kurt</surname>
            <given-names>Jacobson</given-names>
          </string-name>
          , Vidhya Murali, Edward Newett, Brian Whitman, and
          <string-name>
            <given-names>Romain</given-names>
            <surname>Yon</surname>
          </string-name>
          .
          <year>2016</year>
          .
          <article-title>Music Personalization at Spotify</article-title>
          .
          <source>In Proceedings of the 10th ACM Conference on Recommender Systems</source>
          (Boston, Massachusetts, USA) (
          <source>RecSys '16)</source>
          .
          <article-title>Association for Computing Machinery</article-title>
          , New York, NY, USA,
          <volume>373</volume>
          . https://doi.org/10.1145/ 2959100.2959120
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>Marius</given-names>
            <surname>Kaminskas</surname>
          </string-name>
          and
          <string-name>
            <given-names>Derek</given-names>
            <surname>Bridge</surname>
          </string-name>
          .
          <year>2016</year>
          . Diversity, Serendipity, Novelty, and
          <article-title>Coverage: A Survey and Empirical Analysis of Beyond-Accuracy Objectives in Recommender Systems</article-title>
          .
          <source>ACM Trans. Interact. Intell. Syst. 7</source>
          ,
          <issue>1</issue>
          ,
          <string-name>
            <surname>Article 2</surname>
          </string-name>
          (
          <issue>Dec</issue>
          .
          <year>2016</year>
          ),
          <volume>42</volume>
          pages. https://doi.org/10.1145/2926720
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <surname>Kyung-Min</surname>
            <given-names>Kim</given-names>
          </string-name>
          , Donghyun Kwak, Hanock Kwak,
          <string-name>
            <surname>Young-Jin</surname>
            <given-names>Park</given-names>
          </string-name>
          , Sangkwon Sim,
          <string-name>
            <surname>Jae-Han</surname>
            <given-names>Cho</given-names>
          </string-name>
          , Minkyu Kim, Jihun Kwon, Nako Sung, and
          <string-name>
            <surname>Jung-Woo Ha</surname>
          </string-name>
          .
          <year>2019</year>
          .
          <article-title>Tripartite Heterogeneous Graph Propagation for Large-scale Social Recommendation</article-title>
          . arXiv preprint arXiv:
          <year>1908</year>
          .
          <volume>02569</volume>
          (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <surname>Dawen</surname>
            <given-names>Liang</given-names>
          </string-name>
          , Rahul G. Krishnan,
          <string-name>
            <given-names>Matthew D.</given-names>
            <surname>Hofman</surname>
          </string-name>
          , and
          <string-name>
            <given-names>Tony</given-names>
            <surname>Jebara</surname>
          </string-name>
          .
          <year>2018</year>
          .
          <article-title>Variational Autoencoders for Collaborative Filtering</article-title>
          .
          <source>In Proceedings of the 2018 World Wide Web Conference (Lyon, France) (WWW '18)</source>
          .
          <source>International World Wide Web Conferences Steering Committee, Republic and Canton of Geneva, CHE</source>
          ,
          <fpage>689</fpage>
          -
          <lpage>698</lpage>
          . https://doi.org/10.1145/3178876.3186150
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>Rouzbeh</given-names>
            <surname>Meymandpour</surname>
          </string-name>
          and
          <string-name>
            <given-names>Joseph G.</given-names>
            <surname>Davis</surname>
          </string-name>
          .
          <year>2020</year>
          .
          <article-title>Measuring the diversity of recommendations: a preference-aware approach for evaluating and adjusting diversity</article-title>
          .
          <source>Knowledge and Information Systems</source>
          <volume>62</volume>
          ,
          <issue>2</issue>
          (
          <issue>01</issue>
          <year>Feb 2020</year>
          ),
          <fpage>787</fpage>
          -
          <lpage>811</lpage>
          . https://doi.org/10.1007/s10115-019-01371-0
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <surname>Tomas</surname>
            <given-names>Mikolov</given-names>
          </string-name>
          , Kai Chen, Greg Corrado, and
          <string-name>
            <given-names>Jefrey</given-names>
            <surname>Dean</surname>
          </string-name>
          .
          <year>2013</year>
          .
          <article-title>Eficient Estimation of Word Representations in Vector Space</article-title>
          .
          <source>In 1st International Conference on Learning Representations, ICLR</source>
          <year>2013</year>
          , Scottsdale, Arizona, USA, May 2-
          <issue>4</issue>
          ,
          <year>2013</year>
          , Workshop Track Proceedings, Yoshua Bengio and Yann LeCun (Eds.). http://arxiv.org/abs/1301.3781
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <surname>Tien</surname>
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Nguyen</surname>
          </string-name>
          ,
          <string-name>
            <surname>Pik-Mai Hui</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Maxwell</surname>
            <given-names>Harper</given-names>
          </string-name>
          , Loren Terveen, and
          <string-name>
            <surname>Joseph</surname>
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Konstan</surname>
          </string-name>
          .
          <year>2014</year>
          .
          <article-title>Exploring the Filter Bubble: The Efect of Using Recommender Systems on Content Diversity</article-title>
          .
          <source>In Proceedings of the 23rd International Conference on World Wide Web (Seoul</source>
          , Korea) (
          <source>WWW '14)</source>
          .
          <article-title>Association for Computing Machinery</article-title>
          , New York, NY, USA,
          <fpage>677</fpage>
          -
          <lpage>686</lpage>
          . https://doi.org/10.1145/2566486.2568012
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>Eli</given-names>
            <surname>Pariser</surname>
          </string-name>
          .
          <year>2011</year>
          .
          <article-title>The Filter Bubble: What the Internet Is Hiding from You</article-title>
          . Penguin Group , The.
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <surname>Young-Jin</surname>
            <given-names>Park</given-names>
          </string-name>
          , Kyuyong Shin, and
          <string-name>
            <surname>Kyung-Min Kim</surname>
          </string-name>
          .
          <year>2020</year>
          .
          <article-title>Hop Sampling: A Simple Regularized Graph Learning for Non-Stationary Environments</article-title>
          . arXiv preprint arXiv:
          <year>2006</year>
          .
          <volume>14897</volume>
          (
          <year>2020</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [24]
          <string-name>
            <surname>Bryan</surname>
            <given-names>Perozzi</given-names>
          </string-name>
          , Rami Al-Rfou, and
          <string-name>
            <given-names>Steven</given-names>
            <surname>Skiena</surname>
          </string-name>
          .
          <year>2014</year>
          .
          <article-title>DeepWalk: Online Learning of Social Representations</article-title>
          .
          <source>In Proceedings of the 20th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining</source>
          (New York, New York, USA) (
          <source>KDD '14)</source>
          .
          <article-title>Association for Computing Machinery</article-title>
          , New York, NY, USA,
          <fpage>701</fpage>
          -
          <lpage>710</lpage>
          . https://doi.org/10.1145/2623330.2623732
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          [25]
          <string-name>
            <surname>Kyuyong</surname>
            <given-names>Shin</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Young-Jin</surname>
            <given-names>Park</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kyung-Min Kim</surname>
            , and
            <given-names>Sunyoung</given-names>
          </string-name>
          <string-name>
            <surname>Kwon</surname>
          </string-name>
          .
          <year>2020</year>
          .
          <article-title>Multi-Manifold Learning for Large-Scale Targeted Advertising System</article-title>
          . arXiv preprint arXiv:
          <year>2007</year>
          .
          <volume>02334</volume>
          (
          <year>2020</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          [26] Rianne van den Berg, Thomas N Kipf, and
          <string-name>
            <given-names>Max</given-names>
            <surname>Welling</surname>
          </string-name>
          .
          <year>2017</year>
          .
          <article-title>Graph Convolutional Matrix Completion</article-title>
          .
          <source>arXiv preprint arXiv:1706.02263</source>
          (
          <year>2017</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          [27]
          <string-name>
            <given-names>Saúl</given-names>
            <surname>Vargas</surname>
          </string-name>
          .
          <year>2014</year>
          .
          <article-title>Novelty and Diversity Enhancement and Evaluation in Recommender Systems and Information Retrieval</article-title>
          .
          <source>In Proceedings of the 37th International ACM SIGIR Conference on Research Development in Information Retrieval</source>
          (Gold Coast, Queensland, Australia) (
          <source>SIGIR '14)</source>
          .
          <article-title>Association for Computing Machinery</article-title>
          , New York, NY, USA,
          <volume>1281</volume>
          . https://doi.org/10.1145/2600428.2610382
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          [28]
          <string-name>
            <surname>Petar</surname>
            <given-names>Veličković</given-names>
          </string-name>
          , Guillem Cucurull, Arantxa Casanova, Adriana Romero, Pietro Liò, and
          <string-name>
            <given-names>Yoshua</given-names>
            <surname>Bengio</surname>
          </string-name>
          .
          <year>2018</year>
          .
          <article-title>Graph Attention Networks</article-title>
          .
          <source>International Conference on Learning Representations</source>
          (
          <year>2018</year>
          ). https://openreview.net/forum?id= rJXMpikCZ
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          [29]
          <string-name>
            <surname>Jesse</surname>
            <given-names>Vig</given-names>
          </string-name>
          , Shilad Sen,
          <string-name>
            <given-names>and John</given-names>
            <surname>Riedl</surname>
          </string-name>
          .
          <year>2012</year>
          .
          <article-title>The Tag Genome: Encoding Community Knowledge to Support Novel Interaction</article-title>
          .
          <source>ACM Trans. Interact. Intell. Syst. 2</source>
          ,
          <issue>3</issue>
          ,
          <string-name>
            <surname>Article 13</surname>
          </string-name>
          (
          <issue>Sept</issue>
          .
          <year>2012</year>
          ),
          <volume>44</volume>
          pages. https://doi.org/10.1145/2362394.2362395
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          [30]
          <string-name>
            <surname>Hongwei</surname>
            <given-names>Wang</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Miao Zhao</surname>
            , Xing Xie,
            <given-names>Wenjie</given-names>
          </string-name>
          <string-name>
            <surname>Li</surname>
            ,
            <given-names>and Minyi</given-names>
          </string-name>
          <string-name>
            <surname>Guo</surname>
          </string-name>
          .
          <year>2019</year>
          .
          <article-title>Knowledge Graph Convolutional Networks for Recommender Systems</article-title>
          . In The World Wide Web Conference (San Francisco, CA, USA) (
          <source>WWW '19)</source>
          .
          <article-title>Association for Computing Machinery</article-title>
          , New York, NY, USA,
          <fpage>3307</fpage>
          -
          <lpage>3313</lpage>
          . https://doi.org/10.1145/3308558. 3313417
        </mixed-citation>
      </ref>
      <ref id="ref31">
        <mixed-citation>
          [31]
          <string-name>
            <surname>Guanjie</surname>
            <given-names>Zheng</given-names>
          </string-name>
          , Fuzheng Zhang, Zihan Zheng,
          <string-name>
            <surname>Yang</surname>
            <given-names>Xiang</given-names>
          </string-name>
          , Nicholas Jing Yuan, Xing Xie, and
          <string-name>
            <given-names>Zhenhui</given-names>
            <surname>Li</surname>
          </string-name>
          .
          <year>2018</year>
          .
          <article-title>DRN: A Deep Reinforcement Learning Framework for News Recommendation</article-title>
          .
          <source>In Proceedings of the 2018 World Wide Web Conference (Lyon, France) (WWW '18)</source>
          .
          <source>International World Wide Web Conferences Steering Committee, Republic and Canton of Geneva, CHE</source>
          ,
          <fpage>167</fpage>
          -
          <lpage>176</lpage>
          . https://doi.org/10. 1145/3178876.3185994
        </mixed-citation>
      </ref>
      <ref id="ref32">
        <mixed-citation>
          [32]
          <string-name>
            <surname>Guorui</surname>
            <given-names>Zhou</given-names>
          </string-name>
          , Xiaoqiang Zhu, Chenru Song, Ying Fan, Han Zhu, Xiao Ma, Yanghui Yan, Junqi Jin,
          <string-name>
            <given-names>Han</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>and Kun</given-names>
            <surname>Gai</surname>
          </string-name>
          .
          <year>2018</year>
          .
          <article-title>Deep Interest Network for ClickThrough Rate Prediction</article-title>
          .
          <source>In Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery Data Mining (London, United Kingdom) (KDD '18)</source>
          .
          <article-title>Association for Computing Machinery</article-title>
          , New York, NY, USA,
          <fpage>1059</fpage>
          -
          <lpage>1068</lpage>
          . https://doi.org/10.1145/3219819.3219823
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>