<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Online ranking prediction in non-stationary environments</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Erzsébet Frigó</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>András A. Benczúr</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Róbert Pálovics Domokos Kelen Levente Kocsis Institute for Computer Science and Control Hungarian Academy of Sciences</institution>
          ,
          <addr-line>MTA SZTAKI</addr-line>
        </aff>
      </contrib-group>
      <abstract>
        <p>Recommender systems have to serve in online environments which can be highly non-stationary.1. Traditional recommender algorithms may periodically rebuild their models, but they cannot adjust to quick changes in trends caused by timely information. In our experiments, we observe that even a simple, but online trained recommender model can perform significantly better than its batch version. We investigate online learning based recommender algorithms that can eficiently handle non-stationary data sets. We evaluate our models over seven publicly available data sets. Our experiments are available as an open source project2.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>INTRODUCTION</title>
      <p>
        The research of recommender systems became popular since the
Netflix prize [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. As an efect of the competition, batch rating
prediction is considered the standard recommender evaluation task,
with one part of the data used for model training, and the other
for evaluation. In contrast to the Netflix Prize task, recommender
systems are typically not required to predict ratings, rather they
are required to present a ranked top list of relevant items for the
user. Also, most users give no explicit ratings and we have to infer
their preferences from implicit feedback [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ]. Most importantly,
users request one or a few items at a time, and may get exposed
to new information that can change their needs and taste before
they return to the service the next time. In a real application, top
item recommendation by online learning is therefore more relevant
than batch rating prediction usually. There is vast literature on the
batch performance of collaborative filtering algorithms, while the
more realistic sequential or online evaluation received much less
attention. Information on recommendation performance is scarce
when collaborative filtering is evaluated online, sequentially.
      </p>
      <p>
        In this work we intend to consider top recommendation in highly
non-stationary environments similarly to the conceptually
simpler classification task [
        <xref ref-type="bibr" rid="ref22">22</xref>
        ]. Our goal is to promptly update
recommender models after user interactions using online learning
methods.
      </p>
      <p>
        There are some indications [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ] that the performance gains in
batch experiments do not necessary carry over to online
learning environments. One result of the Netflix prize is that matrix
factorization [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ] techniques provide strong baseline results on
non-temporal data sets. In our work we compare the online and
batch versions of the same matrix factorization algorithm. Due
to the highly non-stationary properties of the data set, we apply
1Copyright©2017 for this paper by its authors. Copying permitted for private and
academic purposes.
2https://github.com/rpalovics/Alpenglow
incremental stochastic gradient descent (SGD) based matrix
factorization. We show that the online version strongly outperforms the
batch version on non-stationary data sets.
      </p>
      <p>
        In [
        <xref ref-type="bibr" rid="ref23">23</xref>
        ], the online DCG measure is defined as a side result to
track the performance of recommender algorithms over time. In
our present experiments, we heavily rely on online DCG, introduce
another measure, the online MRR, and show their importance for
evaluating streaming recommenders. Summarized, our main results
are the following:
• We provide some measures suited to non-stationary
environments that are predictive of which algorithm would work
well on a particular data set.
• We show that simpler algorithms that can be updated online—
and therefore use the most recent data as well—often perform
better than more complex algorithms that can only be
updated periodically. This is in contrast to the case when the
testing is batch or the data is stationary.
• We also show that even though initialization by large batch
models may be required for optimal performance, online
matrix factorization combined with a simple sampling strategy
can perform comparably to more computationally intensive
models.
1.1
      </p>
    </sec>
    <sec id="sec-2">
      <title>Related Results</title>
      <p>
        The dificulty of evaluating streaming recommenders was first
mentioned in [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ], although they evaluated models by ofline
training and testing split. Ideas for online evaluation metrics appeared
ifrst in [
        <xref ref-type="bibr" rid="ref23 ref24 ref33">23, 24, 33</xref>
        ]. In [
        <xref ref-type="bibr" rid="ref33">33</xref>
        ], incremental algorithms are evaluated
using recall, which is a batch quality measure. In the other two
results, online evaluation appears as a side result of investigating
social influence. As the starting point of our results, the last paper
notices that some online methods perform better than their batch
counterpart.
      </p>
      <p>
        Since our goal is to recommend diferent items at diferent times,
our evaluation must be based on the quality of the top list produced
by the recommender. This so-called top-k recommender task is
known to be hard [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]. A recent result on evaluating top-k
recommenders is found in [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ].
      </p>
      <p>
        In our experiments, we apply widely used methods of matrix
factorization. The highly successful gradient descent matrix
factorization recommender is described among others by [
        <xref ref-type="bibr" rid="ref11 ref30">11, 30</xref>
        ].
Note that even though typically performing marginally better than
gradient descent, we do not investigate alternating least squares
[
        <xref ref-type="bibr" rid="ref18 ref27">18, 27</xref>
        ], since it is an inherently batch algorithm that cannot be
easily adapted to the online setting.
      </p>
      <p>
        We also evaluate basic item-to-item recommenders. Recent
works [
        <xref ref-type="bibr" rid="ref26">26</xref>
        ] show that most industry recommendation tasks are
item-to-item, since the only information available is the present
u
time
user session. The first item-to-item recommender methods [
        <xref ref-type="bibr" rid="ref20 ref29">20, 29</xref>
        ]
were using similarity information to directly find nearest neighbor
[
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] transactions. Nearest neighbor methods were criticized for two
reasons. First, similarity metrics typically have no mathematical
justification. Second, the confidence of the similarity values is
often not involved when finding the nearest neighbor, which leads
to overfitting in sparse cases. A new method to give session
recommendations was described in [
        <xref ref-type="bibr" rid="ref28">28</xref>
        ] by Rendle et al. that models
the users by factorizing personal Markov chains. Koenigstein and
Koren [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ] improve the basic transition model by computing
latent item factors to represent all items in a low dimensional space.
While certainly giving performance gains in batch testing, these
algorithms are overly costly to be updated online.
      </p>
      <p>
        Finally we mention that item-to-item recommendation was also
considered as a special context aware recommendation problem. In
[
        <xref ref-type="bibr" rid="ref14">14</xref>
        ] sequentiality as context is handled using pairwise associations
as features in an alternating least squares model by Hidasi et al.
They mention that they face the sparsity problem in setting
minimum support, confidence and lift of the associations and they use
the category of last purchased item as fallback. In a follow-up result
[
        <xref ref-type="bibr" rid="ref15">15</xref>
        ], they use the same context-aware ALS algorithm, however they
only consider seasonality as context in that paper. Our result can
be used independently of the ALS based methods and can easily be
combined with user personalization. Most recently, context
information learning was also performed by recurrent neural networks
[
        <xref ref-type="bibr" rid="ref13">13</xref>
        ].
2
      </p>
    </sec>
    <sec id="sec-3">
      <title>TEMPORAL EVALUATION</title>
      <p>
        In the implicit top-k recommendation task [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ], the goal is not to
rate individual items, but to recommend the best candidates. In a
time sensitive or online recommender that potentially re-trains the
model after each new item, we have to generate a new top list for
every recommendation request. The online top-k task is therefore
diferent from the standard recommender evaluation settings, since
there is always only a single relevant item. In an online setting, as
seen in Figure 1, we
(1) request a top-k recommendation for the active user,
(2) evaluate against the single relevant item,
(3) train the model on the revealed user-item pair.
      </p>
      <p>
        Next we introduce natural evaluation techniques for the online
ranking prediction problem, extending the methods of [
        <xref ref-type="bibr" rid="ref23">23</xref>
        ]. In our
setting, model training and evaluation happen simultaneously,
iterating over the data set only once, in chronological order. Whenever
we see a new user-item pair, we assume that the user becomes
active and requests a recommendation. The recommendation is
online, hence it may depend on all events before the exact time of
the interaction. If a user u views item i at time t , our models predict
a score rˆ(u, i ′, t ) for each item i ′ that appears in the data so far, and
recommend the k items with the largest values from those that u
has not seen before.
      </p>
      <p>
        One possible measure for the quality of a recommended top list
could be precision and recall [
        <xref ref-type="bibr" rid="ref35 ref36">35, 36</xref>
        ]. Note that we evaluate against
a single item. Both the number of relevant (1) and the number of
retrieved (K ) items are fixed. Precision is 1/K if we retrieve the
single item viewed and 0 otherwise. Recall is 0 if we do not retrieve
the single relevant item and 1 otherwise. Hence up to a constant
factor K , precision and recall are identical and are binary indicators
of whether the item viewed is on the recommended list.
      </p>
      <p>
        Recently, measures other than precision and recall became
preferred for measuring the quality of top-k recommendation [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. The
most common measure is NDCG, the normalized version of the
discounted cumulative gain (DCG). Since the decrease of DCG as
the function of the rank is smoother than the decrease of precision
and recall, it is more advantageous, since we have a large number of
items of potential interest to each user. Our choice is in accordance
with the observations in [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] as well.
      </p>
      <p>DCG computed individually for each event and averaged in time
is an appropriate measure for real-time recommender evaluation.
If i is the next consumed item by the user, the online DCG@K is
defined as the following function of the rank of i returned by the
recommender system:</p>
      <p>DCG@K(i ) = 0log2 (ran1k(i ) + 1) iofthraenrwk(iis)e&gt;. K ; (1)

The overall evaluation of a model is the average of the DCG@K
values over all events of the evaluation period. Note that in our
unusual temporal setting of DCG evaluation, there is a single
relevant item and hence no normalization is needed as opposed to the
original NDCG measure.</p>
      <p>
        Another possible ranking evaluation measure with a natural
online extension is Mean Reciprocal Rank (MRR), the average of
1/rank of the first relevant item [
        <xref ref-type="bibr" rid="ref34">34</xref>
        ]. In online evaluation, there
is a single relevant item i, therefore online MRR@K is equal to
1/rank(i ) if rank(i ) ≤ K and is 0 otherwise. Hence the diference
between online DCG and MRR the rank discount function, which
is reciprocal for MRR and inverse logarithmic for DCG.
3
      </p>
    </sec>
    <sec id="sec-4">
      <title>ALGORITHMS</title>
      <p>In this section, we describe the algorithms used in our experiments,
both in batch and online setting. The simplest algorithms, for
example the count based ones, are online by nature. Most of the
algorithms discussed have more or less natural online counterparts.
Our goal is to include a variety of count, nearest neighbor, session
and matrix factorization based methods to see how these algorithms
can adapt to non-stationary environments. Many of the algorithms
will have parameters to handle decay in time, i.e. forget older events
and emphasize new ones.</p>
      <p>
        Online recommenders seem more restricted, since they may
not iterate over the data several times, hence we would expect
inferior quality. Online methods however have the advantage of
putting much more emphasis on recent events. In an online setting
[
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], the model needs to be retrained after each new event and
hence reiterations over the earlier parts of the data is ruled out. We
may implement an online recommender algorithm by allowing a
single iteration over the training data only, with this single iteration
processing the events in the order of time.
3.1
      </p>
    </sec>
    <sec id="sec-5">
      <title>Popularity (POP)</title>
      <p>In this non-personalized method, we recommend the most popular
items. The method may be both batch and online depending on the
granularity of the item frequency measurement update. For each
data set, we selected the best time frame, which had a wide range
of variety between 10 minutes and a year.
3.2</p>
    </sec>
    <sec id="sec-6">
      <title>Time sensitive nearest neighbor (t-NN)</title>
      <p>
        One of the earliest and most popular collaborative filtering
algorithms in practice is the item-based nearest neighbor [
        <xref ref-type="bibr" rid="ref29">29</xref>
        ]. For
these algorithms similarity scores are computed between item pairs
based on the co-occurrence of the pairs in the preference of users.
Non-stationarity of the data can be accounted for e.g. with the
introduction of a time-decay [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ].
      </p>
      <p>Describing the algorithm more formally, let us denote by Ui the
set of users that visited item i, by Iu the set of items visited by user
u, and by sui the index of item i in the sequence of interactions of
user u. The frequency based time-weighted similarity function is
defined by
sim(j, i ) =</p>
      <p>Pu ∈Uj ∩Ui f (sui − su j )</p>
      <p>Uj
,
where f (τ ) = γ τ is the time decaying function. For non-stationary
data we sum only over users that visit item j before item i, setting
f (τ ) = 0 if τ &lt; 0. For stationary data the absolute value of τ is used.
The score assigned to item i for user u is
score (u, i ) = X f |Iu | − su j sim(j, i ).</p>
      <p>
        j ∈Iu
The model is represented by the similarity scores. Since computing
the model is time consuming, it is done periodically. Moreover, only
the most similar items are stored for each item. When the prediction
scores are computed for a particular user, all items visited by the
user can be considered, including the most recent ones. Hence, the
algorithm can be considered semi-online in that it uses the most
recent interactions of the current user, but not of the other users.
We note that the time decay function is used here to quantify the
strength of connection between pairs of items depending on how
closely are located in the sequence of a user, and not as a way to
forget old data as in [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ].
3.3
      </p>
    </sec>
    <sec id="sec-7">
      <title>Item-to-item transition model (TRANS)</title>
      <p>A simple algorithm that focuses on the sequence of items a user
has visited is one that records how often users visited item i after
visiting another item j. This can be viewed as particular form of
the item-to-item nearest neighbor with a time decay function that
is non-zero only for the immediately preceding item. While the
algorithm is more simplistic, it is fast to update the transition
frequencies after each interaction, thus all recent information is taken
into account.
(2)
(3)
sampling set S
),i ),i ),i
(u (u (u</p>
      <p>time
instances for learning
(),i
u
(4)
(5)
positive samples</p>
      <p>
        negative samples
One of our methods is the popular gradient descent matrix
factorization recommender described among others by [
        <xref ref-type="bibr" rid="ref11 ref30">11, 30</xref>
        ]. The original
algorithm builds a batch model by iterating over a fixed training
data set a certain I number of times in random order,
performing stochastic gradient descent, until convergence. The algorithm
builds F latent factors, Puf for user u and Qif for item i, to predict
the rating by
rˆ(u, i ) =
      </p>
      <p>F
X Puf Qif .</p>
      <p>f =1
Given an actual rating r (u, i ), the factors Puf are updated by
regularized gradient descent as</p>
      <p>(t +1) = Pu(tf) + η · (r (u, i ) − rˆ(u, i )) · Qif − ηλPu(tf),</p>
      <p>Puf
where η is the learning rate and λ is the regularization coeficient .
Item factors Qif are updated similarly.</p>
      <p>
        In our tasks, all of the ratings are implicit. Whenever we observe
a user-item interaction (u, i ), we may only infer positive feedback
r (u, i ) = 1. We follow the approach of [
        <xref ref-type="bibr" rid="ref27">27</xref>
        ] by treating all unknown
elements (u ′, i ′) of the matrix as negative feedback, i.e we assume
r (u ′, i ′) = 0. In order to avoid the model fitting to r ≡ 1, we
generate negative samples for each item interaction. In other words,
after training on (u, i ), we train according to equation (5) for a
ifxed negative rate nRate number of (u, i ′) pairs. Item i ′ is drawn
uniformly nRate times for user u from the set of items that the u
has not interacted with.
      </p>
      <p>Online matrix factorization takes the same steps, including
negative sample generation, but strictly in the order of the events.
In the update equation (5), we may consider the superscript t as
time, and process interactions (u, i, t ) in the order of t . For negative
samples, we generate items i ′ such that no (u, i ′, t ′) interaction
exists with t ′ ≤ t .</p>
      <p>A natural option to combine the advantages of the batch and the
online model is to periodically (e.g. weekly) build batch factors P
and Q and continue from the same P and Q with the online gradient
descent method. We call this method batch &amp; online.
3.5</p>
    </sec>
    <sec id="sec-8">
      <title>Sampling from the past for matrix factorization</title>
      <p>While batch &amp; online matrix factorization naturally combines the
advantages of the two methods, the periodic re-training of the
batch model may be computationally ineficient. We propose an
online, eficient technique that approximates the batch and online
combination. Similar to the online MF model, we only allow a
single iteration for the model in a temporal order. However, for each
interaction, we generate not only negative but also positive samples.
The positive samples are selected randomly from past interactions,
i.e. we allow the model to re-learn the past. We generate pRate
positive samples along with nRate negative samples, hence for t
events, we only take (1 + nRate + pRate) · t gradient steps.</p>
      <p>The samples are not drawn uniformly from the past, but selected
randomly from pool S with maximum size s. This avoids
oversampling interactions that happened at the beginning of the data set.
More specifically, as seen in Fig. 2, for each observed new training
instance, we
• update the model by pRate random samples from S,
• delete the selected samples from S if |S | &gt; s,
• and add the new instance pRate times to S.
3.6</p>
    </sec>
    <sec id="sec-9">
      <title>Asymmetric Matrix Factorization (AMF)</title>
      <p>
        Asymmetric Matrix Factorization [
        <xref ref-type="bibr" rid="ref25">25</xref>
        ] computes item-to-item
similarity with the help of latent vectors for items. In contrast to
Section 3.4, both latent matrices P and Q correspond to items. Using
the notations and time decaying function from Section 3.2, the score
assigned to item i for user u is:
score (u, i ) = X
f |Iu | − su j Pj Qi .
      </p>
      <p>(6)
j ∈Iu
Sampling negative instances and updating the latent vectors online
using stochastic gradient descent can be done in a similar way to
the one described in Section 3.4.
4</p>
    </sec>
    <sec id="sec-10">
      <title>DATA</title>
      <p>We compare the performance of the batch and the online algorithms
of Section 3 over seven data sets described next. For each data set,
we discard items that appear less than ten times in interactions. We
apply no filtering for the users.</p>
      <p>
        Twitter. We use 400M tweets over four months between
February 1 and May 30, 2012 from a crawl introduced in [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ]. We
recommend new hashtags that the user has not used before.
      </p>
      <p>
        Last.fm We recommend artists and tracks in the 30Music data
set [
        <xref ref-type="bibr" rid="ref31">31</xref>
        ]. New tracks appear more frequently than new artists. We
expect that user-track events are highly non-stationary, more than
user-artists events.
      </p>
      <p>
        MovieLens data sets [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ], first released in 1998, contain
timestamped explicit movie ratings. We use the ML10M data set,
consisting of 10,000,064 ratings of 69,878 users for 10,681 movies. In our
experiments, we consider these records as implicit interactions.
      </p>
      <p>
        Amazon data sets [
        <xref ref-type="bibr" rid="ref21">21</xref>
        ] are browsing and co-purchasing data
crawled from Amazon. We use four categories, Books, Electronics
as well as Movies and TV .
      </p>
    </sec>
    <sec id="sec-11">
      <title>4.1 Statistics</title>
      <p>Before turning to explaining the findings, we show two statistics
that highlight the non-stationary properties of the data sets.</p>
      <p>
        Burstiness. The inter-event time (Fig. 3) is the time elapsed
between consecutive events of the same user. In accordance with
several results [
        <xref ref-type="bibr" rid="ref16 ref3 ref32">3, 16, 32</xref>
        ], the distribution is highly skewed for
all data, which show their burstiness in the approximate order
from strongest to weakest as Twitter, Amazon, Last.fm and finally
MovieLens (Fig. 3, right to left).
      </p>
      <p>Item-to-item transitions. Users of online music services very
often follow playlists, which result in artificially very strong
performance of the item-to-item methods. When listening to playlists,
users do not actually request recommendations and hence
evaluating user sequences for recommendation quality is somewhat
artificial. Next we examine if the data sets contain trivial
item-toitem transitions by computing the item-to-item transition matrix,
i.e. the number of times j followed i in the music listening session of
the same user. For each item i we count the least number of unique
i → j item transitions needed to cover 90% of the item’s transitions.
In Figure 3 we plot the fraction of these against the total number of
unique i → j transitions. We observe that for the Last.fm track data,
the MovieLens data set, and for the Twitter data lower fractions
are possible, hence these data sets may involve trivial transitions.</p>
      <p>Shufling the temporal order. In order to assess the
importance of the time dimension in the streaming recommendation task,
we compare our results over the same data sets but with randomly
shufled order of time. In other words, we evaluate over the shufled
stationary variant of each data, where we expect that the temporal
efects disappear.
5</p>
    </sec>
    <sec id="sec-12">
      <title>RESULTS</title>
      <p>In this section, we describe our results, including the performance
of the algorithms described in Section 3 and our experiments with
matrix factorization training strategies. In Table 1, we summarize
the performance of our models by online DCG. Online MRR behaves
almost identical and is not shown for space limitations.</p>
      <p>Twitter data is very sparse and bursty with the highest fraction
of small inter-events (Fig. 3). As a result, factorization methods
perform worse and popularity models perform well. The Last.fm track
data set is similar, but it incorporates playlist and album listening
data, hence transition and similarity models perform better than
factorization and popularity. The number of items in the Last.fm
artist data are significantly lower than for the track data: as a result
factorization gets close to transition and similarity models on this
data set . In Fig. 3 we can observe that the Last.fm track data, the
MovieLens data set, and the Twitter data contain several predictable
item-to-item transitions. This observation is consistent with
performance of the transition model, as it provides the strongest DCG
values in case of these three data sets.</p>
      <p>For batch rating prediction, MF is known in the literature to
perform very strong. For ranking prediction, however, our results show
the superiority of item-to-item nearest neighbor methods.
Furthermore, performance of batch MF models drop considerably when the
data is non-stationary. For example, for MovieLens, MF is the best
performing algorithm after shufling, but the worst for the original
temporal order. Popularity, the simplest algorithm, performs well
for many data sets in non-stationary setting, although the strong
performance is mainly due to users with few interactions. AMF
performs between MF and NN, slightly better for non-stationary
data. The transition model is constantly inferior to NN, but despite
its simplicity, it is competitive with more complex factorization
models for non-stationary data.
5.1</p>
    </sec>
    <sec id="sec-13">
      <title>Online training strategies for MF</title>
      <p>The high performance of transition models indicates the existence
of trivial item-to-item transitions. In music data, such as the Last.fm
data set, the listening of predefined playlists or albums results
in predictable item-to-item transitions. These are easy to predict,
however they are not valuable for the actual user, therefore they bias
recommender evaluation. We filtered the Last.fm artist data before
comparing diferent factor model training strategies to reduce the
above efects. We only considered records after at least 30 minutes
of user inactivity, i.e. the start of each user session. Statistics of the
ifltered 30M data are shown in Fig. 4.</p>
      <p>We compared multiple diferent learning strategies of MF
models: batch, online, batch &amp; online, and sampling methods. In the
experiments we use the following best parameters. We train batch
MF weekly with η = 0.05, nRate = 69 and 9 iterations in random
order. For online MF we set η = 0.2, nRate = 99. The parameters are
the same for batch &amp; online. For the sampling model, we use lower
learning rate η = 0.05 and nRate = 39. Best results are achieved
by generating pRate = 3 positive samples from the past for each
record with pool size 3,000.
weekly events
40</p>
      <p>50
weekly new users
weekly new items
0
10
20 30
time (weeks)
0
0
10
20 30
time (weeks)
40
50</p>
      <p>Twitter all
Twitter from 10th</p>
      <p>Last.fm track
Last.fm artist
MovieLens10M</p>
      <p>Books all
n Books from 10th
zo Electronics
am Electronics from 10th
A Movies</p>
      <p>Movies from 10th</p>
      <p>In Fig. 5 we plot the average weekly DCG over a one year period.
While sampling appears to keep up with batch &amp; online in the
beginning, its performance drops in the second part. If we start
each factor model from the batch model of week 18, sampling
produces similar results to batch &amp; online. In Fig. 6, as a function
of k, we start with the batch model of week k. Then, we only use
online updates with or without further sampling. It can be seen
that sampling performs roughly the same as batch &amp; online for
k &gt; 10. Note that the number of weekly incoming new users and
items drops after the first 10 weeks as seen in Fig. 4. The single
iteration online model produces a comparable, but slightly worse
results than the sampling version.</p>
      <p>In summary, in a non-stationary environment, multiple passes
(i.e. batch models) are required over the data to fully incorporate
many new users and items into the system. However, if a large
component of the user-item matrix is previously learned by a robust
batch model, a simple online sampling algorithm is suficient to keep
up in performance with periodic batch re-training, while requiring
only 3 additional positive samples from the past per iteration and
thus being much more eficient.
6</p>
    </sec>
    <sec id="sec-14">
      <title>CONCLUSIONS</title>
      <p>Despite the fact that a practical recommender system processes
events and generates top lists in an online sequence, the literature
payed relative little attention to designing and evaluating online
0.040
5
4
20.038
5
s
k
e
e
nw0.036
e
e
w
t
e
b0.034
G
C
D
e
g0.032
a
r
e
v
a
0.030
batch
batch &amp; online
sampling started from batch
online started from batch
5</p>
      <p>10
starting week k
15
20
learning methods in highly non-stationary environments. We
proposed and evaluated a variety of algorithms as well as measures that
are predictive of which algorithm would work well on a particular
data. We showed that simpler algorithms that can use most recent
data by updating their models online perform better than more
complex algorithms that can be updated only periodically. We also
showed that sampling from past events may completely replace
batch modeling needs in a real time recommender system, thus
reducing latency. We released our code as an open source project3.</p>
    </sec>
    <sec id="sec-15">
      <title>REFERENCES</title>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>J.</given-names>
            <surname>Abernethy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Canini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Langford</surname>
          </string-name>
          ,
          <article-title>and</article-title>
          <string-name>
            <given-names>A.</given-names>
            <surname>Simma</surname>
          </string-name>
          .
          <article-title>Online collaborative filtering</article-title>
          . University of California at Berkeley,
          <source>Tech. Rep</source>
          ,
          <year>2007</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>A.</given-names>
            <surname>Al-Maskari</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Sanderson</surname>
          </string-name>
          , and
          <string-name>
            <given-names>P.</given-names>
            <surname>Clough</surname>
          </string-name>
          .
          <article-title>The relationship between ir effectiveness measures and user satisfaction</article-title>
          .
          <source>In Proc. SIGIR</source>
          , pp.
          <fpage>773</fpage>
          -
          <lpage>774</lpage>
          . ACM,
          <year>2007</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>Albert-Laszlo</given-names>
            <surname>Barabasi</surname>
          </string-name>
          .
          <article-title>The origin of bursts and heavy tails in human dynamics</article-title>
          .
          <source>Nature</source>
          ,
          <volume>435</volume>
          (
          <issue>7039</issue>
          ):
          <fpage>207</fpage>
          -
          <lpage>211</lpage>
          ,
          <year>2005</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>J.</given-names>
            <surname>Bennett</surname>
          </string-name>
          and
          <string-name>
            <given-names>S.</given-names>
            <surname>Lanning</surname>
          </string-name>
          .
          <article-title>The netflix prize</article-title>
          .
          <source>In KDD Cup and Workshop in conjunction with KDD</source>
          ,
          <year>2007</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>N.</given-names>
            <surname>Cesa-Bianchi</surname>
          </string-name>
          and
          <string-name>
            <given-names>G.</given-names>
            <surname>Lugosi</surname>
          </string-name>
          .
          <article-title>Prediction, learning, and games</article-title>
          . Cambridge Univ Pr,
          <year>2006</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>P.</given-names>
            <surname>Cremonesi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Koren</surname>
          </string-name>
          , and
          <string-name>
            <given-names>R.</given-names>
            <surname>Turrin</surname>
          </string-name>
          .
          <article-title>Performance of recommender algorithms on top-n recommendation tasks</article-title>
          .
          <source>In Proc. RecSys</source>
          , pages
          <fpage>39</fpage>
          -
          <lpage>46</lpage>
          . ACM,
          <year>2010</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>M.</given-names>
            <surname>Deshpande</surname>
          </string-name>
          and
          <string-name>
            <given-names>G.</given-names>
            <surname>Karypis</surname>
          </string-name>
          .
          <article-title>Item-based top-n recommendation algorithms</article-title>
          .
          <source>ACM TOIS</source>
          ,
          <volume>22</volume>
          (
          <issue>1</issue>
          ):
          <fpage>143</fpage>
          -
          <lpage>177</lpage>
          ,
          <year>2004</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>C.</given-names>
            <surname>Desrosiers</surname>
          </string-name>
          and
          <string-name>
            <given-names>G.</given-names>
            <surname>Karypis</surname>
          </string-name>
          .
          <article-title>A comprehensive survey of neighborhood-based recommendation methods</article-title>
          .
          <source>In Recommender systems handbook</source>
          , pages
          <fpage>107</fpage>
          -
          <lpage>144</lpage>
          . Springer,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Ding</surname>
          </string-name>
          and
          <string-name>
            <given-names>X.</given-names>
            <surname>Li</surname>
          </string-name>
          .
          <article-title>Time weight collaborative filtering</article-title>
          .
          <source>In Proc. CIKM</source>
          , pages
          <fpage>485</fpage>
          -
          <lpage>492</lpage>
          . ACM,
          <year>2005</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>L.</given-names>
            <surname>Dobos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Szule</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Bodnar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Hanyecz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Sebok</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Kondor</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Kallus</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Stéger</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Csabai</surname>
          </string-name>
          , and
          <string-name>
            <given-names>G.</given-names>
            <surname>Vattay.</surname>
          </string-name>
          <article-title>A multi-terabyte relational database for geo-tagged social network data</article-title>
          .
          <source>In Proc. CogInfoCom</source>
          , pages
          <fpage>289</fpage>
          -
          <lpage>294</lpage>
          . IEEE,
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>S.</given-names>
            <surname>Funk</surname>
          </string-name>
          .
          <article-title>Netflix update: Try this at home</article-title>
          . http://sifter.org/˜simon/journal/20061211.html,
          <year>2006</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>F. M.</given-names>
            <surname>Harper</surname>
          </string-name>
          and
          <string-name>
            <given-names>J. A.</given-names>
            <surname>Konstan</surname>
          </string-name>
          .
          <article-title>The movielens datasets: History and context</article-title>
          .
          <source>ACM TiiS</source>
          ,
          <volume>5</volume>
          (
          <issue>4</issue>
          ):
          <fpage>19</fpage>
          ,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>B.</given-names>
            <surname>Hidasi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Karatzoglou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Baltrunas</surname>
          </string-name>
          , and
          <string-name>
            <given-names>D.</given-names>
            <surname>Tikk</surname>
          </string-name>
          .
          <article-title>Session-based recommendations with recurrent neural networks</article-title>
          .
          <source>In ICLR</source>
          ,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>B.</given-names>
            <surname>Hidasi</surname>
          </string-name>
          and
          <string-name>
            <given-names>D.</given-names>
            <surname>Tikk</surname>
          </string-name>
          .
          <article-title>Fast als-based tensor factorization for context-aware recommendation from implicit feedback</article-title>
          .
          <source>In Machine Learning and Knowledge Discovery in Databases</source>
          , pages
          <fpage>67</fpage>
          -
          <lpage>82</lpage>
          . Springer,
          <year>2012</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>B.</given-names>
            <surname>Hidasi</surname>
          </string-name>
          and
          <string-name>
            <given-names>D.</given-names>
            <surname>Tikk</surname>
          </string-name>
          .
          <article-title>Context-aware item-to-item recommendation within the factorization framework</article-title>
          .
          <source>In Proc. 3rd Workshop on Context-awareness in Retrieval and Recommendation</source>
          , pp.
          <fpage>19</fpage>
          -
          <lpage>25</lpage>
          .
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>Mikko</given-names>
            <surname>Kivelä</surname>
          </string-name>
          and
          <article-title>Mason A Porter. Estimating inter-event time distributions from finite observation periods in communication networks</article-title>
          .
          <source>arXiv preprint arXiv:1412.8388</source>
          ,
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>N.</given-names>
            <surname>Koenigstein</surname>
          </string-name>
          and
          <string-name>
            <given-names>Y.</given-names>
            <surname>Koren</surname>
          </string-name>
          .
          <article-title>Towards scalable and accurate item-oriented recommendations</article-title>
          .
          <source>In Proc RecSys</source>
          , pages
          <fpage>419</fpage>
          -
          <lpage>422</lpage>
          . ACM,
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Koren</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Bell</surname>
          </string-name>
          , and
          <string-name>
            <given-names>C.</given-names>
            <surname>Volinsky</surname>
          </string-name>
          .
          <article-title>Matrix factorization techniques for recommender systems</article-title>
          .
          <source>Computer</source>
          ,
          <volume>42</volume>
          (
          <issue>8</issue>
          ):
          <fpage>30</fpage>
          -
          <lpage>37</lpage>
          ,
          <year>2009</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>N.</given-names>
            <surname>Lathia</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Hailes</surname>
          </string-name>
          , and
          <string-name>
            <given-names>L.</given-names>
            <surname>Capra</surname>
          </string-name>
          .
          <article-title>Temporal collaborative filtering with adaptive neighbourhoods</article-title>
          .
          <source>In Proc SIGIR</source>
          , pages
          <fpage>796</fpage>
          -
          <lpage>797</lpage>
          . ACM,
          <year>2009</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>G.</given-names>
            <surname>Linden</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Smith</surname>
          </string-name>
          ,
          <string-name>
            <given-names>and J.</given-names>
            <surname>York</surname>
          </string-name>
          . Amazon.
          <article-title>com recommendations: item-to-item collaborative filtering</article-title>
          .
          <source>Internet Computing, IEEE</source>
          ,
          <volume>7</volume>
          (
          <issue>1</issue>
          ):
          <fpage>76</fpage>
          -
          <lpage>80</lpage>
          ,
          <year>2003</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <surname>J. McAuley</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Pandey</surname>
          </string-name>
          , and
          <string-name>
            <given-names>J.</given-names>
            <surname>Leskovec</surname>
          </string-name>
          .
          <article-title>Inferring networks of substitutable and complementary products</article-title>
          .
          <source>In Proc SIGKDD</source>
          , pages
          <fpage>785</fpage>
          -
          <lpage>794</lpage>
          . ACM,
          <year>2015</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <surname>H. B. McMahan</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          <string-name>
            <surname>Holt</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Sculley</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Young</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Ebner</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Grady</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <string-name>
            <surname>Nie</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Phillips</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          <string-name>
            <surname>Davydov</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Golovin</surname>
          </string-name>
          , et al.
          <article-title>Ad click prediction: a view from the trenches</article-title>
          .
          <source>In Proc SIGKDD</source>
          , pages
          <fpage>1222</fpage>
          -
          <lpage>1230</lpage>
          . ACM,
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <given-names>R.</given-names>
            <surname>Pálovics</surname>
          </string-name>
          and
          <string-name>
            <given-names>A. A.</given-names>
            <surname>Benczúr</surname>
          </string-name>
          .
          <article-title>Temporal influence over the last.fm social network</article-title>
          .
          <source>In Proc. ASONAM</source>
          , pages
          <fpage>486</fpage>
          -
          <lpage>493</lpage>
          . ACM,
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [24]
          <string-name>
            <given-names>R.</given-names>
            <surname>Pálovics</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. A.</given-names>
            <surname>Benczúr</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Kocsis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Kiss</surname>
          </string-name>
          , and
          <string-name>
            <given-names>E.</given-names>
            <surname>Frigó</surname>
          </string-name>
          .
          <article-title>Exploiting temporal influence in online recommendation</article-title>
          .
          <source>In Proc. RecSys</source>
          , pages
          <fpage>273</fpage>
          -
          <lpage>280</lpage>
          . ACM,
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          [25]
          <string-name>
            <given-names>A.</given-names>
            <surname>Paterek</surname>
          </string-name>
          <article-title>Improving regularized singular value decomposition for collaborative ifltering</article-title>
          <source>In Proc. SIGKDD</source>
          , pages
          <fpage>39</fpage>
          -
          <lpage>42</lpage>
          . ACM,
          <year>2007</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          [26]
          <string-name>
            <given-names>I.</given-names>
            <surname>Pilászy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Serény</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Dózsa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Hidasi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Sári</surname>
          </string-name>
          , and
          <string-name>
            <given-names>J.</given-names>
            <surname>Gub</surname>
          </string-name>
          .
          <article-title>Neighbor methods vs matrix factorization - case studies of real-life recommendations</article-title>
          .
          <source>In LSRS2015 at RECSYS</source>
          ,
          <year>2015</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          [27]
          <string-name>
            <given-names>I.</given-names>
            <surname>Pilászy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Zibriczky</surname>
          </string-name>
          , and
          <string-name>
            <given-names>D.</given-names>
            <surname>Tikk</surname>
          </string-name>
          .
          <article-title>Fast als-based matrix factorization for explicit and implicit feedback datasets</article-title>
          .
          <source>In Proc. RecSys</source>
          , pages
          <fpage>71</fpage>
          -
          <lpage>78</lpage>
          . ACM,
          <year>2010</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          [28]
          <string-name>
            <given-names>S.</given-names>
            <surname>Rendle</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Freudenthaler</surname>
          </string-name>
          , and L.
          <string-name>
            <surname>Schmidt-Thieme</surname>
          </string-name>
          .
          <article-title>Factorizing personalized markov chains for next-basket recommendation</article-title>
          .
          <source>In Proc. WWW</source>
          , pages
          <fpage>811</fpage>
          -
          <lpage>820</lpage>
          . ACM,
          <year>2010</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          [29]
          <string-name>
            <given-names>B.</given-names>
            <surname>Sarwar</surname>
          </string-name>
          , G. Karypis,
          <string-name>
            <given-names>J.</given-names>
            <surname>Konstan</surname>
          </string-name>
          , and
          <string-name>
            <given-names>J.</given-names>
            <surname>Reidl</surname>
          </string-name>
          .
          <article-title>Item-based collaborative filtering recommendation algorithms</article-title>
          .
          <source>In Proc. WWW</source>
          , pages
          <fpage>285</fpage>
          -
          <lpage>295</lpage>
          ,
          <year>2001</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          [30]
          <string-name>
            <given-names>G.</given-names>
            <surname>Takács</surname>
          </string-name>
          ,
          <string-name>
            <given-names>I.</given-names>
            <surname>Pilászy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Németh</surname>
          </string-name>
          , and
          <string-name>
            <given-names>D.</given-names>
            <surname>Tikk</surname>
          </string-name>
          .
          <article-title>Investigation of various matrix factorization methods for large recommender systems</article-title>
          .
          <source>In Proc. 2nd KDD Workshop on Large-Scale Recommender Systems and the Netflix Prize Competition</source>
          , pages
          <fpage>1</fpage>
          -
          <lpage>8</lpage>
          . ACM,
          <year>2008</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref31">
        <mixed-citation>
          [31]
          <string-name>
            <given-names>R.</given-names>
            <surname>Turrin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Quadrana</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Condorelli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Pagano</surname>
          </string-name>
          , and
          <string-name>
            <given-names>P.</given-names>
            <surname>Cremonesi</surname>
          </string-name>
          .
          <article-title>30music listening and playlists dataset</article-title>
          .
          <source>In RecSys Posters</source>
          ,
          <year>2015</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref32">
        <mixed-citation>
          [32]
          <string-name>
            <surname>Alexei</surname>
            <given-names>Vazquez</given-names>
          </string-name>
          , Balazs Racz, Andras Lukacs,
          <string-name>
            <given-names>and Albert-Laszlo</given-names>
            <surname>Barabasi</surname>
          </string-name>
          .
          <article-title>Impact of non-poissonian activity patterns on spreading processes</article-title>
          .
          <source>Physical review letters</source>
          ,
          <volume>98</volume>
          (
          <issue>15</issue>
          ):
          <fpage>158702</fpage>
          ,
          <year>2007</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref33">
        <mixed-citation>
          [33]
          <string-name>
            <given-names>J.</given-names>
            <surname>Vinagre</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. M.</given-names>
            <surname>Jorge</surname>
          </string-name>
          , and
          <string-name>
            <given-names>J.</given-names>
            <surname>Gama</surname>
          </string-name>
          .
          <article-title>Evaluation of recommender systems in streaming environments</article-title>
          .
          <source>In Workshop on Recommender Systems Evaluation: Dimensions and Design (REDD)</source>
          ,
          <source>in conjunction with RecSys</source>
          ,
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref34">
        <mixed-citation>
          [34]
          <string-name>
            <given-names>E. M.</given-names>
            <surname>Voorhees</surname>
          </string-name>
          et al.
          <article-title>The TREC-8 question answering track report</article-title>
          .
          <source>In TREC</source>
          , volume
          <volume>99</volume>
          , pages
          <fpage>77</fpage>
          -
          <lpage>82</lpage>
          ,
          <year>1999</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref35">
        <mixed-citation>
          [35]
          <string-name>
            <given-names>X.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Steck</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Guo</surname>
          </string-name>
          , and
          <string-name>
            <given-names>Y.</given-names>
            <surname>Liu</surname>
          </string-name>
          .
          <article-title>On top-k recommendation using social networks</article-title>
          .
          <source>In Proc. RecSys</source>
          , pages
          <fpage>67</fpage>
          -
          <lpage>74</lpage>
          . ACM,
          <year>2012</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref36">
        <mixed-citation>
          [36]
          <string-name>
            <given-names>Q.</given-names>
            <surname>Yuan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Chen</surname>
          </string-name>
          , and
          <string-name>
            <given-names>S.</given-names>
            <surname>Zhao</surname>
          </string-name>
          .
          <article-title>Factorization vs. regularization: fusing heterogeneous social relationships in top-n recommendation</article-title>
          .
          <source>In Proc. RecSys</source>
          , pages
          <fpage>245</fpage>
          -
          <lpage>252</lpage>
          . ACM,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>