<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>October</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Implicit Feedback Recommendation via Implicit-to-Explicit Ordinal Logistic Regression Mapping</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Alexandros Karatzoglou Telefonica Research</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Xavier Amatriain Telefonica Research</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Denis Parra University of Pittsburgh</institution>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Idil Yavuz University of Pittsburgh</institution>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2011</year>
      </pub-date>
      <volume>23</volume>
      <issue>2011</issue>
      <abstract>
        <p>One common dichotomy faced in recommender systems is that explicit user feedback -in the form of ratings, tags, or user-provided personal information- is scarce, yet the most popular source of information in most state-of-the-art recommendation algorithms, and on the other side, implicit user feedback - such as numbers of clicks, playcounts, or web pages visited in a session- is more frequently available, but there are fewer methods well studied to provide recommendations based on this kind of information. Given the current scenario, and under a situation where just implicit user feedback is available, it would be more appropriate either to provide recommendations using the implicit data and implicit-fedback-based methods, or to map implicit user feedback to explicit feedback and then use an explicit-based algorithm? On this paper, we analyze this problem in the context of music recommendation by means of a well-known implicit feedback recommendation method described in Hu et al. [1] by comparing the use of raw playcounts with the use of explicit data - user ratings - obtained by mapping implicit to explicit feedback with a novel mixede ects logistic regression model.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. INTRODUCTION</title>
      <p>
        Recommender Systems (RS) [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] have proved their
business value and impact on many application scenarios that
go from recommending movie rentals to new contacts on a
social network. One of the main features of these systems
is that they rely on understanding user preferences in
order to estimate the utility of items and decide whether they
should be recommended. These user preferences are infered
by taking into account direct feedback from the user, either
in explicit or implicit form.
      </p>
      <p>
        We obtain implicit feedback [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] by measuring the
interaction of the user with the di erent items. We can use signals
such as the number of playcounts in a song, or the clicks on
webpages as implicit feedback. This kind of data is obtained
without incurring into any overhead on the user, since it is
obtained from direct usage [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. However, it is not clear that
we can trust a simple one-to-one mapping between usage
and preference [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. On the other hand, explicit feedback is
obtained by directly querying the user, who is usually
presented with an integer scale where to quantify how much
she likes the items. In principle, explicit feedback is a more
robust way to extract preference, since the user is reporting
directly on this variable, removing the need of an indirect
inference. However, it is also known that this kind of
feedback is a ected by user inconsistencies known as natural
noise [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. Besides, the fact that we are introducing a user
overhead, makes it di cult to have a complete view on the
user preferences [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ].
      </p>
      <p>
        None of the two existing strategies for capturing user
feedback clearly outperforms the other. Ideally, we would like
to use implicit feedback, minimizing the impact on the user,
but having a robust and proven way to map this
information to the actual user preference. In a previous work [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ], we
tested several regression models and we were able to map
implicit user feedback to explicit ratings. Our results were
satisfactory, but we did not compare to state-of-the-art
methods that make use of raw implicit information to provide
recommendations. In this paper we propose an ordinal logistic
regression model that by using a few ratings is able to infer
a generic parametric mapping from implicit to explicit data.
Our mapping model integrates usual implicit user feedback
(playcounts) with contextual information (how recently the
user listened to an album). We compare our approach to a
state-of-the art algorithm for implicit feedback
recommendations and discuss possible extensions.
2.
      </p>
    </sec>
    <sec id="sec-2">
      <title>PRELIMINARIES AND RELATED WORK</title>
      <p>Implicit feedback is much more readily available in
practical scenarios for recommender systems. However, most of
the research literature focuses on the use of explicit feedback
input since this is considered the ground truth on the user
preferences and allows to reduce the recommender problem
to one of predicting ratings.</p>
      <p>
        In one of the few papers addressing the implicit feedback
recommendation problem [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], Hu et al. deal with the implicit
feedback recommendation problem by binarizing it and
introducing the idea of con dence. In our previous work [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ],
however, we presented an analysis of implicit and explicit
feedback that challenged most of the assumptions stated
in [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. In particular: (1) There is no negative
feedback. While it is true that you cannot interpret \no implicit
feedback\ as \negative feedback\ { and this is true also for
explicit feedback{, implicit data can include negative
feedback. You can assume that low feedback is negative
feedback as long as the granularity of the items is comparable,
and there is enough variability. (2) Implicit feedback is
noisy. Implicit feedback is noisy but, as we showed in
previous work [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ], so is explicit feedback. (3) Preference vs.
Con dence. As we showed in our work [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ], the numerical
value of implicit feedback can indeed be directly mapped
to preference, given the appropriate mapping. (4)
Evaluation of implicit feedback. On the other hand, we do
agree that there is no appropriate evaluation approaches for
implicit feedback and this is in fact one of the motivations
of our work: if we nd an appropriate way to map implicit
to explicit feedback we can ensure an evaluation that is as
good as the one we have in the explicit case.
      </p>
      <p>
        Our hypothesis that there is some observable correlation
between implicit and explicit feedback can be tracked in the
literature. Already in 1994, Morita and Shinoda [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] proved
that there was a correlation between reading time on
online news and self-reported preference. Konstan et al. [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ]
did a similar experiment with the larger user base of the
Grouplens project and again found this to be true. Oard
and Kim [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] performed experiments using not only reading
time but also other actions like printing an article to nd a
positive correlation between implicit feedback and ratings.
Koh et al. did a thorough study of rating behavior in two
popular websites [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ]. They hypothesize that the overall
popularity or average rating of an item will in uence raters
and they conclude that while there is an e ect, this depends
on the cultural background of the raters.
      </p>
      <p>
        Lee et al. [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ] implement a recommender system based
on implicit feedback by constructing \pseudo-ratings"
using temporal information. In this work, the authors
introduce the idea that recent implicit feedback should contribute
more positively towards inferring the rating. The authors
also use the idea of distinguishing three temporal bins: old,
middle, and recent.
      </p>
      <p>
        Two recent works approach the issue of implicit feedback
in the music domain. Jawasher et. al analyze the
characteristics of user implicit and explicit feedback in the context
of last.fm music service [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ]. However, their results are not
conclusive due to limitations in the dataset since they only
used explicit feedback available in the last.fm pro les, which
is limitted to the love/ban binary categories. This data is
very sparse and, as the authors report, almost non-existant
for some users or artists. On the other hand, Kurdomova
et. al use a Bayesian approach to learn a classi er on
multiple implicit feedback variables [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ]. Using these features,
the authors are able to classify liked and disliked items with
an accuracy of 0.75, uncovering the potential of mapping
implicit feedback directly to preferences.
      </p>
      <p>
        In our previous work [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ], we showed that it was possible
to create a simple parametric model for implicit feedback
by using linear regression on some available explicit ratings.
However, as we will explain, in the context of user ratings,
it may be more appropriate to use a mixed-e ects ordinal
logistic regression model. In this context, the main
contribution of our present work is to present an ordinal logistic
regression model that allows to map implicit data into explicit
ratings for the task of recommendation. We make our model
context-aware with respect to how recently a user listened
to an album by contextual modeling, i.e., using the
contextual information directly in the modelling technique, unlike
data-driven approaches such as contextual pre- ltering or
post- ltering [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ]. Once the implicit-to-explicit mapping is
performed, we can use the inferred ratings in methods for
explicit or implicit data. We can then compare the
performance of these models to the one by Hu et al. in several
experiments.
3.
3.1
      </p>
    </sec>
    <sec id="sec-3">
      <title>REGRESSION MODELS</title>
    </sec>
    <sec id="sec-4">
      <title>Linear Regression</title>
      <p>
        In [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] we introduce a linear regression model to predict
explicit preference of users on music albums in the form of
ratings based on implicit user behavior variables - (1)
Implicit Feedback (if ): playcount for a user on a given item;
(2) Global Popularity (gp): global playcount for all users
on a given item; (3) Recentness (re) : time elapsed since
user played a given item. In that article, we compare
different linear regression models based on the aforementioned
variables and we nd that the variables implicit feedback
and recentness explain the largest part the variability of the
ratings, while global popularity explained a very small
portion. This result suggested us that the two former variables
would be better predictors of the user preference, and we
supported these assumption by performing a 10-fold cross
validation experiment using the data of our online survey on
music preference as a ground truth. The RMSE values were
consistent with the previously described regression analysis.
3.1.1
      </p>
      <p>Limitations and shortcomings of Linear
Regression</p>
      <p>
        Although the linear regression gives good results, there are
some considerations that must be observed to generalize this
model to other domains and to make it able to be compared
with other approaches. First, depending on the application
we may want the predicted values to fall in the range from
1 to 5, but using linear regression we cannot ensure it.
Second, as in most of recommender systems research, our main
evaluation metric is RMSE. When using this metric, we are
assuming that ratings form an interval scale, i.e. the
distance between any two consecutive values in the rating scale
is the same. However, in a previous study [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ], we have shown
that users have a larger probabilty to be more inconsistent
with some ratings numbers than with others, what give us
the clue that users do not see the rating scale as equally
spaced. Hence, we should consider the ratings as an
ordinal variable rather than an linear or interval one. This also
implies that RMSE is not a good measure alone to predict
user preference, it should be combined, and in some cases
replaced, with other measures coming from Information
Retrieval such as precision, recall, or nDCG.
      </p>
      <p>Given that users present individual variability in their
ratings, a good extension of our model should include the user
as a random factor. Additionally, given that ratings are
actually an ordinal variable, as explained in the previous
paragraph, and the fact that are not normally distributed,
logistic regression is a proper alternative to our linear
regression model. Combining both considerations, our next
model for implicit-to-explicit behavior mapping model will
be a mixed-e ects logistic regression.
3.2</p>
    </sec>
    <sec id="sec-5">
      <title>Mixed-effects Ordinal Logistic Regression</title>
      <p>The multinomial logistic regression is the natural model
for an ordinal scale variable (rating, that ranges from 1 to
5) and a mixed-e ects model will help us to reduce the
variability due to di erences in rating among the users. Our
multinomial logistic regression, that uses cummulative logit
as link function, can be represented as:
logit(P (rui
k)) = k + X
+ gu
(1)
where k = f1; 2; 3; 4g, rui is the rating that user u gives to
item i, P (rui k) is the probability that the rating rui is
less or equal than k, k is the intercept for the cumulative
probability that rating is less than or equal to k, X is a vector
with the actual values of the xed factors (if, re and gp), is
the vector of coe cients of the xed factors, gu iid N (0; g2)
is the random e ect of the users, and
logit(p) = log(
1
p
p</p>
      <p>)</p>
      <p>To obtain the predicted rating of a user u on an item i,
we calculate the expected value of the rating as
k=1
k)
1</p>
      <p>P (rui
P (rui k
P (rui k
k) , k = 1
1) , 1 &lt; k &lt; 5
1) , k = 5
where</p>
      <p>E ect
intercept 1
intercept 2
intercept 3
intercept 4
gp
if
re
gp*if
if*re
concerts</p>
    </sec>
    <sec id="sec-6">
      <title>EXPERIMENTAL SETUP</title>
    </sec>
    <sec id="sec-7">
      <title>Data sets</title>
      <p>We use two datasets in this study. The rst one was
collected by an online user study among users of the last.fm
music service between September and October of 2010,
containing implicit and explicit information, and also demographic
and consumption data. The second one was collected using
the last.fm API during May of 2011, and contains only
implicit information. The characteristics of both datasets are
described in Table 2 .
4.1.1</p>
      <sec id="sec-7-1">
        <title>Generating Explicit Fedback</title>
        <p>
          We conducted an online user study among users of the
last.fm music service. The goal of the study was to gather
explicit feedback on music albums to compare to the user
implicit feedback we obtained by directly crawling the last.fm
page related to the user taking the survey. Explicit
feedback was obtained by asking users to rate albums on a 1
to 5 star scale. The items to rate were obtained from the
(2)
(3)
(4)
list of albums in the user's playlist so that users responded
to a personalized survey. Details of this study, such as the
strategy to sample the items that were rated by users and
the results of user demographics and user consumption, can
be found in our previous article [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ].
4.1.2 Implicit Music Consumption Feedback
        </p>
        <p>We call Implicit Music Consumption Feedback to our Dataset2
since, unlike Dataset1 that has demographic data of each
user, it only has information about implicit behavior of the
users: playcount of albums per each user, how recently each
album was listened to for the last time, and the total
number of listeners of each album in the whole last.fm website.
The statistics of this dataset are described in Table 2.
4.2</p>
      </sec>
    </sec>
    <sec id="sec-8">
      <title>Regression Model Selection</title>
      <p>To select the xed e ects that would be part of our model
we conducted a forward selection on the set of all the main
e ects and their two-way interactions. The main e ects
considered were if , re, gp (as described in section 3.1) plus ten
demographic and consumption variables: gender, age, hours
of music per week, hours of internet per week, buying
physical records, buying online records, interaction style
(preference on listening to tracks or albums), number of concerts
per year, interest on reading specialized music blogs or
magazines, and familiarity rating music online. We have to pick
two models nally because of the nature of our two datasets.
In the smallest one (dataset1) we have all the variables
obtained by a user study, but in the second dataset (dataset2)
we just have implicit information (playcounts per user, how
recently the user listened to each album, and the total
number of listeners of an album in the whole dataset) that can
be reduced to if , re and gp.</p>
      <p>
        After conducting the process of forward selection, the model
obtained for dataset1 considers four xed e ects (if, re, gp
and concerts per year) and the random e ect of the user.
The details of the model are described in Table 1. Although
the main e ects of global popularity (gp) and recentness (re)
are not signi cant, we keep them in the model because their
interaction with implicit feedback (if ) is signi cant [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ].
      </p>
      <p>For dataset2, we consider in the model if , re, and gp as
xed e ects plus the random e ect of the user. For the sake
of space we do not show the details of this model, but the
coe cient and signi cant values are similar to those shown
in Table 1 excepting that the factor number of concerts is
not considered in the model. As in the previous model, we
keep in the model gp and re although they are not signi cant
due to their interaction with if . Under this model, is also
users
albums
entries
density
avg albums/user
avg user/album
not signi cant the intercept for rating equal to 2, which tell
us that this intercept is not signi cantly di erent than 0,
and we may dismiss it from the model.</p>
    </sec>
    <sec id="sec-9">
      <title>4.3 Comparing the different approaches</title>
      <p>After we have done the implicit-to-explicit mapping, we
are in condition to compare the use of impplicit data with
inferred explicit data. In this article, we compare four
approaches using dataset 1 and three aproaches using dataset
2. The methods we compare, as identi ed in the rst column
of Table 3, are:</p>
      <p>
        HK : the implicit feedback method introduced in Hu et al.
[
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] which uses raw playcounts,
HKlog: a variation of the HK method, also introduced in
[
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], that makes a log-transformation of the playcounts,
logit3 : the HK method, where the input values are the
ratings inferred by logistic regression using 3 xed factors
(if, gp, and re)
logit4 : similar to logit3 but adding the factor number of
concerts in the logistic regression model to infer the ratings.
      </p>
      <p>We have this information available just for dataset1.</p>
      <p>
        Description of the HK method. For the implicit
feedback modeling we use the Matrix Factorization method
developed in [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. In this Matrix Factorization method a
weighted least squares error loss function is minimized. To
this end user-item interactions pij are signaled with a 1 and
missing interactions are marked with a 0. The counts of
user-item interactions (e.g. playcounts Yij ) are translated
into a con dence measure wij, which in the case of the HK
method correspond to pij + Yij, and in the case of the
HKlog method a simple log transform is used where:
wij =
1
log(1 + Yijk) Yijk &gt; 0
      </p>
      <p>Yijk = 0</p>
      <p>This "con dence" is then used as a weight in the loss
function and the objective function then becomes
min
U;M;C i j
n m
X X[wij (pij</p>
      <p>hUi Mj i)2
+ n jjUi jj2 + m jjMj jj2]
(5)
(6)</p>
      <p>where the Frobenius norm of the factor matrices is used
for regularization. This minimization problem is then solved
in linear time using Alternate Least Squares and utilizing a
trick to avoid direct optimization over the 0 entries of the
matrix.
4.3.1</p>
      <sec id="sec-9-1">
        <title>Error Measures</title>
        <p>
          RMSE [
          <xref ref-type="bibr" rid="ref18">18</xref>
          ] is probably the most common measure to
evaluate the performance of recommender systems and we used
it to evaluate and compare our linear regression approaches
in [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ]. However, when there are no ratings to assess the
performance of the algorithms we can not use metrics like
RMSE or MAE. Hence, we opt for using Mean Average
Precision (MAP) [
          <xref ref-type="bibr" rid="ref19">19</xref>
          ] and normalized Discounted Cummulative
Gain (nDCG) [
          <xref ref-type="bibr" rid="ref20">20</xref>
          ]. The former gives us an overall sense of
how well we identify relevant items to recommend from a
set of retrieved recommendations, and the latter how well
we rank them in a list.
5.
        </p>
      </sec>
    </sec>
    <sec id="sec-10">
      <title>RESULTS</title>
      <p>In order to evaluate and compare the methods, we split
each dataset into 5 groups in order to perform a 5-fold cross
validation. The result of each run is a list of recommended
items (albums) for each user in the test set, sorted by the
preference that the user would have for that item. We
calculate MAP and nDCG for each list recommended to a
user, judging an item as relevant whether it was consumed
(played) at least once by the user. Results can be seen in
Table 3.</p>
      <p>In the case of dataset 1, the best results of MAP and
nDCG are obtained by recommending the most popular
items. This result is somewhat expected due to the
sparsity of the dataset that a ects the methods based on matrix
factorization. As shown in Table 2, each album was rated in
average by just 1:71 users. This situation is not repeated in
dataset 2, where the average number of users per album is
18:52, and then the popularity method performs the worst.</p>
      <p>We highlight two results on these initial experiments. The
rst one is that the log transformation of raw playcounts
makes HKlog improve clearly over HK on both MAP and
nDCG measures. The second result we higlight is that logit3
and logit4 perform better than HK and there is not a big
di erence in performance with HKlog, leading us investigate
further to con rm this di erence.</p>
    </sec>
    <sec id="sec-11">
      <title>CONCLUSIONS AND FUTURE WORK</title>
      <p>
        In this paper, we continue the work that we started in [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]
to create a model that allows us to map implicit to explicit
user behavior. Using MAP and nDCG metrics, we show
that our method is comparable to state of the art methods
that provides recommendations making use of implicit user
feedback.
      </p>
      <p>The results that we have obtained, part of which we show
on this paper, give us some insights but they mainly open
research questions that we need to analyze further. We have
con rmed in our dataset the bene ts of applying a log
transformation to the raw user feedback in the Hu et al. model,
showing consistently better results than the unmodi ed
version.</p>
      <p>
        In terms of the questions we need to further analyze, up
to this point, we have considered the factors implicit
feedback and global popularity in our logistic regression models
as ordinal variables. We coded these variables on this way
to make sure that we were doing an appropiate diverse
sampling when creating the user survey described in [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ].
However, there is no constraint to rather use the raw playcounts
for both factors aforementioned, and we think that this
modi cation can bene t the results of our implicit-to-explicit
logistic regression model.
      </p>
      <p>On the experiments run on this study, since we are not
predicting user ratings but rather user preference, metrics
such as RMSE or MAE can not be used to compare the
methods so we opt for IR metrics such as MAP and nDCG,
which rely on how we de ne relevancy. We wonder if our
definition of relevance might bias our results and conclusions.
As we have stated it before, we think that low feedback
might be, in fact, negative feedback. For this reason, we are
currently testing di erent user activity (implicit feedback)
thresholds to de ne relevancy in order to analyze how that
in uences the evaluation of the di erent recommendation
approaches.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Hu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Koren</surname>
          </string-name>
          , and
          <string-name>
            <given-names>C.</given-names>
            <surname>Volinsky</surname>
          </string-name>
          .
          <article-title>Collaborative ltering for implicit feedback datasets</article-title>
          .
          <source>In Proceedings of ICDM</source>
          <year>2008</year>
          ,
          <year>2008</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>F.</given-names>
            <surname>Ricci</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Rokach</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Shapira</surname>
          </string-name>
          , and P. B. Kantor, editors.
          <source>Recommender Systems Handbook</source>
          . Springer,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>Douglas</given-names>
            <surname>Oard</surname>
          </string-name>
          and
          <string-name>
            <given-names>Jinmook</given-names>
            <surname>Kim</surname>
          </string-name>
          .
          <article-title>Implicit feedback for recommender systems</article-title>
          .
          <source>In in Proceedings of the AAAI Workshop on Recommender Systems</source>
          , pages
          <fpage>81</fpage>
          {
          <fpage>83</fpage>
          ,
          <year>1998</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>G.</given-names>
            <surname>Potter.</surname>
          </string-name>
          <article-title>Putting the collaborator back into collaborative ltering</article-title>
          .
          <source>In 2nd KDD Workshop on Large-Scale Recommender Systems and the Net ix Prize Competition</source>
          ,
          <year>2008</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>D. M.</given-names>
            <surname>Nichols</surname>
          </string-name>
          .
          <article-title>Implicit rating and ltering</article-title>
          .
          <source>In In Proceedings of the Fifth DELOS Workshop on Filtering and Collaborative Filtering</source>
          , pages
          <volume>31</volume>
          {
          <fpage>36</fpage>
          ,
          <year>1997</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>X.</given-names>
            <surname>Amatriain</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.M.</given-names>
            <surname>Pujol</surname>
          </string-name>
          , and
          <string-name>
            <given-names>N.</given-names>
            <surname>Oliver</surname>
          </string-name>
          .
          <article-title>I like it</article-title>
          ...
          <article-title>i like it not: Evaluating user ratings noise in recommender systems</article-title>
          .
          <source>In Proc. of the 2009 Conference on User Modeling</source>
          , Adaptation, and
          <string-name>
            <surname>Personalization</surname>
          </string-name>
          ,
          <year>2009</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>G.</given-names>
            <surname>Jawaheer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Szomszor</surname>
          </string-name>
          , and
          <string-name>
            <given-names>P.</given-names>
            <surname>Kostkova</surname>
          </string-name>
          .
          <article-title>Characterisation of explicit feedback in an online music recommendation service</article-title>
          .
          <source>In Proceedings of the fourth ACM conference on Recommender systems, RecSys '10</source>
          , pages
          <fpage>317</fpage>
          {
          <fpage>320</fpage>
          ,
          <year>2010</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>D.</given-names>
            <surname>Parra</surname>
          </string-name>
          and
          <string-name>
            <given-names>X.</given-names>
            <surname>Amatriain</surname>
          </string-name>
          .
          <article-title>Walk the talk: Analyzing the relation between implicit and explicit feedback for preference elicitation</article-title>
          .
          <source>In Proc. of the 2011 Conference on User Modeling</source>
          , Adaptation, and
          <string-name>
            <surname>Personalization</surname>
          </string-name>
          ,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>M.</given-names>
            <surname>Morita</surname>
          </string-name>
          and
          <string-name>
            <given-names>Y.</given-names>
            <surname>Shinoda</surname>
          </string-name>
          .
          <article-title>Information ltering based on user behavior analysis and best match text retrieval</article-title>
          .
          <source>In SIGIR '94: Proceedings of the 17th annual international ACM SIGIR conference</source>
          , pages
          <volume>272</volume>
          {
          <fpage>281</fpage>
          , New York, NY, USA,
          <year>1994</year>
          . Springer-Verlag New York, Inc.
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <surname>Joseph</surname>
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Konstan</surname>
            ,
            <given-names>Bradley N.</given-names>
          </string-name>
          <string-name>
            <surname>Miller</surname>
          </string-name>
          , David Maltz, Jonathan L.
          <string-name>
            <surname>Herlocker</surname>
          </string-name>
          ,
          <string-name>
            <surname>Lee R. Gordon</surname>
            ,
            <given-names>and John</given-names>
          </string-name>
          <string-name>
            <surname>Riedl</surname>
          </string-name>
          . Grouplens:
          <article-title>applying collaborative ltering to usenet news</article-title>
          .
          <source>Commun. ACM</source>
          ,
          <volume>40</volume>
          (
          <issue>3</issue>
          ):
          <volume>77</volume>
          {
          <fpage>87</fpage>
          ,
          <year>1997</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>D.</given-names>
            <surname>Oard</surname>
          </string-name>
          and
          <string-name>
            <given-names>J.</given-names>
            <surname>Kim</surname>
          </string-name>
          .
          <article-title>Modeling information content using observable behavior</article-title>
          .
          <source>In Proc. of the ASIST Annual Meeting</source>
          , pages
          <fpage>p481</fpage>
          {
          <fpage>88</fpage>
          ,
          <year>2001</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>N.S.</given-names>
            <surname>Koh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Hu</surname>
          </string-name>
          , and
          <string-name>
            <given-names>E. K.</given-names>
            <surname>Clemons</surname>
          </string-name>
          .
          <article-title>Do online reviews re ect a product's true perceived quality? - an investigation of online movie reviews across cultures</article-title>
          .
          <source>Electronic Commerce Research and Applications</source>
          ,
          <year>2010</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>T.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Park</surname>
          </string-name>
          , and
          <string-name>
            <given-names>Y.</given-names>
            <surname>Park</surname>
          </string-name>
          .
          <article-title>A time-based approach to e ective recommender systems using implicit feedback</article-title>
          .
          <source>Expert Syst. Appl.</source>
          ,
          <volume>34</volume>
          (
          <issue>4</issue>
          ):
          <volume>3055</volume>
          {
          <fpage>3062</fpage>
          ,
          <year>2008</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <surname>Gawesh</surname>
            <given-names>Jawaheer</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>Martin</given-names>
            <surname>Szomszor</surname>
          </string-name>
          , and
          <string-name>
            <given-names>Patty</given-names>
            <surname>Kostkova</surname>
          </string-name>
          .
          <article-title>Comparison of implicit and explicit feedback from an online music recommendation service</article-title>
          .
          <source>In Proceedings of the 1st International Workshop on Information Heterogeneity and Fusion in Recommender Systems</source>
          ,
          <year>2010</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>S.</given-names>
            <surname>Kordumova</surname>
          </string-name>
          , I. Kostadinovska,
          <string-name>
            <given-names>M.</given-names>
            <surname>Barbieri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Pronk</surname>
          </string-name>
          , and
          <string-name>
            <given-names>J.</given-names>
            <surname>Korst</surname>
          </string-name>
          .
          <article-title>Personalized implicit learning in a music recommender system</article-title>
          .
          <source>In UMAP</source>
          <year>2010</year>
          ,
          <year>2010</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>Gediminas</given-names>
            <surname>Adomavicius</surname>
          </string-name>
          and
          <string-name>
            <given-names>Alexander</given-names>
            <surname>Tuzhilin</surname>
          </string-name>
          .
          <article-title>Context-aware recommender systems</article-title>
          . In Francesco Ricci, Lior Rokach, Bracha Shapira, and Paul B. Kantor, editors,
          <source>Recommender Systems Handbook</source>
          , pages
          <volume>217</volume>
          {
          <fpage>253</fpage>
          .
          <string-name>
            <surname>Springer</surname>
            <given-names>US</given-names>
          </string-name>
          ,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>J.</given-names>
            <surname>Neter</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. H.</given-names>
            <surname>Kutner</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C. J.</given-names>
            <surname>Nachtsheim</surname>
          </string-name>
          , and
          <string-name>
            <given-names>W.</given-names>
            <surname>Wasserman</surname>
          </string-name>
          . Applied Linear Statistical Models. Irwin, Chicago,
          <year>1996</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <surname>Jonathan</surname>
            <given-names>L.</given-names>
          </string-name>
          <string-name>
            <surname>Herlocker</surname>
          </string-name>
          , Joseph A.
          <string-name>
            <surname>Konstan</surname>
          </string-name>
          , Loren G. Terveen, and John T. Riedl.
          <article-title>Evaluating collaborative ltering recommender systems</article-title>
          .
          <source>ACM Trans. Inf</source>
          . Syst.,
          <volume>22</volume>
          (
          <issue>1</issue>
          ):5{
          <fpage>53</fpage>
          ,
          <year>2004</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <surname>Christopher</surname>
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Manning</surname>
            , Prabhakar Raghavan, and
            <given-names>Hinrich</given-names>
          </string-name>
          <string-name>
            <surname>Schtze</surname>
          </string-name>
          . Introduction to Information Retrieval. Cambridge University Press, New York, NY, USA,
          <year>2008</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <surname>Kalervo</surname>
            <given-names>Ja</given-names>
          </string-name>
          <article-title>rvelin and Jaana Kekalainen. Cumulated gain-based evaluation of ir techniques</article-title>
          .
          <source>ACM Trans. Inf</source>
          . Syst.,
          <volume>20</volume>
          :
          <fpage>422</fpage>
          {
          <fpage>446</fpage>
          ,
          <year>October 2002</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>