<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Nameling Discovery Challenge - Collaborative Neighborhoods</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Dirk Schafer</string-name>
          <email>dirkschaefer@jivas.de</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Robin Senge</string-name>
          <email>senge@informatik.uni-marburg.de</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Mathematics and Computer Science Department Philipps-Universitat Marburg</institution>
          ,
          <country country="DE">Germany</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>This paper describes a series of experiments designed to solve the \Nameling" challenge. In this task, a recommender should provide suggestions for interesting rst names, based on a set of names in which a user has shown interest. An approach based on dyadic factors is proposed where side-information about names and users were incorporated. Furthermore, factors based on User-based Collaborative Filtering play a central role. The performance considering the neighborhood and binary similarity measures was assessed.</p>
      </abstract>
      <kwd-group>
        <kwd>collaborative ltering</kwd>
        <kwd>implicit feedback</kwd>
        <kwd>dyad</kwd>
        <kwd>competition</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Introduction</title>
      <p>
        Implicit feedback data can be collected whenever users are interacting with
information systems. In connection with recommender systems this data is appealing,
because it is obtainable in large quantities, e.g. from log les, and can be used
as a complementary information source to rating data. On the one hand, the
popularity of this kind of data is re ected by the numerous synonyms1 in
literature [
        <xref ref-type="bibr" rid="ref11 ref12 ref5 ref6">6, 5, 12, 11</xref>
        ]. One the other hand, there are various applications ranging
from basket case analysis [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] to large scale news recommendation [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] and applied
machine learning elds, e.g. Information Retrieval and Recommender Systems
research to name a few.
The Nameling discovery challenge is part of a workshop held at the European
Conference on Machine Learning and Principles and Practice of Knowledge
Discovery in Databases2 2013.
1 0-1 data, one-class data, positive-only feedback, click-stream data, market basket
data, click-through data
2 ECML PKDD
      </p>
      <sec id="sec-1-1">
        <title>Task Description</title>
        <p>
          Given is a set of names that has been recorded as input by users of the Nameling
web platform [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ]. The task consists in recommending further names for a subset
of selected users. The recommender should build an ordered list of one thousand
names for each test user, where the most relevant names are placed at high rank
positions. As performance metric the Mean Average Precision (MAP@1000) had
to be used.
U
I
u,v
i,j
I(u)
U (i)
wuv
suj
N
Symbols De nition
        </p>
        <p>Set of all users
Set of all items
Indices for users
Indices for items (=names)
Item set of user u
Set of users that have an a liation with item i
Similarity between two item sets
Propensity for user u to select item j</p>
        <p>Items that are listed in the le \Namelist.txt"
2.2</p>
      </sec>
      <sec id="sec-1-2">
        <title>Dataset</title>
        <p>The dataset contains a training set which is a sparse matrix consisting of 59764
users (rows) and 17479 names (columns) and a test set with 4140 user IDs.
Furthermore, a namelist le3 is provided consisting of 44k names. For each user
of the test set, two names from the system activity \ENTER SEARCH", that
are also contained in the namelist, have been extracted and are held out by
the challenge organizers. Two Perl scripts were provided, the rst one is able to
build an own training and test set with names from the challenge training set
exactly the same way as the challenge training and test set were built. The second
script can be used for evaluation. In Figure 1 the sparsity of the training set is
re ected by the frequencies of the most often chosen names. It shows a long-tail
distribution, i.e. a small amount of names is very popular and at position 2000
names occur, which were only chosen by few users.
3</p>
      </sec>
    </sec>
    <sec id="sec-2">
      <title>Related Work</title>
      <p>In recommender systems literature, algorithms are classi ed as either being
neighborhood-based (aka memory-based) or model-based. In the rst group,
there are user-based and item-based collaborative ltering methods. For
userbased k-nearest neighbor collaborative ltering (UCF) the idea is to identify a
3 Namelist.txt, see abbreviation N in Table 1
2500
2000
1cy500
n
e
u
1req000
F
500
group of k most similar users for each user to infer a ranking of items that are
new for the user and also most interesting. Item-based nearest neighbor CF in
contrast identify new items that have a relation to the existing item set of a
user. The item-based CF approach has the advantage that a recommendation is
easily explainable to the user and much more e cient to generate compared to
UCF. On the other hand it lacks in accuracy.</p>
      <p>
        For the model-based approaches and especially for this setting of implicit
feedback, various methods based on Matrix Factorization (MF) have been
created. In the One-Class Collaborative Filtering framework from Pan et al. two
extreme assumptions about the data are being made [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ], these are, that all
missing values are all negative (AMAN) and all missing values are unknown
(AMAU). They use weighted matrix factorization and sampling strategies to
cope with the uncertainty about the unobserved data.
      </p>
      <p>
        Another MF method that deals with side-information has been proposed
in [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ], where an embedding of auxiliary information into a weighted MF has
been proposed. In [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ] the Bayesian Personalized Ranking framework for implicit
feedback has been described where also MF in combination with a bootstrap
sampling approach is used to learn from pairwise comparisons.
      </p>
      <p>
        NameRank is an item-based recommendation approach that has been
proposed recently in combination with the Nameling data set and showed very good
performance compared to the above mentioned approaches [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ].
4
      </p>
    </sec>
    <sec id="sec-3">
      <title>Recommendations Using Various Dyadic Factors</title>
      <p>To begin with, we describe how dyadic factors can be used to induce rankings on
names. The rest of the section deals with the engineering of these factors with two
di erent approaches. The rst one aims at improving the most popular items
approach by using side-information for users and names, whereas the second
approach is based on Collaborative Filtering.
4.1</p>
      <sec id="sec-3-1">
        <title>Basic Scoring Scheme</title>
        <p>
          In the following, each user-item pair (a dyad) receives a score based on the
following equation:
sui = dui f u[1;i] f [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ]
i
(1)
Sorting sui scores in descending order for user u provides a ranking of items that
can be recommended. The formula consists of a dyadic factor dui and two lter
factors. The lters are indicator functions that adress the requirements of the
prediction task. f [
          <xref ref-type="bibr" rid="ref1">1</xref>
          ] is 0 for names that are members of the user item-set I(u).
And the purpose of the other lter f [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ] is to comply with the demand to accept
only names that are part of the namelist N . The various possibilites to engineer
the dyadic factor are described below.
4.2
        </p>
      </sec>
      <sec id="sec-3-2">
        <title>Discovering the \Most Popular" Baseline</title>
        <p>
          In rst experiments we found out that by simply considering a constant top
1000 majority vote for all names4, we were already able to outperform some of
the results others contributed to the leaderboard at an early stage. We could
even improve this result by considering the 2000 top items and ltering out
the training names of the individual users, which lead to the de nition of the
lter factor f [
          <xref ref-type="bibr" rid="ref1">1</xref>
          ]. It turned out, that recommending items that way is known
in literature as the Most Popular (MP) approach [
          <xref ref-type="bibr" rid="ref10">10</xref>
          ]. In the course of the
challenge, we could identify other teams that scored equally, and we think they
recommended names using the same approach. Because of this and its simplicity
we say we discovered the \baseline method" within the leaderboards.
4.3
        </p>
      </sec>
      <sec id="sec-3-3">
        <title>Side-Information on Names</title>
        <p>In the given data set, there were no additional attributes on the item objects
provided. Therefore we constructed two features as side-information for names:
length of name and gender. Both features are characterized by having a nite set
of discrete values as co-domain. With that, the now explained general procedure
is applied, where the Most Popular Ranking is used as input. The idea is to
split the MP Ranking into several queues corresponding to the available discrete
values of a feature and later to adjust the MP recommendations for some of the
users on basis of their item sets. From the queues, items are sampled according
to proportions found in the item sets of the user. Since the item sets are typically
small (see Figure 2), some signi cance criteria have to be met, e.g. a minumum
size of the item-sets. Having found a ranking of at least 1000 names that way,
we turned the ordered list of names in to numerical factors by assigning score
values uniformly according to the rank position in an interval [max, min], e.g.
the top ranked names recieved scores of 1; 0:9; 0:8; : : :5.</p>
        <p>Name Length Factor The general strategy of sampling from queues described
above had to be slightly adapted due to the fact that there are 13 discrete values
for name lengths, but item sets of users are too small to capture the proportions
su ciently (see Figure 2). To be able to sample from all those queues, the user
4 that were used by the approx. 60k training users
5 we actually used the interval [1,0.1]
1y0000
c
n
e
u
q
e
rF5000
0 0</p>
        <p>40</p>
        <p>Item Set Size
10
20
30
50
60
70
80
preferences for items, in this case for short or long names, must be signi cantly
explainable. For this reason we decided to join the discrete values in two almost
equally balanced sets: the set of shorter names (length 1 to 5) and longer names
(length 6 to 13), see Table 2.
Gender Factor Since it was o cially allowed and explicitly encouraged to use
additional data sources for the o ine challenge, we decided to include gender
information for the names. By doing this, the Most Popular performance could be
considerably improved (see Table 5). We extracted all names from the training
set which were associated with the activity \ENTER SEARCH" and used an
external program6 to classify names into the following three classes: 0 - unisex
name or not identi able, 1 - male, 2 - female. We used the idea described above
and sampled from three queues according to the item sets' gender distributions
until the new recommendation set reached a size of 1000 names.
4.4</p>
      </sec>
      <sec id="sec-3-4">
        <title>Side-Information on Users</title>
        <p>In previous uses of side-information, one global Most Popular recommendation
was taken as basis and according to item set statistics the recommendation
was modi ed by reordering or ltering. In further experiments, we followed a
di erent approach: for groups of users, speci ed by their attributes, many Most
Popular recommendation lists are created. Approximate geo locations of users
were provided that were derived from IP addresses. We used in particular the
country information for de ning a new factor as described next.
6 gender.c by Jorg Michael, which has been cited in the German computer magazine
c't 2007: Anredebestimmung anhand des Vornamens.
Demographic Factor Additional geographical information was available for
a subset of users. For every country we created a separate Most Popular
recommendation list (see Table 3 for an excerpt). Unfortunately, the performance
could only be increased marginally by this e ort (see Table 5).
User-based collaborative ltering is known as a successful technique to
recommend items based on the ratings of items by di erent users. For this technique,
the choice of a similarity measure that captures how similar users are, and the
choice of neighborhood are the most important factors. In order to cope with
this kind of binary data, the similarity has to be chosen suitably for implicit
feedback data. And, as will be shown later, also the neighborhood size is even
more crucial for the performance. The general user-based CF formula provides
a score for each user-item pair as follows:
(2)
with indicator function
and a normalization term
puj =
jUj
X wuvcvj
v=1
cui =
(1; if i 2 I(u)</p>
        <p>0; else
=</p>
        <p>1</p>
        <p>PjvU=j1 wuv
to ensure that puj 2 [0; 1]. Note that Equation (2) shows a special case, where
the similarities of user u to all other users v are considered. In literature often
only a neighborhood around u of the top N similar users are taken into account.
Similarity Measures A classical similarity measure between two vectors
consisting of binary numbers is the Jaccard Index, which is the proportion between
the sizes of the intersection and the union of two sets.
(3)
wuv = jI(u) \ I(v)j</p>
        <p>jI(u) [ I(v)j
In this context, it means two item-sets are more similar the more common items
they share.</p>
        <p>
          Good and often superior results were reported regarding the \log-likelihood
similarity" which is implemented in the Mahout software package7 for item- and
user-based collaborative ltering [
          <xref ref-type="bibr" rid="ref10 ref3">10, 3</xref>
          ]. In computational linguistics Dunning
proposed the use of the likelihood ratio test to nd rare events as alternative to
traditional contingency table methods [
          <xref ref-type="bibr" rid="ref3">3</xref>
          ]. In a bigram study he aimed at nding
pairs of words that occurred next to each other signi cantly more often than
expected from pure word frequencies. As basis for the log-likelihood statistics he
used the following contingency table (see Table 4).
Both similarity measures can be characterized regarding their use of information
from two binary vectors under consideration. In [
          <xref ref-type="bibr" rid="ref1">1</xref>
          ] a survey of 76 binary
similarity and distance measures had been carried out. All measures were described by
a table called Operational Taxonomic Units that is identical to the contingency
table shown above in Table 4. The measures fall in two groups regarding their
use of negative matches that corresponds to the cell value k22. The Dunning
7 http://mahout.apache.org/ - scalable machine learning libraries
8 for implementation details refer to the Appendix
similarity is not covered in that survey but falls into the category of similarity
measures that make use of that kind of information. In contrast, the Jaccard
similarity belongs to the negative match exclusive measures. The general inclusion
or exclusion of negative matches has been debated over years and the particular
choice of a measure is surely domain dependent.
4.6
        </p>
      </sec>
      <sec id="sec-3-5">
        <title>Results</title>
        <p>
          The MAP performances for the dyadic factor approaches based on MP are given
in Table 5. Among those, the dyadic factor using user geo-locations performs
best. Otherwise, the dyadic factor approaches based on UCF perform clearly
better, if neighborhood sizes of UCF factor pui are larger than 500 (see Figure 3).
The best performance values can be achieved when choosing the neighborhoods
as large as possible. This means to consider all users according to Equation
(2). To conclude, the choice between Dunning and Jaccard similarity for the
UCF factor is not that relevant. In contrast, the neighborhood size is much
more important. However, if not many users are available in the system and
furthermore not much is known about users and items, the MP approach then
provides a solid basis.
We propose to combine the di erent dyadic factors presented in the last section
by extending the basic scheme. Furthermore, experimental results are shown for
di erent factor combinations.
In this section, we show how di erent dyadic factors can be combined into a
uni ed scoring scheme, which is an extension to Equation (1).
(4)
(5)
sui = Dui f u[1;i]fi[
          <xref ref-type="bibr" rid="ref2">2</xref>
          ]
        </p>
        <p>Dui = (d[u1i]) 1 (d[u2i]) 2 : : : (d[umi ]) m
The choice of a dyadic factor combination is governed by the hyperparameter
set = f 1; 2; : : :g, where each parameter has the purpose of weighting the
contribution of a particlar dyadic factor to the score sui.
5.2</p>
      </sec>
      <sec id="sec-3-6">
        <title>Optimization</title>
        <p>The optimization of the hyperparameter set for formula (4) for 8u 2 U and
8i 2 I to maximize the overall MAP value is not trivial, because gradient based
methods are not applicable here. For this reason we used grid search and an
evolutionary algorithm9 to nd a well weighted combination of dyadic factors.
5.3</p>
      </sec>
      <sec id="sec-3-7">
        <title>Experiments and Results</title>
        <p>For a random selection of 1000 test users the following dyadic factors described
in the last section were considered:
{ The demographic factor cui, where the most popular items are recommended
according to the country of a user.
{ The length of a name factor nui.
{ The gender of a name gui.
{ User based Collaborative Filtering factor using Jaccard similarity pui.</p>
        <p>
          According to Table 6 the combination of multiple dyadic factors can be
bene cial. However, regarding the MAP performance only minimal improvements
can be observed.
9 CMA-ES from Apache Commons (http://commons.apache.org).
During the challenge phase we could con rm ndings reported in literature:
The baseline method described in section 4.2 performs well compared to statical
methods using relational data from co-occurrence networks [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ]. We found out
that the choice of input for the recommendation has a large impact on the
performance, e.g. restricting the item sets for the UCF approach a-priori to
names contained in N has a negative e ect. Furthermore, we introduced an
approach based on weighted dyadic factors, which enabled us to combine di erent
information sources and assumptions in a formal way. That allowed us to express
the preferences of users by di erent factors and provides further possibilities, that
are out of the scope of this paper, e.g. to combine User-based with Item-based
Collaborative Filtering. Even though this approach seems to be appealing at rst
sight, the MAP performance improvements are only minimal compared to single
dyadic factors. Introducing the various dyadic factors using side-information,
they performed not as well as factors based on Collaborative Filtering, which
shows the e ectiveness of UCF despite its simplicity.
        </p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Appendix</title>
      <sec id="sec-4-1">
        <title>Calculation of the Entropies for the Dunning Similarity</title>
        <p>The entropies for the Dunning similarities using the notation of Table 4 can be
calculated as follows :</p>
        <p>LLR = 2[H(row) + H(col)</p>
        <p>H(row; col)]
H(row) =
H(col) =
((k11 + k12) log(k11 + k12) + (k21 + k22) log(k21 + k22))
((k11 + k21) log(k11 + k21) + (k12 + k22) log(k12 + k22))
H(row; col) =
((X kij log(X kij )</p>
        <p>(X kij log kij )):</p>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Seung-Seok</surname>
            <given-names>Choi</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sung-Hyuk Cha</surname>
            , and
            <given-names>C</given-names>
          </string-name>
          <string-name>
            <surname>Tappert</surname>
          </string-name>
          .
          <article-title>A survey of binary similarity and distance measures</article-title>
          .
          <source>Journal of Systemics, Cybernetics and Informatics</source>
          ,
          <volume>8</volume>
          (
          <issue>1</issue>
          ):
          <volume>43</volume>
          {
          <fpage>48</fpage>
          ,
          <year>2010</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Abhinandan S Das</surname>
          </string-name>
          ,
          <string-name>
            <surname>Mayur Datar</surname>
            , Ashutosh Garg, and
            <given-names>Shyam</given-names>
          </string-name>
          <string-name>
            <surname>Rajaram</surname>
          </string-name>
          .
          <article-title>Google news personalization: scalable online collaborative ltering</article-title>
          .
          <source>In Proceedings of the 16th international conference on World Wide Web</source>
          , pages
          <volume>271</volume>
          {
          <fpage>280</fpage>
          . ACM,
          <year>2007</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <given-names>Ted</given-names>
            <surname>Dunning</surname>
          </string-name>
          .
          <article-title>Accurate methods for the statistics of surprise and coincidence</article-title>
          .
          <source>Computational linguistics</source>
          ,
          <volume>19</volume>
          (
          <issue>1</issue>
          ):
          <volume>61</volume>
          {
          <fpage>74</fpage>
          ,
          <year>1993</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <given-names>Yi</given-names>
            <surname>Fang</surname>
          </string-name>
          and
          <string-name>
            <given-names>Luo</given-names>
            <surname>Si</surname>
          </string-name>
          .
          <article-title>Matrix co-factorization for recommendation with rich side information and implicit feedback</article-title>
          .
          <source>In Proceedings of the 2nd International Workshop on Information Heterogeneity and Fusion in Recommender Systems</source>
          , pages
          <fpage>65</fpage>
          {
          <fpage>69</fpage>
          . ACM,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <given-names>Z.</given-names>
            <surname>Gantner</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Drumond</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Freudenthaler</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Rendle</surname>
          </string-name>
          , and L.
          <string-name>
            <surname>Schmidt-Thieme</surname>
          </string-name>
          .
          <article-title>Learning attribute-to-feature mappings for cold-start recommendations</article-title>
          .
          <source>In Data Mining (ICDM)</source>
          ,
          <year>2010</year>
          IEEE 10th International Conference on, pages
          <volume>176</volume>
          {
          <fpage>185</fpage>
          . IEEE,
          <year>2010</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <given-names>Michael</given-names>
            <surname>Hahsler</surname>
          </string-name>
          .
          <article-title>Developing and testing top-n recommendation algorithms for 0-1 data using recommenderlab</article-title>
          ,
          <year>2010</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <given-names>Andreas</given-names>
            <surname>Mild</surname>
          </string-name>
          and
          <string-name>
            <given-names>Thomas</given-names>
            <surname>Reutterer</surname>
          </string-name>
          .
          <article-title>An improved collaborative ltering approach for predicting cross-category purchases based on binary market basket data</article-title>
          .
          <source>Journal of Retailing and Consumer Services</source>
          ,
          <volume>10</volume>
          (
          <issue>3</issue>
          ):
          <volume>123</volume>
          {
          <fpage>133</fpage>
          ,
          <year>2003</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <given-names>Folke</given-names>
            <surname>Mitzla</surname>
          </string-name>
          and
          <string-name>
            <given-names>Gerd</given-names>
            <surname>Stumme</surname>
          </string-name>
          .
          <article-title>Namelings - discover given name relatedness based on data from the social web</article-title>
          . In Karl Aberer, Andreas Flache, Wander Jager, Ling Liu,
          <string-name>
            <given-names>Jie</given-names>
            <surname>Tang</surname>
          </string-name>
          , and Christophe Guret, editors,
          <source>SocInfo</source>
          , volume
          <volume>7710</volume>
          of Lecture Notes in Computer Science, pages
          <volume>531</volume>
          {
          <fpage>534</fpage>
          . Springer,
          <year>2012</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <given-names>Folke</given-names>
            <surname>Mitzla</surname>
          </string-name>
          and
          <string-name>
            <given-names>Gerd</given-names>
            <surname>Stumme</surname>
          </string-name>
          .
          <article-title>Relatedness of given names</article-title>
          .
          <source>Human Journal</source>
          ,
          <volume>1</volume>
          (
          <issue>4</issue>
          ):
          <volume>205</volume>
          {
          <fpage>217</fpage>
          ,
          <year>2012</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <given-names>Folke</given-names>
            <surname>Mitzla</surname>
          </string-name>
          and
          <string-name>
            <given-names>Gerd</given-names>
            <surname>Stumme</surname>
          </string-name>
          .
          <article-title>Recommending given names</article-title>
          .
          <source>arXiv preprint arXiv:1302.4412</source>
          ,
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>Rong</surname>
            <given-names>Pan</given-names>
          </string-name>
          , Yunhong Zhou, Bin Cao, Nathan N Liu, Rajan Lukose,
          <string-name>
            <given-names>Martin</given-names>
            <surname>Scholz</surname>
          </string-name>
          , and
          <string-name>
            <given-names>Qiang</given-names>
            <surname>Yang</surname>
          </string-name>
          .
          <article-title>One-class collaborative ltering</article-title>
          .
          <source>In Data Mining</source>
          ,
          <year>2008</year>
          . ICDM'08. Eighth IEEE International Conference on, pages
          <volume>502</volume>
          {
          <fpage>511</fpage>
          . IEEE,
          <year>2008</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <string-name>
            <surname>Ste</surname>
          </string-name>
          en Rendle, Christoph Freudenthaler, Zeno Gantner, and
          <string-name>
            <surname>Lars</surname>
          </string-name>
          Schmidt-Thieme.
          <article-title>Bpr: Bayesian personalized ranking from implicit feedback</article-title>
          .
          <source>In Proceedings of the Twenty-Fifth Conference on Uncertainty in Arti cial Intelligence</source>
          , pages
          <fpage>452</fpage>
          {
          <fpage>461</fpage>
          . AUAI Press,
          <year>2009</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>