<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Incorporating Context Correlation into Context-aware Matrix Factorization</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Yong Zheng</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Bamshad Mobasher</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Robin Burke</string-name>
          <email>rburkeg@cs.depaul.edu</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Center for Web Intelligence, DePaul University Chicago</institution>
          ,
          <addr-line>Illinois</addr-line>
          ,
          <country country="US">USA</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Context-aware recommender systems (CARS) go beyond traditional recommender systems, that only consider users' profiles, by adapting their recommendations also to users' contextual situations. Several contextual recommendation algorithms have been developed by incorporating context into recommendation algorithms in different ways. The most effective approaches try to model deviations in ratings among contexts, but ignore the correlations that may exist among these contexts. In this paper, we highlight the importance of contextual correlations and propose a correlationbased context-aware matrix factorization algorithm. Through detailed experimental evaluation we demonstrate that adopting contextual correlations leads to improved performance.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>Recommender systems (RS) are an effective ways to alleviate
information overload by tailoring recommendations to users’
personal preferences. Context-aware recommender systems
(CARS) emerged to go beyond user preferences and also take
into account the contextual situation of users in generating
recommendations.</p>
      <p>The standard formulation of the recommendation problem
begins with a two-dimens-ional (2D) matrix of ratings,
organized by user and item: Users Items ! Ratings.
The key insight of context-aware recommender systems is
that users’ preferences for items may be a function of the
context in which those items are encountered. Incorporating
contexts requires that we estimate user preferences using
a multidimensional rating function – R: Users Items
Contexts ! Ratings [Adomavicius et al., 2011].</p>
      <p>In the past decade, a number of context-aware
recommendation algorithms have been developed that attempt to
integrate context with recommendation algorithms. Some
of the most effective methods, such as context-aware matrix
factorization (CAMF) [Baltrunas et al., 2011c] and
contextual sparse linear methods (CSLIM) [Zheng et al., 2014b;
2014c], incorporate a contextual rating deviation component
into the recommendation algorithms. However, these
methods generally ignore contextual correlations that may
have bearing on how the rating behavior in different contexts
is modeled. In this paper, we highlight the importance
of contextual correlation, and propose a correlation-based
context-aware matrix factorization algorithm.
2</p>
    </sec>
    <sec id="sec-2">
      <title>Related Work</title>
      <p>In context-aware recommender systems, context is usually
defined as, “any information that can be used to characterize
the situation of an entity [Abowd et al., 1999]”, e.g., time and
companion may be two influential contexts in movie domain.</p>
      <p>According to the two-part classification of contextual
information by Adomavicius, et al. [Adomavicius et al.,
2011], we are concerned with static and fully observable
contexts, where we already have a set of known contextual
variables at hand which remain stable over time, and
try to model users’ contextual preferences to provide
recommendations.</p>
      <p>Several context-aware recommendation algorithms have
been developed in the past decade. Typically, context can be
taken into account using three basic strategies: pre-filtering,
post-filtering and contextual modeling [Adomavicius et al.,
2011]. Pre-filtering techniques, such as context-aware
splitting approaches [Baltrunas and Ricci, 2014; Zheng et
al., 2014a], simply apply contexts as filters beforehand to
filter out irrelevant rating profiles. Post-filtering, on the
other hand, applies the context as filters to the recommended
items after the recommendation process. Or, contexts can
be used as filters in the recommendation process, such as
differential context modeling [Zheng et al., 2012; 2013].
In contextual modeling, predictive models are learned using
the full contextual data, and context information is used
in the recommendation process. Most of the recent work,
such as CAMF [Baltrunas et al., 2011c], tensor factorization
(TF) [Karatzoglou et al., 2010] and CSLIM [Zheng et al.,
2014b; 2014c], belong to contextual modeling.</p>
      <p>The most effective context-aware recommendation
algorithms, such as CAMF and CSLIM, usually incorporate a
contextual rating deviation term which is used to estimate
users’ rating deviations associated with specific contexts.
Alternatively, the contextual correlation could be another
way to incorporate contextual information. This idea has
been introduced in the context of sparse linear method in
our previous work [Zheng et al., 2015], but it has not been
explored as part of recommendation models based on matrix
factorization.
3</p>
      <p>Preliminary: Matrix Factorization and
Context-aware Matrix Factorization
In this section, we introduce matrix factorization used in
recommender systems, as well as the existing research
on Context-Aware Matrix Factorization which utilizes the
contextual rating deviations.
3.1</p>
      <sec id="sec-2-1">
        <title>Matrix Factorization</title>
        <p>Matrix factorization (MF) [Koren et al., 2009] is one of the
most effective recommendation algorithm in the traditional
recommender systems. Simply, both users and items are
represented by vectors. For example, p!u is used to denote
a user vector, and !qi as an item vector. The values in
those vectors indicate the weights on K (e.g., K = 5) latent
factors. As a result, the rating prediction can be described by
Equation 1.</p>
        <p>r^ui = p!u !qi (1)</p>
        <p>More specifically, the weights in p!u can be viewed as how
much users like those latent factors, and the weights in !qi
represent how this specific item obtains those latent factors.
Therefore, the dot product of those two vectors can be used
to indicate how much the user likes this item, where users’
preferences on items are captured by the latent factors.</p>
        <p>In addition, user and item rating biases can be added to the
prediction function, as shown in Equation 2, where denotes
the global average rating in the data set, bu and bi represent
the user bias and item bias respectively.</p>
        <p>r^ui =
+ bu + bi + p!u !qi
(2)
3.2</p>
      </sec>
      <sec id="sec-2-2">
        <title>Context-aware Matrix Factorization</title>
        <p>Consider the movie recommendation example in Table 1.
There is one user U 1, one item T 1, and three contextual
dimensions – Time (weekend or weekday), Location (at
home or cinema) and Companion (alone, girlfriend, family).
In the following discussion, we use contextual dimension
to denote the contextual variable, e.g. “Location”. The
term contextual condition refers to a specific value in a
dimension, e.g. “home” and “cinema” are two contextual
conditions for “Location”. A context or contextual situation
is, therefore, a set of contextual conditions, e.g. fweekend,
home, familyg.</p>
        <p>More specifically, let ck and cm denote two different
contextual situations each of which is composed of a set
of contextual conditions. We use ck;l to denote the lth
contextual condition in the context ck. For example, assume
ck = fweekend, home, aloneg, then ck;2 is the contextual
condition “home”. In the following discussion, we continue
to use those terms and symbols to describe corresponding
contextual dimensions and conditions in the algorithms or
equations.</p>
        <p>The CAMF algorithm was proposed by Baltrunas et
al. [Baltrunas et al., 2011c]. The CAMF rating prediction
function is shown in Equation 3.</p>
        <p>L
+ bu + X</p>
        <p>j=1
r^uick;1ck;2:::ck;L =
Bijck;j + p!u !qi
(3)</p>
        <p>Assume there are L contextual dimensions in total, then
ck = fck;1ck;2:::ck;Lg describes a contextual situation, where
ck;j denotes the contextual condition in the jth context
dimension. Therefore, Bijck;j indicates the contextual rating
deviation associated with item i and the contextual condition
in the jth dimension.</p>
        <p>A comparison between Equation 2 and Equation 3 reveals
that CAMF simply replaces the item bias bi by a contextual</p>
        <p>L
rating deviation term P Bijck;j , where it assumes that the
j=1
contextual rating deviation is dependent on items. Therefore,
this approach is named as CAMF CI. Alternatively, this
deviation can also be viewed as being dependent on users,
which replaces bu by the contextual rating deviation term
resulting in the CAMF CU variant. In addition, CAMF C
algorithm assumes that the contextual rating deviation is
independent of both users and items.</p>
        <p>The parameters, such as the user and item vectors,
user biases and rating deviations, can be learned by the
stochastic gradient descent (SGD) method to minimize
the rating prediction errors. In early work [Baltrunas et
al., 2011c], CAMF was demonstrated to outperform other
contextual recommendation algorithms, such as the tensor
factorization [Karatzoglou et al., 2010].
4</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Correlation-Based CAMF</title>
      <p>Introducing the contextual rating deviation term is an
effective way to build context-aware recommendation
algorithms. Our earlier work [Zheng et al., 2014b; 2014c]
has successfully incorporated the contextual rating deviations
into the sparse linear method (SLIM) to develop contextual
SLIM (CSLIM) which was demonstrated to outperform
the state-of-the-art contextual recommendation algorithms,
including CAMF and tensor factorization.</p>
      <p>As mentioned before, contextual correlation is an
alternative way to build context-aware recommendation algorithms,
other than modeling the contextual deviations. Our recent
work [Zheng et al., 2015] has made the first attempt to
introduce the contextual correlation (or context similarity)
into SLIM. In this paper, we further explore ways to develop
correlation-based CAMF.</p>
      <p>The underlying assumption behind the notion of
“contextual correlation (or context similarity)” is that the more
similar or correlated two contexts are, the more similar two
recommendation lists for the same user within those two
contextual situations should be. When integrated into matrix
Time=Weekend
Time=Weekday
Location=Home
Location=Cinema
where cE denotes the empty contextual situation – the value
in each contextual dimension is empty or “NA”; that is, cE;1 =
cE;2 = ... = cE;L = N/A. Therefore, the function Corr(ck; cE )
estimates the correlation between the cE and the contextual
situation ck where at least one contextual condition is not
empty or “NA”. Note that in Equation 3, the contextual rating
deviation can be viewed as the deviation from the empty
contextual situation to a non-empty contextual situation.</p>
      <p>Accordingly, the user and item vectors, as well as the
contextual correlations can be learned based using stochastic
gradient decsent by minimizing the rating prediction errors.
The loss function is described in Equation 5. Note that this is
the general formulation for the loss function, where the term
“Corr2” should be specified and adjusted accordingly when
the similarity of context is modeled in different ways. We will
introduce three context similarity models in the next section.</p>
      <p>The remaining challenge is how to represent or model the
correlation function in Equation 4. Different representations
may directly influence the recommendation performance.
In our recent work [Zheng et al., 2015], we considered
four strategies: Independent Context Similarity (ICS),
Latent Context Similarity (LCS), Multidimensional Context
Similarity (MCS) and Weighted Jaccard Context Similarity
(WJCS) 1. And those strategies can also be reused for
the correlation-based CAMF. The prediction function in
Equation 4 and loss function in Equation 5 can be updated
accordingly when the correction is represented by different
ways.</p>
      <p>In this paper, we will not consider WJCS since it only
uses the contextual dimensions with the same values, and
the correlation in Equation 4 is measured between a context
and the empty context. WJCS is therefore not applicable to
the correlation-based CAMF, since it only works when the
contextual conditions are the same. In the following, we
introduce and compare the ICS, LCS and MCS approaches
based on CAMF. Note that context correlation or similarity
is assumed to be measured between any two contextual
situations ck and cm. However, in the correlation-based
CAMF, the correlation is actually measured between ck and
cE , where cE is the empty context.</p>
      <p>1Here, context similarity is identical to context correlation
4.1</p>
      <sec id="sec-3-1">
        <title>Independent Context Similarity (ICS)</title>
        <p>An example of a correlation matrix can be seen in Table 2.
With Independent Context Similarity, we only measure the
context correlation or similarity between two contextual
conditions when they lie on the same contextual dimension,
e.g., we never measure the correlation between “Time =
Weekend” and “Location = Home”, since they are from two
different dimensions. Each pair of contextual dimensions
are assumed to be independent. In this case, the correlation
between two contexts can be represented by the product of
the correlations among different dimensions. For example,
assume ck is fTime = Weekend, Location = Homeg, and cm
is fTime = Weekday, Location = Cinemag, the correlation
between ck and cm can be represented by the correlation of
&lt;Time = Weekend, Time = Weekday&gt; multiplied by the
correlation of &lt;Location = Home, Location = Cinema&gt;,
since those two dimensions are assumed as independent.</p>
        <p>Assuming there are L contextual dimensions in total,
the correlations can be depicted by Equation 6, where
ck;l is used to denote the value of contextual condition
in the lth dimension in context ck, and the “correlation”
function is used to represent the correlation between two
contextual conditions, which is also what to be learnt in
the optimization. In other words, the correlation between
two contexts is represented by the multiplication of the
individual correlations between contextual conditions on each
dimension.</p>
        <p>Corr(ck; cm) =</p>
        <p>L
Y correlation(ck;l; cm;l)
l=1
(6)</p>
        <p>These correlation values (i.e., correlation(ck;l; cm;l)) can
be learned by the optimization process accordingly. The
risk of this representation is that some information may be
lost, if correlations are not in fact independent in different
dimensions. For example, if users usually go to cinema
to see romantic movies with their partners, the “Location”
(e.g. at cinema) and “Companion” (e.g. partners) may be
significantly correlated as a result.
4.2</p>
      </sec>
      <sec id="sec-3-2">
        <title>Latent Context Similarity (LCS)</title>
        <p>As noted earlier, contextual rating data is often sparse, since
it is somewhat unusual to have users rate items repeatedly
within multiple contextual situations. This poses a difficulty
when new contexts are encountered. For example, the
correlation between a new pair of contextual conditions
&lt;“Time=Weekend”, “Time=Holiday”&gt; may be required in
the testing set, but it may not have been learned from the
training data due to the sparsity problem. But, the correlation for
two existing pairs, &lt;“Time=Weekend”, “Time=Weekday”&gt;
and &lt;“Time=Weekday”, “Time=Holiday”&gt;, may have been
!"#$%!&amp;
!"#$%!&amp;
%&amp;'#"($(0,0,4)
%&amp;'#"($(0,0,4)
/("&amp;0,$(0,6,0)
*!(+#&amp;%!&amp;
)##*+(,$(2.5,0,0)
!!"#$%&amp;(2.5,0,0)
'%()
)##*#'+$(5,0,0)
'%()
!!"!#$%(5,0,0)
learned. In this case, this representation suffers from the
contextual rating sparsity problem. Treating each dimension
independently prevents the algorithm from taking advantage
of comparisons that might be made across dimensions.</p>
        <p>To alleviate this situation, we represent each contextual
condition by a vector of weights over a set of latent
factors (we use 5 latent factors in our experiments), where
the weights are initialized at the beginning and learnt by
the optimization process. The dot product between two
vectors can be used to denote the correlation between each
pair of contextual conditions. As a result, even if the
newly observed pair does not exist in the training data, the
weights in the vectors representing the two conditions (i.e.,
“Time=Weekend” and “Time=Holiday”) will be learned and
updated by the learning process over existing pairs, and the
correlation for the new pair can be easily computed using the
dot product. The correlation is given by
correlation(ck;l; cm;l) = Vck;l
Vcm;l
(7)
where Vck;l and Vcm;l denote the vector representation for
the contextual condition ck;l and cm;l, respectively, over the
space of latent factors. We then use the same correlation
calculation as shown in Equation 6. We call this approach the
Latent Context Similarity (LCS) model. This approach was
able to improve the performance of deviation-based CSLIM
algorithms too [Zheng, 2015]. In contrast to the independent
context similarity approach, LCS provides more flexibility,
but it also has the added computational costs associated with
learning the latent factors. In LCS, what has to be learnt in the
optimization process are the vectors of weights representing
each contextual condition.
4.3</p>
        <p>Multidimensional Context Similarity (MCS)
In the multidimensional context similarity model, we
assume that contextual dimensions form a multidimensional
coordinate system. An example is depicted in Figure 1.</p>
        <p>Let us assume that there are three contextual dimensions:
time, location and companion. We assign a real value to
each contextual condition in those dimensions, so that each
condition can locate a position in the corresponding axis.
In this case, a context (as a set of contextual conditions)
can be viewed as a point in the multidimensional space.
Accordingly, the distance between two such points can
be used as the basis for a correlation measure. In this
approach, the real values for each contextual condition are
the parameters to be learned in the optimization process. For
example, the values for “family” and “kids” are updated in the
right-hand side of the figure. Thus, the position of the data
points associated to those two contextual conditions will be
changed, as well as the distance between the corresponding
two contexts. Therefore, the correlation can be measured
as the inverse of the distance between two data points. In
our experiments, we use Euclidean distance to measure the
distances, though other distance measures can also be used.
The computational cost is directly associated with the number
of contextual conditions in the data set, which may make
this approach the highest-cost model. Again, the number of
contextual conditions can be reduced by context selection.
5</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Experimental Evaluation</title>
      <p>In this section, we present our experimental evaluation and
discuss the results.
5.1</p>
      <sec id="sec-4-1">
        <title>Data Sets</title>
        <p>We select three context-aware data sets with different
numbers of contextual dimensions and conditions.
Restaurant data [Ramirez-Garcia and Garca-Valdez, 2014] is
comprised of users’ ratings on restaurants in city of Tijuana,
Mexico. Music data [Baltrunas et al., 2011a] captures
users’ ratings on music tracks in different driving and
traffic conditions. The Tourism data [Baltrunas et al.,
2011b] collects users’ places of interest (POIs) from mobile
applications. The characteristics of these data sets are
summarized in Table 3. For more specific information about
the contextual dimensions and conditions, please refer to the
original papers using those data sets.
!"%#
!"%
We use a five-fold cross validation on our data sets,
performing top 10 recommendation task and using precision
and mean average precision (MAP) as the evaluation metrics.
Precision is defined as the ratio of relevant items selected to
number of items recommended in a specific context. MAP
is another popular ranking metric which additional takes
the ranks of the recommended items into consideration. It
is calculated by Equation 8, where M denotes the number
of the users, and N is the size of the recommendation
list, where P (k) means the precision at cut-off k in the
item recommendation list, i.e., the ratio of number of users
followed up to the position k over the number k, where m in
ap@N denotes the number of relevant items.</p>
        <p>N
P P (k)
k=1
min(m; N )
(8)</p>
        <p>We choose the CAMF [Baltrunas et al., 2011c] as
baseline in our experiments. We tried all three versions:
CAMF CI, CAMF CU and CAMF C, and only present the
best performing one which is denoted as CAMF-Dev in
the following sections. In addition, we also add tensor
factorization (TF) [Karatzoglou et al., 2010] as baseline.
5.3</p>
      </sec>
      <sec id="sec-4-2">
        <title>Analysis and Findings</title>
        <p>The results are depicted by Figure 2, where the
correlationbased CAMF are represented by three algorithms
accordingly: CAMF-ICS, CAMF-LCS and CAMF-MCS.</p>
        <p>In terms of the best performing algorithm, CAMF-MCS
outperforms the other ones for the restaurant and tourism data
sets. It is able to obtain comparable results with
CAMFLCS in the music data. Compared with the deviation-based
CAMF, the correlation-based CAMF can always outperform
CAMF-Dev if the appropriate correlation modeling is
applied. For example, in the restaurant data, CAMF-Dev
works better than CAMF-ICS in precision, but CAMF-MCS</p>
        <p>Restaurant
CAMF-Dev CAMF-MCS CSLIM-Dev CSLIM-MCS CAMF-Dev CAMF-MCS
0.1309 0.1586 0.2044 0.2151 0.0210 0.0385
0.1130 0.1276 0.1496 0.1723 0.0226 0.0361
0.2001 0.2765 0.2889 0.2993 0.0474 0.0661
0.2122 0.2840 0.3128 0.3187 0.0542 0.0747
outperforms CAMF-Dev significantly, which states that MCS
is the better representation to model contextual correlations
than the ICS for this data set.</p>
        <p>Tensor Factorization (TF) only works better than some
CAMF algorithms in the tourism data set, but CAMF-MCS
is still the best one for this data. In TF, contextual dimensions
are considered as extra dimensions in addition to the user and
item dimensions, and they are assumed as independent with
each other, where CAMF algorithms are actually dependent
algorithms which further measures either contextual rating
deviations or contextual correlations.</p>
        <p>In our previous work [Zheng et al., 2014c; 2015], we have
incorporated contextual deviations and correlations into
SLIM respectively to formulate CSLIM algorithms. Therefore,
it is necessary to compare the CAMF and CSLIM, which
can be viewed from Table 4, where we only present the best
performing deviation and correlation models. From the table,
we can see that CSLIM outperforms CAMF significantly
when the same strategy (either deviation or correlation
modeling) is applied. And CSLIM-MCS works the best
in general. It is not surprising, since earlier work [Ning
and Karypis, 2011] has demonstrated that SLIM is able to
outperform the state-of-the-art traditional recommendation
algorithms, including the matrix factorization. Those results
based on contextual recommendations further confirms this
pattern, since CSLIM outperforms CAMF in view of the
Table 4.</p>
        <p>In short, those experimental results demonstrate that
correlation-based CAMF is able to outperform
deviationbased CAMF and the TF algorithm, but the representation
for contextual correlation should be carefully selected.
Generally, the multidimensional context similarity is the best
choice to represent contextual correlations.
6</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>Conclusions and Future Work</title>
      <p>In this paper, we highlighted the importance of
contextual correlation and incorporate this notion into the
matrix factorization technique to formulate correlation-based
Context-Aware Matrix Factorization (CAMF) algorithm. Our
experimental results reveal that correlation-based CAMF is
able to outperform the standard deviation-based CAMF and
the tensor factorization algorithms. In our future work, we
plan to incorporate contextual correlation modeling strategies
into more recommendation algorithms, such as the slope one
recommender.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [Abowd et al.,
          <year>1999</year>
          ]
          <string-name>
            <given-names>Gregory D</given-names>
            <surname>Abowd</surname>
          </string-name>
          , Anind K Dey,
          <string-name>
            <surname>Peter J Brown</surname>
            , Nigel Davies,
            <given-names>Mark</given-names>
          </string-name>
          <string-name>
            <surname>Smith</surname>
            ,
            <given-names>and Pete</given-names>
          </string-name>
          <string-name>
            <surname>Steggles</surname>
          </string-name>
          .
          <article-title>Towards a better understanding of context and context-awareness</article-title>
          .
          <source>In Handheld and ubiquitous computing</source>
          , pages
          <fpage>304</fpage>
          -
          <lpage>307</lpage>
          . Springer,
          <year>1999</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [Adomavicius et al.,
          <year>2011</year>
          ]
          <string-name>
            <given-names>Gediminas</given-names>
            <surname>Adomavicius</surname>
          </string-name>
          , Bamshad Mobasher, Francesco Ricci, and
          <string-name>
            <given-names>Alexander</given-names>
            <surname>Tuzhilin</surname>
          </string-name>
          .
          <article-title>Context-aware recommender systems</article-title>
          .
          <source>AI Magazine</source>
          ,
          <volume>32</volume>
          (
          <issue>3</issue>
          ):
          <fpage>67</fpage>
          -
          <lpage>80</lpage>
          ,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          <source>[Baltrunas and Ricci</source>
          , 2014]
          <string-name>
            <given-names>Linas</given-names>
            <surname>Baltrunas</surname>
          </string-name>
          and
          <string-name>
            <given-names>Francesco</given-names>
            <surname>Ricci</surname>
          </string-name>
          .
          <article-title>Experimental evaluation of context-dependent collaborative filtering using item splitting. User Modeling</article-title>
          and
          <string-name>
            <surname>User-Adapted</surname>
            <given-names>Interaction</given-names>
          </string-name>
          ,
          <volume>24</volume>
          (
          <issue>1-2</issue>
          ):
          <fpage>7</fpage>
          -
          <lpage>34</lpage>
          ,
          <year>February 2014</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [Baltrunas et al., 2011a]
          <string-name>
            <given-names>Linas</given-names>
            <surname>Baltrunas</surname>
          </string-name>
          , Marius Kaminskas, Bernd Ludwig, Omar Moling, Francesco Ricci, Aykan Aydin,
          <article-title>Karl-Heinz Lu¨ke, and Roland Schwaiger</article-title>
          . Incarmusic:
          <article-title>Context-aware music recommendations in a car</article-title>
          .
          <source>In E-Commerce and Web Technologies</source>
          , pages
          <fpage>89</fpage>
          -
          <lpage>100</lpage>
          . Springer,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [Baltrunas et al., 2011b]
          <string-name>
            <given-names>Linas</given-names>
            <surname>Baltrunas</surname>
          </string-name>
          , Bernd Ludwig, Stefan Peer, and
          <string-name>
            <given-names>Francesco</given-names>
            <surname>Ricci</surname>
          </string-name>
          .
          <article-title>Context-aware places of interest recommendations for mobile users</article-title>
          .
          <source>In Design, User Experience, and Usability. Theory, Methods, Tools and Practice</source>
          , pages
          <fpage>531</fpage>
          -
          <lpage>540</lpage>
          . Springer,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [Baltrunas et al., 2011c]
          <string-name>
            <given-names>Linas</given-names>
            <surname>Baltrunas</surname>
          </string-name>
          , Bernd Ludwig, and
          <string-name>
            <given-names>Francesco</given-names>
            <surname>Ricci</surname>
          </string-name>
          .
          <article-title>Matrix factorization techniques for context aware recommendation</article-title>
          .
          <source>In Proceedings of the fifth ACM conference on Recommender systems</source>
          , pages
          <fpage>301</fpage>
          -
          <lpage>304</lpage>
          . ACM,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [Karatzoglou et al.,
          <year>2010</year>
          ]
          <string-name>
            <given-names>Alexandros</given-names>
            <surname>Karatzoglou</surname>
          </string-name>
          , Xavier Amatriain, Linas Baltrunas, and
          <string-name>
            <given-names>Nuria</given-names>
            <surname>Oliver</surname>
          </string-name>
          .
          <article-title>Multiverse recommendation: n-dimensional tensor factorization for context-aware collaborative filtering</article-title>
          .
          <source>In Proceedings of the fourth ACM conference on Recommender systems</source>
          , pages
          <fpage>79</fpage>
          -
          <lpage>86</lpage>
          . ACM,
          <year>2010</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [Koren et al.,
          <year>2009</year>
          ]
          <string-name>
            <given-names>Yehuda</given-names>
            <surname>Koren</surname>
          </string-name>
          , Robert Bell, and
          <string-name>
            <given-names>Chris</given-names>
            <surname>Volinsky</surname>
          </string-name>
          .
          <article-title>Matrix factorization techniques for recommender systems</article-title>
          .
          <source>IEEE Computer</source>
          ,
          <volume>42</volume>
          (
          <issue>8</issue>
          ):
          <fpage>30</fpage>
          -
          <lpage>37</lpage>
          ,
          <year>2009</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          <source>[Ning and Karypis</source>
          , 2011]
          <string-name>
            <given-names>Xia</given-names>
            <surname>Ning</surname>
          </string-name>
          and
          <string-name>
            <given-names>George</given-names>
            <surname>Karypis</surname>
          </string-name>
          . SLIM:
          <article-title>Sparse linear methods for top-n recommender systems</article-title>
          .
          <source>In 2011 IEEE 11th International Conference on Data Mining</source>
          , pages
          <fpage>497</fpage>
          -
          <lpage>506</lpage>
          . IEEE,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [
          <string-name>
            <surname>Ramirez-Garcia and</surname>
          </string-name>
          Garca-Valdez,
          <year>2014</year>
          ]
          <article-title>Xochilt RamirezGarcia and Mario Garca-Valdez. Post-filtering for a restaurant context-aware recommender system</article-title>
          .
          <source>In Recent Advances on Hybrid Approaches for Designing Intelligent Systems</source>
          , volume
          <volume>547</volume>
          , pages
          <fpage>695</fpage>
          -
          <lpage>707</lpage>
          . Springer,
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [Zheng et al.,
          <year>2012</year>
          ]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zheng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Burke</surname>
          </string-name>
          , and
          <string-name>
            <given-names>B.</given-names>
            <surname>Mobasher.</surname>
          </string-name>
          <article-title>Differential context relaxation for context-aware travel recommendation</article-title>
          .
          <source>In E-Commerce and Web Technologies</source>
          , pages
          <fpage>88</fpage>
          -
          <lpage>99</lpage>
          . Springer,
          <year>2012</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [Zheng et al.,
          <year>2013</year>
          ]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zheng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Burke</surname>
          </string-name>
          , and
          <string-name>
            <given-names>B.</given-names>
            <surname>Mobasher</surname>
          </string-name>
          .
          <article-title>Recommendation with differential context weighting</article-title>
          .
          <source>In User Modeling, Adaptation, and Personalization</source>
          , pages
          <fpage>152</fpage>
          -
          <lpage>164</lpage>
          .
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [Zheng et al., 2014a]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zheng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Burke</surname>
          </string-name>
          , and
          <string-name>
            <given-names>B.</given-names>
            <surname>Mobasher</surname>
          </string-name>
          .
          <article-title>Splitting approaches for context-aware recommendation: An empirical study</article-title>
          .
          <source>In Proceedings of the 29th Annual ACM Symposium on Applied Computing</source>
          , pages
          <fpage>274</fpage>
          -
          <lpage>279</lpage>
          . ACM,
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [Zheng et al., 2014b]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zheng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Mobasher</surname>
          </string-name>
          , and
          <string-name>
            <given-names>R.</given-names>
            <surname>Burke</surname>
          </string-name>
          . CSLIM:
          <article-title>Contextual SLIM recommendation algorithms</article-title>
          .
          <source>In Proceedings of the 8th ACM Conference on Recommender Systems</source>
          , pages
          <fpage>301</fpage>
          -
          <lpage>304</lpage>
          . ACM,
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [Zheng et al., 2014c]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zheng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Mobasher</surname>
          </string-name>
          , and
          <string-name>
            <given-names>R.</given-names>
            <surname>Burke</surname>
          </string-name>
          .
          <article-title>Deviation-based contextual SLIM recommenders</article-title>
          .
          <source>In Proceedings of the 23rd ACM Conference on Information and Knowledge Management</source>
          , pages
          <fpage>271</fpage>
          -
          <lpage>280</lpage>
          . ACM,
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [Zheng et al.,
          <year>2015</year>
          ]
          <string-name>
            <given-names>Yong</given-names>
            <surname>Zheng</surname>
          </string-name>
          , Bamshad Mobasher, and
          <string-name>
            <given-names>Robin</given-names>
            <surname>Burke</surname>
          </string-name>
          .
          <article-title>Integrating context similarity with sparse linear recommendation model</article-title>
          .
          <source>In User Modeling, Adaptation, and Personalization</source>
          , volume
          <volume>9146</volume>
          of Lecture Notes in Computer Science, pages
          <fpage>370</fpage>
          -
          <lpage>376</lpage>
          . Springer Berlin Heidelberg,
          <year>2015</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [Zheng, 2015]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zheng</surname>
          </string-name>
          .
          <article-title>Improve general contextual SLIM recommendation algorithms by factorizing contexts</article-title>
          .
          <source>In Proceedings of the 30th Annual ACM Symposium on Applied Computing</source>
          , pages
          <fpage>929</fpage>
          -
          <lpage>930</lpage>
          . ACM,
          <year>2015</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>