=Paper=
{{Paper
|id=None
|storemode=property
|title=The Role of Emotions in Context-aware Recommendation
|pdfUrl=https://ceur-ws.org/Vol-1050/paper4.pdf
|volume=Vol-1050
|dblpUrl=https://dblp.org/rec/conf/recsys/ZhengMB13
}}
==The Role of Emotions in Context-aware Recommendation==
The Role of Emotions in Context-aware Recommendation Yong Zheng, Robin Burke, Bamshad Mobasher Center for Web Intelligence School of Computing, DePaul University Chicago, Illinois, USA {yzheng8, rburke, mobasher}@cs.depaul.edu ABSTRACT recommender system domain are the ones relevant to users’ subject Context-aware recommender systems try to adapt to users’ pref- moods or feelings, for example, user mood when listening to music erences across different contexts and have been proven to provide tracks (e.g. happy, sad, aggressive, relaxed, etc) [19, 9], or users’ better predictive performance in a number of domains. Emotion is emotions after seeing a movie (e.g. user may feel sad after seeing one of the most popular contextual variables, but few researcher- a tragic movie) [13]. s have explored how emotions take effect in recommendations – Emotional context was first exploited by Gonzalez et al [8] for especially the usage of the emotional variables other than the ef- the recommender system domain. This work was followed by oth- fectiveness alone. In this paper, we explore the role of emotions in ers [9, 18, 13, 17] considering emotions as contexts in CARS re- context-aware recommendation algorithms. More specifically, we search. While the effectiveness of emotions as contextual variables evaluate two types of popular context-aware recommendation algo- is therefore well-established, there is little research that has exam- rithms – context-aware splitting approaches and differential context ined specifically the role that these emotional variables play. We modeling. We examine predictive performance, and also explore define "the role of emotions" as the concerns from two aspects – the usage of emotions to discover how emotional features interact whether emotions are useful or effective to improve recommenda- with those context-aware recommendation algorithms in the rec- tion performance? And, the usage of emotions in the recommen- ommendation process. dation process – how emotions are used in the recommendation algorithms, e.g. which emotional variables are selected? which algorithm components are they applied to? and so forth. Current- Categories and Subject Descriptors ly, most research are focused on demonstrating the effectiveness of H.3.3 [Information Search and Retrieval]: Information filtering emotional variables, and few research further explore the usage of the emotions in the recommendation process. General Terms In this paper, we explore the role of emotions by two classes of popular context-aware recommendation algorithms – context- Algorithm, Experiment, Performance aware splitting and differential context modeling. Since both of these approaches require that the algorithm learn the importance Keywords and utility of different contextual features, they help reveal how Recommendation, Context, Context-aware recommendation, Emo- emotions work in each algorithm and what roles they can play in tion, Affective recommender system recommendation. The purpose of this study is therefore to address the following 1. INTRODUCTION research questions: Advances in affective computing have enabled recommender sys- 1. Emotional Effect: Are emotions useful contextual variables tems to take advantage of emotions and personality, leading to the for context-aware recommendation? development of affective recommender systems (ARS) [18]. At the same time, the emerging technique of context-aware recom- 2. Algorithm Comparison: What algorithms are best suited to mender systems (CARS) takes contexts into consideration, convert- make use of emotional variables? Do they outperform the ing a two-dimensional matrix of ratings organized by user and item: baseline algorithms? Users × Items → Ratings, into a multidimensional rating space [1]. 3. Usage of Contexts: How do emotional variables compete CARS have been demonstrated to be effective in a variety of appli- with other contextual variables, such as location and time? cations and domains [4, 15, 10, 24, 13]. Emotional variables are often included as contexts in CARS, which enables the further de- 4. Roles: How can we understand the specific roles emotional velopment of both ARS and CARS. Typical emotional contexts in variables can play in those context-aware recommendation algorithms? 2. RELATED WORK RecSys’13, October 12–16, 2013, Hong Kong, China. Gonzalez et al [8] explored emotional context in recommender sys- Paper presented at the 2013 Decisions@RecSys workshop in conjunc- tems in 2007. They pointed out that,"emotions are crucial for us- tion with the 7th ACM conference on Recommender Systems. Copyright er’s decision making in recommendation process. The users always c 2013 for the individual papers by the papers’ authors. Copying permitted for private and academic purposes. This volume is published and copyright- transmit their decisions together with emotions." With the rapid de- ed by its editors. velopment of context-aware recommender systems, emotion turns 21 out to be one of important and popular contexts in different kinds (the first two rows) in the training data and one unknown rating that of domains, especially in the music and movie domain. For mu- we are trying to predict (the third row). There are three contextual sic recommendation, the emotional context is appealing because it dimensions – time (weekend or weekday), location (at home or cin- can be used to establish a bridge between music items and item- ema) and companion (friend, girlfriend or family). In the following s from other different domains, and perform cross-media recom- discussion, we use contextual dimension to denote the contextu- mendations [6, 1, 9]. Movie recommendation is another domain al variable, e.g. "Location" in this example. The term contextual where emotion turns out to be popular in recent years. In 2010, condition refers to a specific value in a contextual dimension, e.g. Yue et al [16] produced the overall winner in a recent challenge "home" and "cinema" are two contextual conditions for "Location". on context-aware movie recommendation by mining mood-specific movie similarity with matrix factorization. More recent research 3.1 Item Splitting [18, 13] motivates the tendency of taking emotions as contexts Item splitting tries to find a contextual condition on which to split to assist contextual recommendations. Research has demonstrat- each item. The split should be performed once the algorithm iden- ed that emotions can be influential contextual variables in making tifies a contextual condition in which items are rated significantly recommendations, but few of them explore how emotions interact differently. In the movie example above, there are three contextu- with recommendation algorithm – the usage of emotional variables al conditions in the dimension companion: friend, girlfriend and in the recommendation process. family. Correspondingly, there are three possible alternative con- As described in [1], there are three basic approaches to devel- ditions: "friend and not friend", "girlfriend and not girlfriend", op context-aware recommendation algorithms: pre-filtering, post- "family and not family". Impurity criteria [4] are used to deter- filtering, and contextual modeling. A pre-filtering approach applies mine whether and how much items were rated differently in these a context-dependent criterion to the list of items, selecting those alternative conditions, for example a t-test or other statistical met- appropriate to a given context. Only the filtered items are con- ric can be used to evaluate if the means differ significantly across sidered for recommendation. A post-filtering approach is similar conditions. but applies the filter after recommendations have been computed. Item splitting iterates over all contextual conditions in each con- Contextual modeling takes contextual considerations into the rec- text dimension and evaluates the splits based on the impurity cri- ommendation algorithm itself. In this paper, we explore context- teria. It finds the best split for each item in the rating matrix and aware splitting approaches (a class of pre-filtering algorithms), and then items are split into two new ones, where contexts are elimi- differential context modeling (a contextual modeling approach.) nated from the original matrix – it transforms the original multi- The remainder of this paper is organized as follows. In Sec- dimensional rating matrix to a 2D matrix as a result. Assume that tion 3 and Section 4, we formally introduce the two types of recom- the best contextual condition to split item T1 in Table 1 is "Loca- mendation algorithms we are studying, including their capability to tion = home and not home", T1 can be split into T11 (movie T1 capture the role of emotional contexts in recommendation process. being seen at home) and T12 (movie T1 being seen not at home). Sections 5 and 6 discuss the experimental evaluations and mining Once the best split has been identified, the rating matrix can be the role of emotions through those context-aware recommendation transformed as shown by Table 2(a). algorithms, followed by the conclusions and future work in Sec- This example shows a simple split, in which a single contextual tion 7. condition is used to split the item. It is also possible to perform a complex split using multiple conditions across multiple context di- 3. CONTEXT-AWARE SPLITTING mensions. However, as discussed in [5], there are significant costs of sparsity and potential overfitting when using multiple conditions. APPROACHES We use only simple splitting in this work. Contextual pre-filtering is popular as it is straightforward to imple- ment, and can be applied with most recommendation techniques. Table 2: Transformed Rating Matrix Item splitting [4, 5] is considered one of the most efficient pre- filtering algorithms and it has been well developed in recent re- (a) by Item Splitting (b) by User Splitting search. The underlying idea of item splitting is that the nature of an item, from the user’s point of view, may change in different User Item Rating User Item Rating contextual conditions, hence it may be useful to consider it as t- U1 T11 3 U12 T1 3 wo different items [5]. User splitting [2, 15] is based on a similar U1 T12 5 U12 T1 5 intuition – it may be useful to consider one user as two different U1 T11 ? U11 T1 ? users, if he or she demonstrates significantly different preferences (c) by UI Splitting across contexts. In this section, we introduce those two approaches User Item Rating and also propose a new splitting approach – UI splitting, a simple combination of item and user splitting. U12 T11 3 To better understand and represent the splitting approaches, con- U12 T12 5 sider the following movie recommendation example: U11 T11 ? Table 1: Movie Ratings in Contexts 3.2 User Splitting User Item Rating Time Location Companion Similarly, user splitting tries to split users instead of items. It can U1 T1 3 Weekend Home Friend be easily derived from item splitting as introduced above. Similar U1 T1 5 Weekend Cinema Girlfriend impurity criteria can also be used for user splits. Assume that the U1 T1 ? Weekday Home Family best split for user U 1 in Table 1 is "Companion = family and not family", U 1 can be split into U 11 (U 1 saw the movie with family) In Table 1, there are one user U 1, one item T 1 and two ratings and U 12 (U 1 saw the movie with others). The rating matrix can 22 be transformed as shown by Table 2(b). The first two rows contain UBCF, where the four components we separate in UBCF can be the same user U 12 because U 1 saw this movie with others (i.e. not described as follows: family) rather than family as shown in the original rating matrix. 3.3 UI Splitting UI splitting is a new approach proposed in this paper – it applies item splitting and user splitting together. Assuming that the best split for item and user splitting are the same as described above, the rating matrix based on UI splitting can be shown as Table 2(c). Here we see that both users and items were transformed, creating new users and new items. Thus, given n items, m users, k contextual dimensions and d dis- tinct conditions for each dimension, the time complexity for those three splitting approaches is the same as O(nmkd) [5]. The pro- Figure 1: Algorithm Components in UBCF cess in UI splitting is simply an application of item splitting fol- lowed by user splitting on the resulting output. We keep the con- textual information in the matrix after the transformation by item Figure 1 shows the original rating prediction by the well-known splitting so that user splitting can be performed afterwards. Resnick’s UBCF algorithm [14], where a is a user, i is an item, and N is a neighborhood of K users similar to a. The algorithm 3.4 Role of Emotions in Splitting calculates Pa,i , which is the predicted rating that user a is expected Context-aware splitting approaches have been demonstrated to im- to assign to item i. Then we decompose it to four components: prove predictive performance in previous research, but few of those research results report on the details of the splitting process. More Neighborhood selection UBCF applies the well-known k nearest- specifically, it is useful to know which contextual dimensions and neighbor (kNN) approach, where we can select the top-k conditions are selected, and the statistics related to those selection- neighbors from users who have rated on the same item i. If s. For our purposes, it is possible to explore how emotions interact contexts are taken into consideration, the neighborhood can with the splitting process by the usage of contexts in the splitting be further restricted so that users have also rated the item i process, which may help discover the role of emotions from this in the same contexts. This gives a context-specific recom- perspective. In this paper, we try all three context-aware splitting mendation computation. However, the strict application of approaches, empirically compare their predictive performance, and such a filter greatly increases the sparsity associated with us- also explore the distributions of the usage of contexts (emotional er comparisons. There may only be a small number of cases variables included) for splitting operations. in which recommendations can be made. DCM offers two potential solutions to this problem. DCR searches for the optimal relaxation of the context, generalizing the set of con- 4. DIFFERENTIAL CONTEXT MODELING textual features and contextual conditions to reduce sparsity. Differential context modeling (DCM) is a general contextual rec- DCW uses weighting to increase the influence of neighbors ommendation framework proposed in [23]. It can be applied to in similar contexts. any recommendation algorithm. The "differential" part of the tech- nique tries to break down a recommendation algorithm into differ- Neighbor contribution The neighbor contribution is the difference ent functional components to which contextual constraints can be between a neighbor’s rating on an item i and his or her av- applied. The contextual effect for each component is maximized, erage rating over all items. Context relaxation and contex- and the joint effects of all components contribute the best perfor- t weighting can be applied to the computation of r̄u . This mance for the whole algorithm. The "modeling" part is focused on computation is replaced by one in which this average is com- how to model the contextual constraints. There are two approach- puted over ratings from a relaxed set of contexts in DCR, or es: context relaxation and context weighting, where context relax- the average rating can be aggregated by a weighted average ation uses an optimal subset of contextual dimensions, and context across similar contexts under DCW. weighting assigns different weights to each contextual factor. Ac- cordingly, we have two approaches in DCM: differential context User baseline The computation of r̄a is similar to the neighbor’s relaxation (DCR) [21, 22, 20] and differential context weighting average rating and can be made context-dependent in the (DCW) [24]. same way. 4.1 DCM in UBCF User similarity The computation of neighbor similarity sim(a, u) We have successfully applied DCM to user-based collaborative fil- involves identifying ratings ru,i and ra,i where the users tering (UBCF), item-based collaborative filtering (IBCF) and Slope- have rated items in common. For context-aware recommen- One recommendation algorithms in our previous research [23], show- dation, we can additionally add contextual constraints to this ing that DCM is an efficient approach to improve context-aware part: use ratings given in matching contexts with context re- prediction accuracy. However, we did not explore much about the laxation or ratings weighted by contextual similarity using modeling part in our previous work; that is, how contexts are se- context weighting. lected, relaxed or weighted. Recall that we separate different functional components from the With these considerations in mind, we can derive a new rating recommendation algorithm, therefore, how the algorithms relax or prediction formula by applying DCR or DCW to UBCF. More de- weigh contextual variables can tell the role of the contexts in diff- tails about the prediction equations and technical specifications can erent components. In this paper, we continue to apply DCM to be found in [24]. 23 4.2 The Role of Emotions in DCM Usually, a threshold for the splitting criteria should be set so that One advantage of DCM is that it allows us to explore the role of users or items are only be split when the criteria meets the signifi- contexts in each algorithm component. DCM seeks to optimize the cance requirement. We use an arbitrary value of 0.2 in the tIG case. contribution of context in each component, and so, the output of For tmean , tchi and tprop , we use 0.05 as the p-value threshold. A the optimization phase, in which contextual dimensions are select- finer-grained operation is to set another threshold for each impurity ed and/or weighted, can reveal the relative importance of different value and each data set. We deem it as a significant split once the contextual features in different algorithm components. In this pa- p-value is no larger than 0.05. We rank all significant splits by the per, we provide comparisons and also explore the role of emotions impurity value, and we choose the top first (highest impurity) as the in these two classes of context-aware recommendation algorithms. best split. Items or users without qualified splitting criteria are left unchanged. In DCM, we used the same configuration in our previous work 5. EXPERIMENTAL SETUP [24]: we select the top-10 neighbors in UBCF, choose 100 as the In this section, we introduce the data sets, evaluation protocols and maximal iteration in the optimization process, and Pearson Corre- the specific configurations in our experiments. lation is used as the user similarity measure. See our previous work for more details. 5.1 Data Sets We examine context-aware splitting approaches and DCM with the 5.3 Evaluation Protocols real-world data set LDOS-CoMoDa, which is a movie dataset col- For the evaluation of predictive performance, we choose the root lected from surveys and used by Odic et al [12, 13, 17]. After fil- mean square error (RMSE) evaluated using the 5-fold cross vali- tering out subjects with less than 5 ratings and rating records with dation. The data set is relatively small and some subjects in the incomplete feature information, we got the final data sets contain- survey were required to rate specific movies, thus precision and re- ing 113 users, 1186 items, 2094 ratings, 12 contextual dimensions call may be not a good metric, but we plan to evaluate them and and the rating scale is 1 to 5. We created five folds for cross val- other metrics in our future work. idation purposes and all algorithms are evaluated on the same five We applied DCM only to user-based collaborative filtering. For folds. Specific descriptions of all the contextual dimensions and splitting approaches, there are more options – we evaluate the per- conditions can be found in previous work [12] and the description formance of three recommendation algorithms: user-based collab- of the data set online 1 . orative filtering (UBCF), item-based collaborative filtering (IBCF) In our experiments, we use all 12 contextual dimensions, includ- and matrix factorization (MF) for each splitting approach. Ko- ing three emotional contexts: Mood, DominantEmo and EndEmo, ren [11] introduced three MF techniques: MF with rating bias (Bi- and non-emotional contexts, including Location, time, etc. The asMF), Asymmetric SVD (AsySVD) and SVD++; we found that goal is to compare emotional contexts with others. For example, BiasMF was the best choice for this data. in DCR, only influential contextual variables for a specific compo- We used the open source recommendation engine MyMediaLite nent are selected – whether emotional contexts are selected or not v3.07 [7] to evaluate UBCF, IBCF and BiasMF in our experiments. can indicate the significance of their impacts for this component. (We choose K=30 for KNN-based UBCF and IBCF.) In order to For a comprehensive comparison, we also performed experiments better evaluate MF techniques, we tried a range of different fac- without emotional contexts to better evaluate the importance of the tors (5 ≤ N ≤ 60, increment 5) and training iteration T (10 ≤ emotional dimensions. T ≤ 100, increment 10). Other parameters like learning and reg- In this data, there are the three contextual dimensions that con- ularization factors are handled by MyMediaLite, where stochastic tain emotional information. "EndEmo" is the emotional state expe- gradient descent is used as the optimization method. rienced at the end of the movie. "DominantEmo" is the emotion- For comparison purposes, we choose the well-known context- al state experienced the most during watching, i.e. what emotion aware matrix factorization algorithm (CAMF) [3], i.e. CAMF_C, was induced most times during watching. "Mood" is what mood CAMF_CI and CAMF_CU 2 as the baseline. We tried our best the user was in during that part of the day when the user watched to fine-tune the configurations (e.g. learning parameters, training the movie. Mood has lower maximum frequency than emotions, iterations, etc) for CAMF in order to make a reasonable and com- it changes slowly, so we assumed that it does not change during prehensive comparison. watching. "EndEmo" and "DominantEmo" contain the same sev- en conditions: Sad, Happy, Scared, Surprised, Angry, Disgusted, Neutral, where "Mood" only has simple three conditions: Positive, 6. EXPERIMENTAL RESULTS Neutral, Negative. In this section, the comparisons of predictive performance are in- troduced first, followed by the discussion of emotional roles dis- 5.2 Configurations covered in our experiments. To evaluate the performance of context-aware splitting approach- es, we used four splitting criteria described in [5]: tmean , tchi , 6.1 Prediction tprop and tIG . tmean estimates the statistical significance of the We use three treatments of the context information – one is the difference in the means of ratings associated to each alternative data with All Contexts where both emotional contexts and non- contextual condition using a t-test. tchi and tprop estimates the emotional variables are included, and the second one is the data statistical significance of the difference between two proportions – omitting the emotional context dimensions marked by No Emo- high ratings (>R) and low ratings (≤R) by chi square test and z- tions in the table below. The third one is the data with Emotions test respectively, where we choose R = 3 as in [5]. tIG measures Only emitting all non-emotional variables. The overall experimen- the information gain given by a split to the knowledge of the item i tal results are shown in Table 3, where the numbers in underlined rating classes which are the same two proportions as above. in italic are the best RMSEs by each approach (i.e. the best one in 1 2 http://212.235.187.145/spletnastran/raziskave/um/comoda/ CAMF_CU is a CAMF approach which utilizes the interaction comoda.php between contexts and users. [12] 24 Table 3: Overall Comparison of RMSE (The results for splitting approaches are based on BiasMF.) Algorithms All Contexts No Emotions Emotions Only CAMF_C 1.012 1.066 0.968 CAMF [3] CAMF_CI 1.032 1.083 1.019 CAMF_CU 0.932 1.021 0.902 DCR 1.043 1.057 1.046 DCM [24] DCW 1.017 1.037 1.036 Item Splitting 1.011 1.014 1.014 Splitting User Splitting 0.913 0.971 0.932 Approaches UI Splitting 0.892 0.942 0.903 each row), and the bold numbers are the best RMSEs by each data Table 4: Comparison of Predictive Performance (RMSE) Among forms (i.e. the best one in each column). Splitting Approaches The comparison of performances of those approaches over the three forms of data can be visualized by Figure 2. We see that in- LDOS-CoMoDa cluding emotions as contexts can improve RMSE compared with Algorithms tmean tchi tprop tIG the situation we only use non-emotional contexts (i.e. No Emotion- UBCF 1.040 1.021 1.028 1.043 Item s). The results differ when we switch our attention to the "Emo- IBCF 1.030 1.024 1.026 1.034 Splitting tions Only" one. In CAMF, it helps achieve the lowest RMSE with BiasMF 1.020 1.011 1.016 1.020 Emotions Only data. But, including all of the contextual informa- UBCF 1.026 0.987 0.999 1.052 User tion yielded the best RMSE for the other approaches: DCM and IBCF 0.985 0.967 0.985 1.039 Splitting context-aware splitting algorithms. BiasMF 0.934 0.913 0.928 1.011 UBCF 1.012 0.956 0.989 1.042 1.11 UI IBCF 0.972 0.946 0.972 1.020 1.08 Splitting 1.05 BiasMF 0.927 0.892 0.915 0.998 1.02 0.99 0.96 tion algorithm with tchi as the splitting criteria. 0.93 0.9 LDOS−CoMoDa 0.87 CAMF_C CAMF_CI CAMF_CU DCR DCW Item User UI Splitting Splitting Splitting 1.02 All Contexts No Emotions Emotions Only 1.00 Figure 2: Comparison of RMSE Over Three Contextual Situations 0.98 This result is not surprising because both DCM and splitting ap- proaches choose among contextual features, deciding which to ap- RMSE 0.96 ply in recommendation. CAMF, on the other hand, uses all avail- able context information in performing its factorization. Without 0.94 the benefit of feature selection, adding additional features to CAM- F may increase noise. 0.92 From an overall view, UI splitting has the best RMSE across all data treatments among those context-aware recommendation algo- 0.90 rithms. Table 4 shows more details of the predictive performance among the three splitting approaches if we include emotions as the Item Splitting User Splitting UI Splitting contexts (i.e. using all contexts). The numbers are shown as RMSE values, where the numbers in bold are the best performing RMSE values for each recommendation algorithm across all three splitting approaches. The numbers in underlined in italic are the best RMSE Figure 3: Boxplot of RMSE Among Splitting Approaches values achieved for the data set using each splitting approach. The numbers in underlined in bold are the global best RMSE for the Generally, MF is the best performing recommendation algorith- data set. m. This is not surprising because splitting increases sparsity and These tables show that adding emotions to contexts is able to MF approaches are designed to handle sparsity data. Because the provide improvement in terms of RMSE, thus answering research difference in RMSE is small, we show the boxplot of RMSEs a- question 1 in the affirmative. It also suggests an answer to question mong the best performing item splitting, user splitting and UI s- 2: that the UI splitting approach outperforms other two splitting ap- plitting approaches (i.e. the configuration as the underlined values proaches if it is configured optimally. In particular, the best RMSE in Table 4) in Figure 3. The data in the figure are the 120 RMSE values are achieved by UI splitting using MF as the recommenda- values which comes from the training iterations – different factors 25 30% 28% 36% Time 25% 33% DayType 30% 23% Season 27% 20% 24% Time 18% Location 21% Season 15% Weather 18% 13% EndEmo 15% 10% Social DominantEmo 12% 8% EndEmo 9% Mood 5% 6% DominantEmo 3% 3% 0% Interaction 0% Tmean Tchi Tprop Tmean Tchi Tprop Figure 4: Item Splitting Figure 5: User Splitting Table 5: Context Relaxation and Weighting by DCM Algorithm Components Context Relaxation By DCR Context Weighting By DCW Neighbor Selection N/A Day, Mood Neighbor Contribution Movie Year, Genre Movie Genre User Baseline DominantEmo, EndEmo, Movie Language DominantEmo, EndEmo, Interaction User Similarity EndEmo, Location DominantEmo (5 ≤ N ≤ 60, with 5 increment in each step) and training iter- EndEmo denotes the emotion of the users after seeing the movie, ation T (10 ≤ T ≤ 100, with 10 increment in each step). The and this result is consistent with previous work [12] on this da- figure confirms the comparative effectiveness of UI splitting – the ta. Obviously, emotion is a personal quality and can be considered box is significantly lower than the other two approaches with the as more dependent with users other than items – we conjecture that same training iterations. this may be the underlying clue explaining why user splitting works better than item splitting for the LDOS-CoMoDa data. And in user 6.2 The Role of Emotions splitting, the top two selected contextual dimensions are the two As mentioned before, one reason we sought to explore the usage emotional variables: EndEmo and DominantEmo – emotions are of emotional contexts to discover how emotions interact within generated and owned by users, therefore they are more dependent context-aware recommendation algorithms. Both splitting and D- with users other than items. This pattern is confirmed by the com- CM offer insights into how contextual dimensions and features are parison between two CAMF approaches: CAMF_CI and CAM- influential for recommendation. F_CU – contexts are more dependent with users than with items, which results in better performances by CAMF_CU. 6.2.1 Context-aware Splitting In short, the statistics based on splitting approaches reveal the In splitting approaches, the splitting statistics help discover how importance of emotions – at least the top first selected contextual contexts were applied in our experiments. In Figure 4 and 5, we dimension is the "EndEmo" for both item splitting and user split- show the top selected contextual dimensions for item splitting and ting. user splitting3 . The right legend indicates the contextual dimen- sions, and the y axis denotes the percentage of splits (item splits 6.2.2 Differential Context Modeling or user splits) using each dimension. For a clearer representation in In DCM, the context selection and context weighting can be ex- the figures, we only show contextual dimensions used on more than amined and show, in a detailed way, the contribution of each con- 5% (item splitting) or 6% (user splitting) of the recommendations. textual dimension in the final optimized algorithm. The results are We do not show results of tIG because this splitting criterion had shown in Table 5, where the four components in UBCF were de- the worst performance. scribed in the previous section. "N/A" in the table indicates no In general, the top two dimensions are consistent across those contextual constrains were placed. For clearer representation, we three impurity criteria: EndEmo and Time for item splitting and did not show specific weights of contexts in DCW; instead, we on- EndEmo and DominantEmo for user splitting. However, the per- ly list variables which were assigned weights above a threshold of centages as y axis in the figures are different, not to mention that 0.7. The weights are normalized to 1, and 0.7 therefore represents the selected condition in each dimension differs too, which result- a very influential dimension. s in different performance by using various impurity criteria. In We can see in the table that emotional variables are selected in terms of the specific selected emotional conditions, the results are DCR and weighted significantly in DCW and for which compo- not consistent – the top context dimension for item and user split- nents. Emotion is influential for a specific component but may not ting is the same – EndEmo, but the most frequent selected condi- for other ones. In DCR for example, EndEmo turns out to be influ- tion in this dimension is "Happy" for item splitting and "Neutral" ential when measuring user similarities and user baselines, but it is for user splitting. not that significant in computing the neighbor contribution. 3 The actual splitting is based on a specific contextual condition It is not surprising that the results from DCW are not fully consis- in a dimension, but the results of selected contextual conditions tent with ones from DCR. DCW is a finer-grained approach. Emo- are fuzzy, thus the distribution of selected contextual dimensions is tional variables are assigned to neighbor selection in DCW but not explored and reported here. for the same component in DCR, and the specific selected emotions 26 Table 6: Comparison of Context-aware Splitting and Differential Context Weighting Answers Context-aware Splitting Approaches Differential Context Weighting Q1. Emotional Effect Yes. RMSE is improved with emotional context Yes. RMSE is improved with emotional context di- dimensions. mensions. Q2. Algorithm Comparison UI splitting is the best and outperforms DCM and DCW is better than DCR and also works better than CAMF approaches. some CAMF approaches in specific contextual situ- ations. Q3. Usage of Contexts Emotional dimension is the top selected dimension Emotional dimensions get more weight than other for context splitting. contextual variables in most algorithm components. Q4. Roles EndEmo is used as the top first selected context for Emotions are significantly influential for some algo- both user and item splitting. DominantEmo is the rithm components, e.g. EndEmo and DominantEmo top second one in user splitting. for user similarities and baselines in DCW. are different too, e.g. it is DominantEmo selected in DCW for user question, by pointing out which components make good use of d- similarity calculation, but it is the EndEmo in DCR. In DCW, the ifferent contextual variables – showing here that DominantEmo is weights for EndEmo and DominantEmo are close to 1 (the weights important for calculating the user’s baseline. Note that the neigh- are all above 0.92) in the component of user baseline, which im- bor contribution component in DCW is alone in making significant plies significant emotional influence on this component. Emotional use of non-emotional context dimensions. This is interesting be- contexts are important in DCM, where it can be further confirmed cause in this component we are determining which movies to use by Table 3 – the RMSE values are increased if we remove emotions to compute the neighbor’s baseline for prediction. It appears that from the contexts. the system prefers to use non-emotional considerations in making The result by DCM is useful for further applications, such as af- use of others’ ratings, even though the user’s baseline is computed fective computing or marketing purposes. Take the results of DCW based on emotional dimensions. in Table 5 for example, Mood is influential for selecting neighbors, In addition, we can get extra information from our splitting ex- which implies that if two users rated the same item under the same periments because splitting is performed based on specific contex- mood situation, it is highly possible that they are the good neighbor tual conditions. As described above, we found in particular that candidates for each other (though neighbor selection also depends "Happy" vs not-"Happy" was the important split on the EndEmo on the user similarities). Similarly, DominantEmo is useful to mea- dimension for item splitting. Essentially, the algorithm is indicat- sure user similarities, which infers that users rated items similarly ing that there is an important difference between users who feel with the same dominant emotions are probably the good neighbors happy at the conclusion of a given movie and those that do not. in user-based collaborative filtering. In our future work, we plan to continue our exploration of these algorithms and more data using additional metrics, such as recall 7. CONCLUSION AND FUTURE WORK and/or normalized discounted cumulative gain. We are also look- ing at additional data sets to see if similar effects are found with In conclusion, both context-aware splitting approaches and DCM respect to emotions, as well as the empirical comparison among are able to reveal how emotions interact with algorithms to im- those context-aware recommendation algorithms. We also plan to prove recommendation performance by exploring the usage of con- examine the effects by the correlations between different contex- texts in the recommendation or splitting process. More specifically, tual variables, e.g. how emotional effects change if emotions are in context-aware splitting approaches, the percentage of emotional significantly dependent with other contexts or features. In addition, contexts used by item splits or user splits can tell the importance it is interesting to explore the association among emotions, user of emotions in distinguishing different user rating behavior. DCM profiles, item features and users’ ratings. For example, user may provides a way to see very specifically which emotional contexts feel "sad" after seeing a tragedy movie, but the emotion could be are influential for which components in the recommendation pro- "happy" because he or she saw such a good movie even if it is a cess. tragedy. Therefore, which specific emotions will result in a higher Table 6 examines how our research questions are answered for rating? the "sad" or the "happy"? each type of context-aware recommendation. As discussed above, we find that contextual dimensions keyed to emotions are very use- ful in recommendation (Question 1) and that UI splitting outper- 8. REFERENCES forms the DCM and CAMF approaches (Question 2). [1] G. Adomavicius, B. Mobasher, F. Ricci, and A. Tuzhilin. Question 3 asks about how emotional dimensions compare with Context-aware recommender systems. AI Magazine, other contextual information. Our results show that emotion-linked 32(3):67–80, 2011. context makes an important contribution to context-aware recom- [2] L. Baltrunas and X. Amatriain. Towards time-dependant mendation, although other dimensions are also certainly important. recommendation based on implicit feedback. In ACM For example, we see that EndEmo is the top selected contextual RecSys’ 09, Proceedings of the 4th International Workshop dimension in item and user splitting, with Time running the sec- on Context-Aware Recommender Systems (CARS 2009), ond in item splitting, and DominantEmo is the top second selected 2009. contextual dimension in user splitting. [3] L. Baltrunas, B. Ludwig, and F. Ricci. Matrix factorization Our fourth research question asks about how those two class- techniques for context aware recommendation. In es of context-aware recommendation algorithms can tell the roles Proceedings of the fifth ACM conference on Recommender of emotions by usage of contexts. Context-aware approaches are systems, pages 301–304. ACM, 2011. able to infer the roles by the distribution of usage of emotions se- [4] L. Baltrunas and F. Ricci. Context-based splitting of item lected for the splitting process. DCM techniques help answer this ratings in collaborative filtering. In Proceedings of the third 27 ACM conference on Recommender systems, pages 245–248. factors affecting music recommender success. In ISMIR, ACM, 2009. volume 2, pages 204–208, 2002. [5] L. Baltrunas and F. Ricci. Experimental evaluation of [20] Y. Zheng, R. Burke, and B. Mobasher. Assist your context-dependent collaborative filtering using item splitting. decision-making in various situations: Differential context User Modeling and User-Adapted Interaction, pages 1–28, relaxation for context-aware recommendations. In Research 2013. Colloquium, School of Computing, DePaul University, [6] S. Berkovsky, T. Kuflik, and F. Ricci. Cross-domain Chicago IL, USA, 2012. mediation in collaborative filtering. In User Modeling 2007, [21] Y. Zheng, R. Burke, and B. Mobasher. Differential context pages 355–359. Springer, 2007. relaxation for context-aware travel recommendation. In 13th [7] Z. Gantner, S. Rendle, C. Freudenthaler, and International Conference on Electronic Commerce and Web L. Schmidt-Thieme. Mymedialite: A free recommender Technologies (EC-WEB 2012), pages 88–99, 2012. system library. In Proceedings of the fifth ACM conference [22] Y. Zheng, R. Burke, and B. Mobasher. Optimal feature on Recommender systems, pages 305–308. ACM, 2011. selection for context-aware recommendation using [8] G. Gonzalez, J. L. de la Rosa, M. Montaner, and S. Delfin. differential relaxation. In ACM RecSys’ 12, Proceedings of Embedding emotional context in recommender systems. In the 4th International Workshop on Context-Aware Data Engineering Workshop, 2007 IEEE 23rd International Recommender Systems (CARS 2012). ACM, 2012. Conference on, pages 845–852. IEEE, 2007. [23] Y. Zheng, R. Burke, and B. Mobasher. Differential context [9] B.-j. Han, S. Rho, S. Jun, and E. Hwang. Music emotion modeling in collaborative filtering. In School of Computing classification and context-based music recommendation. Research Symposium. DePaul University, USA, 2013. Multimedia Tools and Applications, 47(3):433–460, 2010. [24] Y. Zheng, R. Burke, and B. Mobasher. Recommendation [10] N. Hariri, B. Mobasher, R. Burke, and Y. Zheng. with differential context weighting. In The 21st Conference Context-aware recommendation based on review mining. In on User Modeling, Adaptation and Personalization (UMAP IJCAI’ 11, Proceedings of the 9th Workshop on Intelligent 2013), pages 152–164, 2013. Techniques for Web Personalization and Recommender Systems (ITWP 2011), pages 30–36, 2011. [11] Y. Koren. Factorization meets the neighborhood: a multifaceted collaborative filtering model. In Proceedings of the 14th ACM SIGKDD international conference on Knowledge discovery and data mining, pages 426–434. ACM, 2008. [12] A. Odic, M. Tkalcic, J. F. Tasic, and A. Košir. Relevant context in a movie recommender system: Users’ opinion vs. statistical detection. In ACM RecSys’ 12, Proceedings of the 4th International Workshop on Context-Aware Recommender Systems (CARS 2012), 2012. [13] A. Odić, M. Tkalčič, J. F. Tasič, and A. Košir. Predicting and detecting the relevant contextual information in a movie-recommender system. Interacting with Computers, 25(1):74–90, 2013. [14] P. Resnick, N. Iacovou, M. Suchak, P. Bergstrom, and J. Riedl. Grouplens: an open architecture for collaborative filtering of netnews. In Proceedings of the 1994 ACM conference on Computer supported cooperative work, pages 175–186. ACM, 1994. [15] A. Said, E. W. De Luca, and S. Albayrak. Inferring contextual user profiles – improving recommender performance. In ACM RecSys’ 11, Proceedings of the 4th International Workshop on Context-Aware Recommender Systems (CARS 2011), 2011. [16] Y. Shi, M. Larson, and A. Hanjalic. Mining mood-specific movie similarity with matrix factorization for context-aware recommendation. In Proceedings of the Workshop on Context-Aware Movie Recommendation, pages 34–40. ACM, 2010. [17] M. Tkalčič, U. Burnik, A. Odić, A. Košir, and J. Tasič. Emotion-aware recommender systems–a framework and a case study. In ICT Innovations 2012, pages 141–150. Springer, 2013. [18] M. Tkalcic, A. Kosir, and J. Tasic. Affective recommender systems: the role of emotions in recommender systems. In ACM RecSys Workshop on Human Decision Making, 2011. [19] A. L. Uitdenbogerd and R. G. van Schyndel. A review of 28