=Paper=
{{Paper
|id=Vol-3268/Ran
|storemode=property
|title=PM K-LightGCN: Optimizing for Accuracy and Popularity Match in Course Recommendation
|pdfUrl=https://ceur-ws.org/Vol-3268/paper3.pdf
|volume=Vol-3268
|authors=Yiding Ran,Hengchang Hu,Min-Yen Kan
|dblpUrl=https://dblp.org/rec/conf/recsys/RanHK22
}}
==PM K-LightGCN: Optimizing for Accuracy and Popularity Match in Course Recommendation==
PM K-LightGCN: Optimizing for Accuracy and Popularity Match in Course Recommendation YIDING RAN, HENGCHANG HU, and MIN-YEN KAN, National University of Singapore A growing body of literature on educational recommenders focuses on accuracy but neglects how it can marginalize user experience. Accuracy optimization fits the interaction data, whereas user experience optimization recognizes students’ limited knowledge and recommends better alternatives. We propose a multi-objective course recommender that balances the optimization of both objectives: 1) accuracy, and 2) student experience. For the first objective, we take inspiration from K-Nearest Neighbors (KNN) model’s success in course recommendation, even outperforming neural network based models. KNN’s focus on pairwise relation between close neighbors aligns with the nature of course consumption. Hence, we propose K-LightGCN which uses KNN models to supervise embedding learning in state-of-the-art LightGCN and achieves a 12.8% accuracy improvement relative to LightGCN. For the second objective, we introduce metric PP-Mismatch@K to quantify user experience. We propose PM K-LightGCN which post-filters K-LightGCN’s outputs to optimize PP-Mismatch@K and achieves a 17% improvement in student experience with minimal drop in accuracy. Additional Key Words and Phrases: Multi-Objective Recommender, Course Recommender, Course Popularity 1 INTRODUCTION Course selection in tertiary education are high-stakes decisions that influences both the student’s subsequent academic plan and the future career path [9]. The development of artificial intelligence (AI) inspires exploration of AI-enabled course recommender to assist students’ decision making. Existing studies largely focuses on accuracy optimization [10, 18], where the sole pursuit is fitting the underlying student selections; i.e., course enrollment records. In contrast, student-centric course recommenders should also optimize user experience, which is not guaranteed by high accuracy. Pardos and Jiang proposed a course recommender designed for serendipity that recommends previously unknown but relevant courses [26]. Their experiments show that when presented with such new courses, students opt to change to take these courses instead. Hence, accuracy is not the best metric for user experience, as when given more information, students improve their experience by adjusting selections. Thus, accuracy optimization trained using student selections — the ground truth — is insufficient to optimize such high-stakes student experiences. Students share similar course selection criteria like course content and future career value [7]. But how they assess one course in these aspects differs. Formulating student experience optimization as an objective requires measuring the factors leading to student satisfaction. In a small-scale user study conducted at our institution, students demonstrated disparate interests towards elective courses: when shown detailed course information, some students consistently preferred popular courses whereas others favored niche ones. Hence, student satisfaction is related to whether course popularity matches their interests. Thus, we propose a proxy for student experience, preference-popularity match, that describes the ideal situation where students are recommended with courses of their desired popularity. Both accuracy optimization and preference-popularity match are important objectives of course recommendation. However, one often comes at the cost of another. The former closely fits the data whereas the latter believes students are better off with alternative course plans. To optimize for both and achieve balance, we present a multi-objective course recommender with accuracy optimization as its primary objective and preference-popularity match as a secondary one. ∗ Copyright 2022 for this paper by its authors. Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0). Presented at the MORS workshop held in conjunction with the 16th ACM Conference on Recommender Systems (RecSys), 2022, in Seattle, USA. 1 MORS 2022, September 2022, Seattle, USA Ran, Hu and Kan We first conduct a preliminary study that explores the performance of traditional recommenders and modern neural network based models on course recommendation. We then observe the unusual superiority of ItemKNN in accuracy optimization, which is explained by the tight alignment between its focus on pairwise relations among close neighbors and the unique characteristics of course consumption. Inspired by this, we propose a revised version of the state-of-the-art recommendation algorithm, LightGCN. In our revision, pairwise similarity computed by KNN models are used to supervise embedding learning in LightGCN. Then, predictions by KNN models and LightGCN are combined as the final recommendation. The resultant K-LightGCN model optimizes for our first objective — accuracy. We then introduce a metric for student experience, PP-Mismatch@K, which quantifies the level of preference-popularity mismatch in recommendations. To achieve our second objective — preference-popularity match — we propose PM K-LightGCN, which takes in recommendations of K-LightGCN and selects 10 courses that best fit the defined objective. Through evaluation, our multi-objective course recommender demonstrates a good balance between both objectives. 2 RELATED WORK 2.1 Educational Recommendation Systems. The education domain is relatively underexplored in recommender research. Only 44 works were associated with this field in a 2018 mapping study [29]. The unique characteristics of course-taking behavior add complications to course recommendation. First, in institutes of higher learning (hereafter, IHLs), students only take 4 to 6 courses per semester, worsening sparsity issues already challenging recommender implementation [14]. Also, students’ course selection is influenced by both intrinsic and extrinsic motivations which differ across individuals [17, 23]. Additionally, IHL courses are designed to be complementary and often have complex relations such as prerequisites and preclusions [11]. An ideal recommender should learn students’ varied interests and consider the underlying constraints in course choice [14]. Approaches to course recommendation include content-based, collaborative filtering and knowledge-based techniques [14]. Content-based approaches learn courses’ characteristics from their description and recommend those similar to students’ previous enrollments [6, 25, 30]. Collaborative filtering is the most common approach that uses only the enrollment history, represented as an interaction matrix, from which student and course vector representations are learned [8, 13, 22, 24]. Knowledge-based methods use domain knowledge like students’ faculty and grades and course prerequisite relationships. Jiang et al. proposed a Goal-Based Course Recommender that uses past semester grades [19]. 2.2 Use of Popularity in Recommendation Systems. In this paper, we introduce the novel concept of preference-popularity match in course recommendation, which aligns popularity of recommended courses with student preference. One scenario that violates this ideal is recommending popular courses to students preferring niche ones. This popularity bias arises when popular items are recommended with higher accuracy and frequency, but less popular items are recommended with lower accuracy [4]. Abdollahpouri et al. show that the 3% items involved in the top 20% of interactions, i.e., the popular items, make up 60–100% of recommended items sourced from different recommendation algorithms [4]. Recommending popular items caters to the needs of the majority, leading to high accuracy. However, users enjoy popular items to different extents, and those preferring niche ones would suffer from inaccurate recommendations [4]. Existing popularity bias mitigation includes unbiased learning [5, 32] which adjusts the data distribution by re-weighting interaction samples or uses bias-free uniform data to learn unbiased embedding and ranking adjustment [1, 2, 33] which includes a regularization term to ensure exposure of less popular items or re-ranks recommendation lists based on certain conditions, similar to our approach. However, 2 PM K-LightGCN: Optimizing for Accuracy and Popularity Match in Course Recommendation MORS 2022, September 2022, Seattle, USA Course ID wing_modede529dfcbb2907e9760eea0875cdd12 Student ID wing_mod412b5c6d4a88a03e91dfc16dd4d494ff Faculty School of Computing Interaction Semester 1910 Enrollment Semester 1710 Table 1. Enrollment Record Sample popularity bias mitigation neglects the undesirable scenario where niche courses are recommended to students enjoying popular ones, which motivates our proposal of a more comprehensive concept — preference-popularity match. 3 EXPERIMENT SETTINGS We use course enrollment records provided by a leading tertiary educational institution to explore both objectives with accuracy optimization in Section 4 and preference-popularity match in Section 5. Due to privacy concerns, IHLs normally do not disclose their students’ enrollment records, making it difficult to test our algorithm on multiple datasets. 3.1 Dataset and Data Split Our dataset is an anonymized, institutional review board (IRB) approved listing of per-semester course-taking histories of graduated undergraduates from the year 2010 to 2020. There are 41,304 unique students and 5,179 unique courses, which together generate 1.4 million enrollments. Due to privacy considerations, each student and course is assigned a unique masked ID to anonymize student and course. For each enrollment record, we also have information on the student’s faculty, the semester when this enrollment occurred (interaction semester) and the semester when the student matriculated (enrollment semester). A sample of this dataset is shown in Table 1. The dataset is split into train, validation and test sets mimicking real-life scenario. Course recommendations are usually based on students’ past enrollment with reference to the records of graduated students. Hence, we randomly select 35% of students to form Group A (the graduated seniors), and the remaining 65% form Group B (students seeking recommendations). All of Group A’s records are allocated for training, whereas Group B’s are split into train, validation and test sets, based on the interaction semester. The resultant train–validation–test ratio is approximately 7:1:2. 3.2 Evaluation The recommendation algorithms are implemented using the RecBole library [31], which provides a unified framework to develop recommendation algorithms. We evaluate recommender performance on two objectives: accuracy by metrics HR@K and NDCG@K as well as preference-popularity match using our self-proposed metric Preference-Popularity Mismatch@K, explained in the upcoming sections. We set 𝐾 = 10 to evaluate the top 10 recommendations. 4 OBJECTIVE 1: ACCURACY OPTIMIZATION We first optimize the recommender for accuracy. We conduct a preliminary study to identify the unique nature of course consumption which motivates our K-LightGCN proposal. We first lay out notations we used: individual students 𝑠𝑖 and courses 𝑐 𝑗 are referred to by subscripts. The set of all students and courses are of size 𝑥 and 𝑦, respectively. The symmetric item–item (user–user) similarity matrix generated by ItemKNN (UserKNN) is represented by 𝑤 𝑖𝑡𝑒𝑚 (𝑤 𝑢𝑠𝑒𝑟 ). 4.1 Preliminary Study We shortlist the following non-domain specific models as baselines. Existing course recommendation algorithms are not selected as they do not share standardized input. Many require course description or grade information that are absent from this dataset. The following algorithms take in only course enrollment records. 3 MORS 2022, September 2022, Seattle, USA Ran, Hu and Kan ItemKNN is an item-based, KNN model [12] which computes recommendation scores based on pairwise item–item similarity. Courses the most similar to those taken by the student are recommended. UserKNN is a user-based model that recommends courses taken by students similar to the target student. Multi-Layer Perceptron Model (MLP) employs a basic neural network structure and represents students and courses as latent vectors for non-linear relation capture [16]. Neural Matrix Factorization (NeuMF) is a specific implementation of Neural Collaborative Filtering (NCF [16]) which fuses Matrix Factorization (MF) and MLP methods so that they reinforce each other. Light Graph Convolution Network (LightGCN) simplifies Graph Convolution Network (GCN) model to best serve the recommendation task by including only neighborhood aggregation [15]. In LightGCN, the embeddings for each student and course are learned by aggregating that of its neighbors. The prediction for the interaction between Student 𝑠𝑖 and Course 𝑐 𝑗 is the dot product of the embeddings for the student and course, 𝑦𝑠𝑖ˆ,𝑐 𝑗 . LightGCN uses Bayesian Personalized Ranking (BPR) loss [27], a pairwise loss that encourages the prediction of an observed interaction to be higher than unobserved ones and penalizes otherwise. The BPR loss is defined in Equation 1 where 𝑛𝑠𝑖 represents courses Student 𝑠𝑖 has taken previously and E (0) 2 is the 𝐿2 regularization term to prevent overfitting. 𝑥 ∑︁ ∑︁ ∑︁ 𝐿𝐵𝑃𝑅 = − ln 𝜎 ( 𝑦ˆ𝑠𝑖 ,𝑐 𝑗 − 𝑦ˆ𝑠𝑖 ,𝑐𝑝 ) + 𝜆 E (0) 2 (1) 𝑖=1 𝑗 ∈𝑛𝑠𝑖 𝑝∉𝑛𝑠𝑖 4.1.1 Results. Table 2a shows baseline performance. Both KNN models outperform the rest despite their much simpler structure, highly uncommon in other recommendation domains [16, 20, 21]. It is possible the nature of course consumption aligns with the underlying inductive bias of KNN models — identifying neighbors of students and courses. Among the deep models, only LightGCN incorporates neighborhood information, outperforming others. This suggests that localized relations captured in neighborhoods are more important than global relations in course recommendation. Model HR@10 NDCG@10 a) ItemKNN 0.7762 0.3337 UserKNN 0.7294 0.2521 # layers HR@10 NDCG@10 MLP 0.6013 0.1946 1 0.6588 0.2260 NeuMF 0.6458 0.2156 2 0.6950 0.2502 LightGCN 0.7008 0.2542 3 0.6938 0.2491 4 0.6897 0.2434 b) CN-LightGCN 0.7287 (+3.98%) 0.2896 (+13.9%) 6 0.6798 0.2365 c) K-LightGCN 0.7905 (+1.84%) 0.3346 (+0.27%) Table 3. LightGCN with Different Layer Table 2. Accuracy Optimization Results: a) Baseline Performance (§ 4.1.2), b) Numbers. (Best performance is bolded.) Revised LightGCN Performance (performance relative to LightGCN) (§ 4.1.3) and c) K-LightGCN Performance (performance relative to ItemKNN) (§ 4.2). 4.1.2 Superiority of ItemKNN. Accurately explaining such observed superiority of ItemKNN in course recommendation assists in understanding the course recommendation task and designing models accordingly. ItemKNN computes the pairwise item–item similarity for all available courses using cosine similarity. For each course, only the top 𝑘 highest similarity values are kept, which correspond to its 𝑘 nearest neighbors. We thus compare the structure of the best performing baseline ItemKNN and the best performing deep model LightGCN. By checking whether the identified structural differences contribute to ItemKNN’s success, we gain a deeper understanding on the ItemKNN’s strength as applied to course recommendation. Importance of Pairwise Relation. Neighborhood information is incorporated in LightGCN through neighbor con- volution. Each layer propagates information of neighbors from varied distances away: the first layer passes information only to those adjacent; i.e., for each course, only information of students who have taken the course is passed. With 4 PM K-LightGCN: Optimizing for Accuracy and Popularity Match in Course Recommendation MORS 2022, September 2022, Seattle, USA two layers, additional information transmitted includes that of students sharing interaction history overlap with the target student. Hence, at layer two, LightGCN learns pairwise relations, equivalent to ItemKNN’s pairwise similarity. ItemKNN’s high accuracy suggests the possibility that pairwise relation captures the essential patterns for accurate recommendation. Hence, we hypothesize that in course recommendation, adding more layers over-smooths the learned representations. To validate this, we experiment with LightGCN of different layer numbers. Table 3’s results confirm that 2-layer LightGCN focusing on pairwise relation is most effective in learning the underlying relations and generating accurate recommendations, which differs from the e-commerce and social network setting studied previously [15]. This relates to the unique nature of course consumption. Students who took the same courses are likely from the same academic program; hence, they share common curriculum requirements and similar interaction patterns [17]. For a target student, what others from same program have taken previously largely indicates the courses (s)he will take next. By learning pairwise information, we are capturing these hidden relations among students and courses. Focus on Close Neighbors. We also observe that ItemKNN only considers the top 𝑘 neighbors whereas in LightGCN, information from all neighbors contributes to embedding learning and the final recommendation. It is possible that considering more than necessary neighbors brings in noise and lowers accuracy. To test this, we restrain LightGCN at the second layer to only perform neighbor propagation using the closest 𝑘 neighbors where 𝑘 is the number of neighbors used in KNN models. This revised LightGCN is called Constrain-Neighbor LightGCN (CN-LightGCN). Table 2b’s results demonstrate the improvement in CN-LightGCN’s performance relative to LightGCN, validating our claim. For a target student, (s)he may have taken the same course with many others. Among them, only a few are from the same academic program, whose course-taking histories are of key importance in suggesting what the target student would take next. Hence, focusing on these strong neighbors and discarding noise from weak ones is essential to course recommendation. 4.2 K-LightGCN: Accuracy Optimization For Course Recommendation The close alignment between ItemKNN’s focus on pairwise relations between close neighbors and the nature of course consumption partially explains the superiority of ItemKNN. Despite its overall higher accuracy, LightGCN provides accurate recommendations for some cases that ItemKNN fails. Hence, we propose K-LightGCN which combines ItemKNN and LightGCN and modifies it to reinforce the two features identified as key to accurate course recommendation. Fig. 1. An illustration of K-LightGCN model architecture. Pairwise similarity matrices computed by ItemKNN are used to supervise the embedding learning in LightGCN. Predictions generated by LightGCN and ItemKNN are combined as final recommendation. 5 MORS 2022, September 2022, Seattle, USA Ran, Hu and Kan 4.2.1 K-LightGCN Structure. K-LightGCN consists of paired LightGCN and ItemKNN components (Fig. 1). K-LightGCN enhances the strength of KNN by having the LightGCN component learn student and course embeddings with similarity values close to that computed in KNN models. Thus, we introduce a new term to the original BPR loss. At each training 𝑢𝑠𝑒𝑟 batch, user–user cosine similarity matrix 𝑤 𝐿𝑖𝑔ℎ𝑡𝐺𝐶𝑁 𝑖𝑡𝑒𝑚 and item–item cosine similarity matrix 𝑤 𝐿𝑖𝑔ℎ𝑡𝐺𝐶𝑁 are computed respectively using the learned student and course embedding in LightGCN. The squared differences between the similarity matrices computed by LightGCN and KNN are added to the loss. The resultant loss function in K-LightGCN is defined in Equation 2 where 𝐿𝐵𝑃𝑅 is the original BPR loss defined in Equation 1. The final prediction is the weighted sum of LightGCN and ItemKNN predictions, with different weights learned for each student. 𝑦 𝑦 𝑥 𝑥 ′ ∑︁ 1 ∑︁ ∑︁ 1 ∑︁ 𝐿𝐵𝑃𝑅 = 𝐿𝐵𝑃𝑅 + 𝜇 ( 𝑖𝑡𝑒𝑚 (𝑤𝐿𝑖𝑔ℎ𝑡𝐺𝐶𝑁𝑐 ,𝑐 − 𝑤𝑐𝑖𝑡𝑒𝑚 𝑖 ,𝑐 𝑗 )2 + 𝑢𝑠𝑒𝑟 (𝑤𝐿𝑖𝑔ℎ𝑡𝐺𝐶𝑁𝑠 ,𝑠 − 𝑤𝑠𝑢𝑠𝑒𝑟 𝑖 ,𝑠 𝑗 )2) (2) 𝑖=1 𝑦 𝑗 =1 𝑖 𝑗 𝑖=1 𝑥 𝑗 =1 𝑖 𝑗 4.2.2 K-LightGCN Performance. The experiment results of K-LightGCN is shown in Table 2c. As K-LightGCN outper- forms both ItemKNN and LightGCN, the effectiveness of the underlying design for course recommendation is validated. By learning embeddings with pairwise similarity close to that computed by KNN models, LightGCN is sharper in capturing the essential pairwise relations. For example, for the test case between Student s37 and Course 4f6, ItemKNN recommends accurately but LightGCN misses the hit. Previously, Student s37 has interacted with eight courses and identified by ItemKNN, the target Course 4f6 is highly similar to three of them, including Courses 0d, f5, and z9. Hence, ItemKNN successfully recommends the target course. However, LightGCN misses this relation as the cosine similarity of the LightGCN embedding between the four courses is close to zero. With ItemKNN supervising the embedding learning, K-LightGCN learns similar embeddings for the four courses identified, leading more accurate recommendation. 5 OBJECTIVE 2: PREFERENCE-POPULARITY MATCH Accuracy solely learns student selections. However, with more information, students can adjust their plans to improve experience. There are multiple ways to quantify student experience. We previously identify that students have varied tendency to take popular or niche courses in a user study. Hence, we measure student preference and course popularity to further optimize student experience by matching the two values, inspiring our Popularity-Match K-LightGCN model. 5.1 Preliminary Study We propose a continuous measure for course popularity and student preference. Course popularity is computed as 𝑝𝑜𝑝 (𝑐 𝑗 ) = 𝑙𝑜𝑔( #𝑖𝑛𝑡𝑒𝑟𝑎𝑐𝑡𝑖𝑜𝑛𝑠 #𝑠𝑒𝑚𝑒𝑠𝑡𝑒𝑟𝑠 ) which is the average number of enrollments during semesters when the module is offered. Higher pop indicates on average, more students take the course, implying it is more highly desired in terms of course selection criteria like course content, instructor quality, and course difficulty [7]. Student preference is computed as Í 𝑝𝑟𝑒 𝑓 (𝑠𝑖 ) = 1 𝑐 𝑗 ∈𝑛𝑠𝑖 𝑝𝑜𝑝 (𝑐 𝑗 ) |𝑛𝑠 | . We make a first step by assuming a uniform distribution of student preference, hence, 𝑖 courses from past interaction are given equal weights regardless of recency. Higher pref indicates the student tends to take more popular courses. Fig. 2 shows the distribution for course popularity and student preference in our dataset. For courses, a limited few have pop above 4 but as they dominate the enrollment records, they are recommended more, even when the target student prefers niche courses. As required by IHLs, a typical student takes several extremely popular core courses and some relatively niche electives, resulting in most with pref above 4. 5.1.1 Preference-Popularity Match Definition and Evaluation. The distribution of student preference demonstrates large variation: some prefer popular courses whereas others enjoy niche ones. Hence, to provide quality user experience for all, 6 PM K-LightGCN: Optimizing for Accuracy and Popularity Match in Course Recommendation MORS 2022, September 2022, Seattle, USA Fig. 2. Violin plots of course popularity (l) and student preference (r). Wider sections indicate most courses (students) have pop (pref ) within that range and those beyond the 75th percentile are in orange. our goal is to expose students with low pref to niche courses as well as recommending popular courses that better fit the interests of students with high pref [3]; i.e., preference-popularity matching. In contrast, preference-popularity mismatch is a loss: a mismatch between student preference and popularity of recommended courses. This occurs when popular (niche) courses are recommended to students with low (high) pref. With the former, students are not exposed to the niche courses they are looking for; with the latter, students are likely recommended with courses they already know of or those that only loosely fit their preference. Such recommendations do not value add to students’ decision making. We propose Preference-Popularity Mismatch@K (PP-Mismatch@K) loss to measure preference-popularity match performance, defined as 𝑧1 1 Í Í𝑧 𝑎=1 |𝐾 | 𝑐 𝑗 ∈𝑟 𝑎 𝑝𝑟𝑒 𝑓 (𝑠𝑖 ) − 𝑝𝑜𝑝 (𝑐 𝑗 ) . For test case 𝑎, it takes in corresponding top 𝐾 = 10 recommendation list 𝑟𝑎 , computing the absolute difference between student’s preference and recommendations’ average popularity. PP-Mismatch@10 is the average mismatch of all 𝑧 test cases and higher values indicate poorer performance. 5.2 PM K-LightGCN: Multi-Objective Course Recommender Matching popularity of recommended courses with student preference is key to meeting students’ varied needs and providing quality user experience. We propose Popularity-Match K-LightGCN (PM K-LightGCN) to explicitly capture this criterion by aligning model recommendations with students’ preference on top of accuracy optimization. PM K-LightGCN takes the top 50 recommendations by K-LightGCN and selects 10 courses that minimize the mismatch defined as |𝐾1 | 𝑐 𝑗 ∈𝑟𝑎 𝑝𝑟𝑒 𝑓 (𝑠𝑖 ) − 𝑝𝑜𝑝 (𝑐 𝑗 ) by choosing those with popularity closest to the target student’s preference. Í We do not perform another round of optimization using all available courses to achieve the second objective as we believe accuracy is the primary criterion to be fulfilled. After K-LightGCN generates accurate recommendations, we select the 10 courses that fit the second objective while still ensuring the accuracy of the final recommendations. 5.2.1 Experiment Results. Recommenders’ performance on both objectives is shown in Table 4, which indicates that recommendations by PM K-LightGCN better match the students’ preference (a mismatch reduction of 17%). The only difference between PM K-LightGCN and K-LightGCN is the filtering component that selects the 𝐾 = 10 courses out of the top recommendations by K-LightGCN. PM K-LightGCN successfully mitigates preference-popularity mismatch, but K-LightGCN generates recommendations with the highest mismatch. To optimize solely for accuracy, K-LightGCN picks up the pattern that most students enjoy popular courses as indicated by the wider sections between 𝑝𝑟𝑒 𝑓 = 4 − 5 in the violin plot for student preference (Fig. 2) and recommends popular courses more often to cater to the majority. Model PP-Mismatch@10 HR@10 NDCG@10 ItemKNN 1.050 0.7762 0.3337 UserKNN 1.071 0.7294 0.2521 LightGCN 1.077 0.7008 0.2542 K-LightGCN 1.109 0.7905 0.3346 PM K-LightGCN 0.920 (-17.0%)† 0.7570 (-4.23%)* 0.3000 (-10.3%)* Table 4. Performance in Two Objectives (relative to K-LightGCN). † : For PP-Mismatch, lower values are desirable and negative percentage change indicates better performance. * : For HR and NDCG, 7 negative percentage change indicates worse performance. MORS 2022, September 2022, Seattle, USA Ran, Hu and Kan 5.2.2 Case Studies. We select three test cases, listed in Table 5 to examine PP-Mismatch mitigation in PM K-LightGCN. For simplification, we list the first five courses in the top 10 recommendations. In the first case study, both PM K- LightGCN and K-LightGCN recommend accurately. In addition to accuracy optimization, with PM K-LightGCN filtering the outputs of K-LightGCN to minimize PP-Mismatch, popularity of its recommendations all align with the student’s preference. Hence, the student is likely to get inspiration from PM K-LightGCN’s recommendations as they have the relevant characteristics considered by the student. In the second case study, the connection between target course and the student’s preference is missed by K-LightGCN, which is corrected with the additional filtering process in PM K-LightGCN. However, both models fail at accuracy optimization in the third case study. Despite it, PM K-LightGCN recommends several niche courses that are of popularity similar to the student’s preference which offers the student more possibilities for making a more informed decision. Target Case Student Course Popularity of Top 5 Courses Recommended Study Preference Popularity PM K-LightGCN K-LightGCN #1 #2 #3 #4 #5 #1 #2 #3 #4 #5 1 4.381 3.883 3.946 3.883 3.876 3.588 3.419 3.946 3.377 3.883 3.411 2.508 2 3.343 3.434 3.332 3.377 3.434 3.583 3.588 3.876 1.872 3.030 4.200 2.085 3 1.749 1.398 1.621 1.477 1.407 1.376 1.301 3.030 3.168 1.477 2.347 2.684 Table 5. Case Study of PM K-LightGCN (Accurate recommendations are bolded, #k represents the 𝑘𝑡ℎ ranked course.) 6 CONCLUSION We propose a multi-objective course recommender that optimizes for accuracy and preference-popularity match. Typical IHL curriculum requires students to read core courses in junior years and electives in senior years. Hence, as students mature, they will benefit more from PM K-LightGCN which matches their personalized preference. In our study, the PP-Mismatch@10 improvement in PM K-LightGCN mainly comes from recommendation in later years. Course recommendation remains a developing area. This project sheds lights on the importance of neighborhood information and multi-objective recommender development. Future studies could improve on PM K-LightGCN with alternative student preference measure. Considering students’ limited knowledge on niche courses, current preference measure might be an overestimation. Besides, it is computed as the average popularity of courses taken previously which neglects the popularity variations in the enrollment history. Such limitation will result in poor performance when the student takes extremely popular and niche courses together. It could be mediated with next-basket recommendation [28] which matches the popularity distribution of the recommended basket with that of courses taken previously. Though PM K-LightGCN’s post-filtering step does not involve intricate machine learning, it is sufficiently lightweight and saves space and time when deployed. Besides, this lightweight design is both effective in PP-Mismatch mitigation and balancing both objectives. Although the pursuit of preference-popularity match lowers accuracy, compared to K-LightGCN, PM K-LightGCN achieves a 17% improvement in preference-popularity match with the sacrifice of only 4% in accuracy. As enrollment records may not optimize student experience due to their limited knowledge [26], the fall in accuracy can improve user satisfaction. However, due to the inherent constraints of our dataset not including actual course information, it is infeasible to conduct a user study for better model evaluation from user’s perspective. Besides, the design of PM K-LightGCN allows scaling to additional criteria. When other needs like the promotion of certain courses to better prepare students for the workplace arise, they can be formulated as additional loss functions or post-filtering criteria to be added to the recommender in the same lightweight and efficient manner. 8 PM K-LightGCN: Optimizing for Accuracy and Popularity Match in Course Recommendation MORS 2022, September 2022, Seattle, USA REFERENCES [1] Himan Abdollahpouri, Robin Burke, and Bamshad Mobasher. 2017. Controlling Popularity Bias in Learning-to-Rank Recommendation. In Proceedings of the Eleventh ACM Conference on Recommender Systems (Como, Italy) (RecSys ’17). Association for Computing Machinery, New York, NY, USA, 42–46. https://doi.org/10.1145/3109859.3109912 [2] Himan Abdollahpouri, Robin Burke, and Bamshad Mobasher. 2019. Managing Popularity Bias in Recommender Systems with Personalized Re-Ranking. https://www.aaai.org/ocs/index.php/FLAIRS/FLAIRS19/paper/view/18199/17319 [3] Himan Abdollahpouri, Masoud Mansoury, Robin Burke, and Bamshad Mobasher. 2020. Addressing the multistakeholder impact of popularity bias in recommendation through calibration. arXiv preprint arXiv:2007.12230 (2020). [4] Himan Abdollahpouri, Masoud Mansoury, Robin Burke, Bamshad Mobasher, and Edward Malthouse. 2021. User-centered Evaluation of Popularity Bias in Recommender Systems. In Proceedings of the 29th ACM Conference on User Modeling, Adaptation and Personalization. 119–129. [5] Qingyao Ai, Keping Bi, Cheng Luo, Jiafeng Guo, and W Bruce Croft. 2018. Unbiased learning to rank with unbiased propensity estimation. In The 41st International ACM SIGIR Conference on Research & Development in Information Retrieval. 385–394. [6] Rel Guzman Apaza, Elizabeth Veronica Vera Cervantes, Laura Vanessa Cruz Quispe, and José Eduardo Ochoa Luna. 2014. Online Courses Recommendation based on LDA. In SIMBig. [7] Elisha Babad. 2001. Students’ Course Selection: Differential Considerations for First and Last Course. Research in Higher Education 42, 4 (2001), 469–492. http://www.jstor.org/stable/40196437 [8] Alican Bozyiğit, Fatma Bozyiğit, Deniz Kilinç, and Efendi Nasiboğlu. 2018. Collaborative Filtering based Course Recommender using OWA operators. In 2018 International Symposium on Computers in Education (SIIE). 1–5. https://doi.org/10.1109/SIIE.2018.8586681 [9] Hui-Tzu Chang, Chia-Yu Lin, Li-Chun Wang, and Fang-Ching Tseng. 2022. How Students can Effectively Choose the Right Courses: Building a Recommendation System to Assist Students in Choosing Courses Adaptively. Educational Technology Society 25, 1 (2022), 61–74. https: //www.jstor.org/stable/48647030 [10] Pei-Chann Chang, Cheng-Hui Lin, and Meng-Hui Chen. 2016. A hybrid course recommendation system by integrating collaborative filtering and artificial immune systems. Algorithms 9, 3 (2016), 47. [11] Glenn M Davis, Abdallah A AbuHashem, David Lang, and Mitchell L Stevens. 2020. Identifying Preparatory Courses that Predict Student Success in Quantitative Subjects. In Proceedings of the Seventh ACM Conference on Learning@ Scale. 337–340. [12] Mukund Deshpande and George Karypis. 2004. Item-Based Top-N Recommendation Algorithms. ACM Trans. Inf. Syst. 22, 1 (jan 2004), 143–177. https://doi.org/10.1145/963770.963776 [13] Asmaa Elbadrawy and George Karypis. 2016. Domain-Aware Grade Prediction and Top-n Course Recommendation. In Proceedings of the 10th ACM Conference on Recommender Systems (Boston, Massachusetts, USA) (RecSys ’16). Association for Computing Machinery, New York, NY, USA, 183–190. https://doi.org/10.1145/2959100.2959133 [14] Deepani B. Guruge, Rajan Kadel, and Sharly J. Halder. 2021. The State of the Art in Methodologies of Course Recommender Systems‚ÄîA Review of Recent Research. Data 6, 2 (2021). https://doi.org/10.3390/data6020018 [15] Xiangnan He, Kuan Deng, Xiang Wang, Yan Li, YongDong Zhang, and Meng Wang. 2020. LightGCN: Simplifying and Powering Graph Convolution Network for Recommendation. 639–648. https://doi.org/10.1145/3397271.3401063 [16] Xiangnan He, Lizi Liao, Hanwang Zhang, Liqiang Nie, Xia Hu, and Tat-Seng Chua. 2017. Neural collaborative filtering. In Proceedings of the 26th international conference on world wide web. 173–182. [17] Mary R. Hedges, Gail A. Pacheco, and Don J. Webber. 2014. What determines students‚Äô choices of elective modules? International Review of Economics Education 17 (2014), 39–54. https://doi.org/10.1016/j.iree.2014.05.003 [18] Chung-Yi Huang, Rung-Ching Chen, and Long-Sheng Chen. 2013. Course-recommendation system based on ontology. In 2013 International Conference on Machine Learning and Cybernetics, Vol. 3. IEEE, 1168–1173. [19] Weijie Jiang, Zachary A Pardos, and Qiang Wei. 2019. Goal-based course recommendation. In Proceedings of the 9th International Conference on Learning Analytics & Knowledge. 36–45. [20] Santosh Kabbur, Xia Ning, and George Karypis. 2013. Fism: factored item similarity models for top-n recommender systems. In Proceedings of the 19th ACM SIGKDD international conference on Knowledge discovery and data mining. 659–667. [21] Giannis Karamanolakis, Kevin Raji Cherian, Ananth Ravi Narayan, Jie Yuan, Da Tang, and Tony Jebara. 2018. Item recommendation with variational autoencoders and heterogeneous priors. In Proceedings of the 3rd Workshop on Deep Learning for Recommender Systems. 10–14. [22] Yehuda Koren, Steffen Rendle, and Robert Bell. 2022. Advances in Collaborative Filtering. Springer US, New York, NY, 91–142. https://doi.org/10. 1007/978-1-0716-2197-4_3 [23] Y.-M Lee and Sung-Hsien Sun. 2010. The relationship between autonomous motivation of course-taking and learning engagement on college students. 55 (03 2010), 155–182. [24] Bhaskar Mondal, Om Patra, Sanket Mishra, and Priyadarshan Patra. 2020. A course recommendation system based on grades. In 2020 international conference on computer science, engineering and applications (ICCSEA). IEEE, 1–5. [25] Raphaël Morsomme and Sofia Vazquez Alferez. 2019. Content-Based Course Recommender System for Liberal Arts Education. International educational data mining society (2019). 9 MORS 2022, September 2022, Seattle, USA Ran, Hu and Kan [26] Zachary A. Pardos and Weijie Jiang. 2020. Designing for Serendipity in a University Course Recommendation System. In Proceedings of the Tenth International Conference on Learning Analytics amp; Knowledge (Frankfurt, Germany) (LAK ’20). Association for Computing Machinery, New York, NY, USA, 350–359. https://doi.org/10.1145/3375462.3375524 [27] Steffen Rendle, Christoph Freudenthaler, Zeno Gantner, and Lars Schmidt-Thieme. 2009. BPR: Bayesian Personalized Ranking from Implicit Feedback. In Proceedings of the Twenty-Fifth Conference on Uncertainty in Artificial Intelligence (Montreal, Quebec, Canada) (UAI ’09). AUAI Press, Arlington, Virginia, USA, 452–461. [28] Steffen Rendle, Christoph Freudenthaler, and Lars Schmidt-Thieme. 2010. Factorizing Personalized Markov Chains for Next-Basket Recommendation. In Proceedings of the 19th International Conference on World Wide Web (Raleigh, North Carolina, USA) (WWW ’10). Association for Computing Machinery, New York, NY, USA, 811–820. https://doi.org/10.1145/1772690.1772773 [29] Abdon Rivera, Mariela Tapia-Leon, and Sergio Luján-Mora. 2018. Recommendation Systems in Education: A Systematic Mapping Study. 937–947. https://doi.org/10.1007/978-3-319-73450-7_89 [30] S. Venugopalan, M. V. Srinath, and Paul Rodrigues. 2016. Recommender System for E-Learning through Content and Profile Based Approach. In Proceedings of the Second International Conference on Information and Communication Technology for Competitive Strategies (Udaipur, India) (ICTCS ’16). Association for Computing Machinery, New York, NY, USA, Article 45, 5 pages. https://doi.org/10.1145/2905055.2905103 [31] Wayne Xin Zhao, Shanlei Mu, Yupeng Hou, Zihan Lin, Kaiyuan Li, Yushuo Chen, Yujie Lu, Hui Wang, Changxin Tian, Xingyu Pan, Yingqian Min, Zhichao Feng, Xinyan Fan, Xu Chen, Pengfei Wang, Wendi Ji, Yaliang Li, Xiaoling Wang, and Ji-Rong Wen. 2020. RecBole: Towards a Unified, Comprehensive and Efficient Framework for Recommendation Algorithms. arXiv preprint arXiv:2011.01731 (2020). [32] Ziwei Zhu, Yun He, Yin Zhang, and James Caverlee. 2020. Unbiased implicit recommendation and propensity estimation via combinational joint learning. In Fourteenth ACM Conference on Recommender Systems. 551–556. [33] Ziwei Zhu, Yun He, Xing Zhao, Yin Zhang, Jianling Wang, and James Caverlee. 2021. Popularity-Opportunity Bias in Collaborative Filtering. In Proceedings of the 14th ACM International Conference on Web Search and Data Mining (Virtual Event, Israel) (WSDM ’21). Association for Computing Machinery, New York, NY, USA, 85–93. https://doi.org/10.1145/3437963.3441820 10