<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Personalized Learning in K-12 Education: Exploring Weak-Labels for a Random Forest-based Collaborative Filtering Approach</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Pedro</forename><surname>Ilídio</surname></persName>
							<email>pedro.ilidio@kuleuven.be</email>
							<affiliation key="aff0">
								<orgName type="department">Department of Public Health and Primary Care</orgName>
								<orgName type="institution">KU Leuven</orgName>
								<address>
									<addrLine>Campus Kulak, Etienne Sabbelaan 53</addrLine>
									<postCode>8500</postCode>
									<settlement>Kortrijk</settlement>
									<country key="BE">Belgium</country>
								</address>
							</affiliation>
							<affiliation key="aff1">
								<orgName type="department">Itec</orgName>
								<orgName type="laboratory">imec research group</orgName>
								<orgName type="institution">at KU Leuven</orgName>
								<address>
									<addrLine>Etienne Sabbelaan 53</addrLine>
									<postCode>8500</postCode>
									<settlement>Kortrijk</settlement>
									<country key="BE">Belgium</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Alireza</forename><surname>Gharahighehi</surname></persName>
							<email>alireza.gharahighehi@kuleuven.be</email>
							<affiliation key="aff0">
								<orgName type="department">Department of Public Health and Primary Care</orgName>
								<orgName type="institution">KU Leuven</orgName>
								<address>
									<addrLine>Campus Kulak, Etienne Sabbelaan 53</addrLine>
									<postCode>8500</postCode>
									<settlement>Kortrijk</settlement>
									<country key="BE">Belgium</country>
								</address>
							</affiliation>
							<affiliation key="aff1">
								<orgName type="department">Itec</orgName>
								<orgName type="laboratory">imec research group</orgName>
								<orgName type="institution">at KU Leuven</orgName>
								<address>
									<addrLine>Etienne Sabbelaan 53</addrLine>
									<postCode>8500</postCode>
									<settlement>Kortrijk</settlement>
									<country key="BE">Belgium</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Felipe</forename><forename type="middle">Kenji</forename><surname>Nakano</surname></persName>
							<email>felipekenji.nakano@kuleuven.be</email>
							<affiliation key="aff0">
								<orgName type="department">Department of Public Health and Primary Care</orgName>
								<orgName type="institution">KU Leuven</orgName>
								<address>
									<addrLine>Campus Kulak, Etienne Sabbelaan 53</addrLine>
									<postCode>8500</postCode>
									<settlement>Kortrijk</settlement>
									<country key="BE">Belgium</country>
								</address>
							</affiliation>
							<affiliation key="aff1">
								<orgName type="department">Itec</orgName>
								<orgName type="laboratory">imec research group</orgName>
								<orgName type="institution">at KU Leuven</orgName>
								<address>
									<addrLine>Etienne Sabbelaan 53</addrLine>
									<postCode>8500</postCode>
									<settlement>Kortrijk</settlement>
									<country key="BE">Belgium</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Celine</forename><surname>Vens</surname></persName>
							<email>celine.vens@kuleuven.be</email>
							<affiliation key="aff0">
								<orgName type="department">Department of Public Health and Primary Care</orgName>
								<orgName type="institution">KU Leuven</orgName>
								<address>
									<addrLine>Campus Kulak, Etienne Sabbelaan 53</addrLine>
									<postCode>8500</postCode>
									<settlement>Kortrijk</settlement>
									<country key="BE">Belgium</country>
								</address>
							</affiliation>
							<affiliation key="aff1">
								<orgName type="department">Itec</orgName>
								<orgName type="laboratory">imec research group</orgName>
								<orgName type="institution">at KU Leuven</orgName>
								<address>
									<addrLine>Etienne Sabbelaan 53</addrLine>
									<postCode>8500</postCode>
									<settlement>Kortrijk</settlement>
									<country key="BE">Belgium</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Personalized Learning in K-12 Education: Exploring Weak-Labels for a Random Forest-based Collaborative Filtering Approach</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">FE828DC384E6EB56F61186CD9EE0C04E</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T19:14+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Random Forest, weak-label learning, k-12 education, educational recommendation, collaborative filtering (C. Vens) 0000-0002-5750-0204 (P. Ilídio)</term>
					<term>0000-0003-1453-1155 (A. Gharahighehi)</term>
					<term>0000-0002-4884-9420 (F. K. Nakano)</term>
					<term>0000-0003-0983-256X (C. Vens)</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Education, a cornerstone of human development, increasingly leverages digital learning tools, generating valuable data from student interactions. This data can enhance learning efficiency through adaptive and personalized systems, moving beyond the traditional "one-size-fits-all" model. Recommendation systems learn user profiles to suggest relevant items and personalize students' learning experiences. In the context of implicit feedback, binary interactions are used in weak-label learning, where negative label annotations are unreliable. This paper proposes a weak-label learning method for recommending learning materials and trajectories, combining local and global Random Forests in a multi-step collaborative filtering process. The proposed approach is named PentaForest, and outperforms other popular collaborative filtering methods in terms of NDCG and recall.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>One of the most important aspects of human development and sustainable development goals is education. In the age of digitization, the use of digital learning tools has become more popular among students, either as a primary means of completing learning assignments or as an auxiliary learning environment. This facilitates the creation of data based on learners' behaviors and interactions with these learning platforms. Such data can be used to provide data-driven interventions, making learning more efficient and effective. There is a growing interest in adaptive and personalized learning systems due to their presumed benefits on cognitive and non-cognitive learning outcomes <ref type="bibr" target="#b1">[2,</ref><ref type="bibr" target="#b2">3]</ref>. Education is continuously moving from the classic "one-size-fits-all" model to more adaptive and personalized learning approaches. As learners have different needs and preferences, they can be served accordingly.</p><p>Recommendation systems are machine learning methods that learn users' profiles based on their previous interactions and recommend items that best fit these profiles. These systems are generally categorized into two main types: content-based filtering and collaborative filtering. The former aims to match item features with user profiles, recommending items whose features best match the profiles, while the latter models user preferences and needs based on collaborative information between users and items-i.e., their interactions-to recommend items. User feedback on items is usually implicit, meaning that, in most of the cases, we do not have explicit ratings from users. In such cases, binary feedback (user interaction with an item) is the only available signal to learn user profiles. This setting is known as one-class collaborative filtering and is also referred to as positive-unlabeled (PU) learning or weak-label learning in machine learning. In this context, the given labels are all positive and missing labels are either from the negative class (i.e., the user observed the item but deliberately did not interact with it) or are missing positive labels (i.e., the user did not observe the item, otherwise (s)he would have interacted with it). Furthermore, because the given labels are based solely on simple interactions between users and items, they may be unreliable and weak positive labels. In this paper, we propose a weak-label learning method to recommend learning materials and learning trajectories to students based on their past interactions within the system. We combine local and global Random Forests in a multi-step procedure for collaborative filtering, utilizing self-learned label probabilities to address label unreliability. The resulting procedure is named PentaForest.</p><p>The following sections are organized as follows. Section 2 provides an overview of related studies. Section 3 then defines the specific learning problem being approached, and Section 4 presents our method proposal. We present our experimental setup in Section 5, and experimental results are provided in Section 6. Finally, Section 7 concludes our work and proposes future research directions.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Literature Review</head><p>We propose using ensembles of randomized decision trees, called Random Forests <ref type="bibr" target="#b3">[4]</ref>, to perform collaborative filtering in weakly-supervised settings. Decision tree-based methods for weaklysupervised tasks have not received much attention in recent years <ref type="bibr" target="#b4">[5]</ref>. Two general approaches are usually proposed: i) impute new labels in a self-supervised manner; or ii) consider the structure of the feature space when growing each tree. Here, we focus on the first approach, where the label matrix is either completed to yield a dense representation or new positive annotations are added before training the final estimator. In this context, Tanha et al. <ref type="bibr" target="#b5">[6]</ref> proposed an iterative approach, where the most confident predictions are imputed before building the next tree. Wang et al. <ref type="bibr" target="#b6">[7]</ref> focused specifically on the problem of weak-labels, where only negative annotations are unreliable. Their proposal is based on a deep forest model, in which multiple decision forests are trained sequentially and the predicted probabilities at each layer are appended as new features for the next one. The authors adapted this procedure to weak-labels by performing label imputation after each layer of the deep forest. We extended this idea and addressed limitations of the original proposal in a recent work <ref type="bibr" target="#b7">[8]</ref>. In the context of PU interaction prediction, Pliakos and Vens <ref type="bibr" target="#b8">[9]</ref> used Neighborhood-Regularized Logistic Matrix Factorization to complete the label annotations, converting the binary label matrix to a dense representation. This representation was then used to build a global multi-output forest to serve as the final estimator. Global estimators consider both item-related and user-related information during training <ref type="bibr" target="#b8">[9]</ref>. In contrast, local models take either users or items as input instances, predicting its interactions as multiple outputs <ref type="bibr" target="#b8">[9]</ref>. Gharahighehi et al. <ref type="bibr" target="#b9">[10]</ref> proposed a two-step approach to address the cold-start problem in a PU learning context. First, the interaction matrix between users and warm items is reconstructed using SLIM <ref type="bibr" target="#b10">[11]</ref>. Then, an inductive multi-target regressor is trained on this reconstructed interaction matrix to predict interactions for new items that enter the system. In the context of Massive Open Online Course (MOOC) recommendations, Gharahighehi et al. <ref type="bibr" target="#b11">[12]</ref> considered censored time-to-event data (time to dropout from MOOCs) as weak-labels and extended Bayesian Personalized Ranking <ref type="bibr" target="#b12">[13]</ref>, which is a learning-to-rank collaborative filtering approach, to incorporate these weak-labels in training the model.</p><p>In the field of recommendation systems, Random Forests were employed by Li et al. <ref type="bibr" target="#b13">[14]</ref> as a dimensionality reduction strategy, preprocessing the data before employing similarity-based collaborative filtering. Panagiotakis et al. <ref type="bibr" target="#b14">[15]</ref>, on the other hand, completed the label matrix with Synthetic Coordinate Recommendations and then trained a local Random Forest as an item recommender given a user as input. However, in both cases, Random Forests are employed to explore side-information or context-information on the problem, in contrast to the current scenario where only interactions are utilized.</p><p>We combine multi-output local forests to complete the label matrix, and then leverage the completed annotations to train a global single output forest. We now define the learning problem under study (Section 3), and a detailed description of our method is presented in Section 4.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Problem Definition</head><formula xml:id="formula_0">Let 𝒰 be a set of users {𝑢 1 , 𝑢 2 , • • • , 𝑢 𝑛 } and ℐ be a set of items {𝑖 1 , 𝑖 2 , • • • , 𝑖 𝑚 }.</formula><p>The 𝑛 × 𝑚 binary matrix 𝑌 = (𝑌 𝑙𝑝 ) then represents known interactions between the user 𝑢 𝑙 and item 𝑖 𝑝 . In this matrix, 1 is assumed to be a confirmed interaction. 0-valued entries, however, can be either non-occurring relationships or unannotated positive values, which characterizes the weak-label scenario. Each item is considered a different label and each user is considered a different sample.</p><p>Our goal is to indicate 𝑘 new items for each user 𝑢 𝑙 in 𝒰, representing the annotations that are most-likely to be missing or interactions that are most-likely to occur in the future. The set of 𝑘 indicated items for 𝑢 𝑙 is called the recommendations for the user 𝑢 𝑙 . To generate recommendations, we only receive the label matrix 𝑌 , no side-information is assumed. Having only 𝑌 characterizes collaborative filtering techniques based on implicit feedback.</p><p>In the present setting, users represent students in an online learning platform. For items, two scenarios are separately explored:</p><p>• Learning materials: items are learning activities available at the platform; • Learning trajectories: items are sets of learning materials. Each trajectory represents a path of materials manually defined by a teacher to be followed by the students.</p><p>We say a user interacted with a learning material if the material was accessed by the user, independently of the activity being concluded or not, and independently of the number of accesses. For learning trajectories, a positive value indicates the trajectory was concluded.</p><p>Weak-labels arise from the user not knowing the item, or having not interacted with it yet at the time that the dataset was built. Step 1.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Proposed Method: PentaForest</head><p>Step 2.</p><p>Step 3. The proposed procedure<ref type="foot" target="#foot_0">1</ref> employs five Random Forest (RF) <ref type="bibr" target="#b3">[4]</ref> models to perform collaborative filtering. Usually, training each RF requires a feature matrix describing the input samples and a label matrix to be modeled. In this case, however, we use 𝑌 as both the feature and the label matrices. The algorithm is divided into three main steps:</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Final probabilities</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Predicted</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Train primary item and user recommenders</head><p>• Forest 1: In the first step, a RF is trained to predict the probabilities of a given user to interact with each of the items. We call this RF an "item recommender". • Forest 2: Also in this step, another RF is built to solve the transposed problem: it predicts the probabilities of a given item to interact with each of the users. It is called an "user recommender".</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Train secondary item and user recommenders</head><p>• Forest 3: In the second step, the probabilities predicted by the item recommender are used as targets for a second user recommender. • Forest 4: We also train a second item recommender using the probabilities predicted by the user recommender.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Train single output predictor</head><p>• In the third step, we average the probabilities predicted by the secondary user recommender and the secondary item recommender. • Forest 5: The averaged probabilities are used as targets for a fifth and final RF. This RF considers concatenated user-item interaction profiles as input, and predicts the single probability of interaction between the given pair. Its predictions are the final probabilities we output. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Return 𝑌 ˜final</head><p>We note that the concatenations in the third step are not performed as such. Instead, the Bipartite Global Single Output procedure <ref type="bibr" target="#b15">[16]</ref> is used to generate the forests more efficiently from both user and item "feature matrices" directly. In each step, the user feature matrix is always the original binary label matrix itself (𝑌 ), and the transposed label matrix (𝑌 𝑇 ) is taken as the item feature matrix. We call local the forests that utilize either users or items as input samples, predicting multiple outputs for each of them <ref type="bibr" target="#b8">[9]</ref>. The global forest, on the other hand, considers both at the same time <ref type="bibr" target="#b8">[9]</ref>. The whole procedure is summarized in algorithm 1 and illustrated in Figure <ref type="figure" target="#fig_0">1</ref>.</p><p>The reason for using multiple steps of reconstruction is to generate several diverse models. We then encourage the different types of learned information to be exchanged between the forests, controlling overfitting and improving their ability to generalize. Further discussion is presented by Section 6. We also provide four reasons for utilizing the original label matrix as features instead of its completed version:</p><p>1. the original (confirmed) positive annotations are prioritized, being used to define the decision boundaries; 2. the lower cardinality of the binary labels reduces the tendency of one forest to overfit the predictions of the previous; 3. the lower cardinality also allows for faster training of the forests; 4. relying on the original annotations makes the model more easily interpretable, allowing us to look only at the final estimator's structure to gain insights on the learning task.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Dataset and Experimental Setup</head><p>We used two datasets from an educational K-12 platform in Belgium. The first dataset contains students' interactions with learning materials, while the second dataset includes students' interactions with learning trajectories, each comprising a series of learning materials defined by teachers. We excluded students and materials with fewer than 10 interactions from the first dataset and students and tracks with fewer than 5 interactions from the second dataset. This is done to ensure that a reasonable amount of information remains on the training set after masking the interactions to be used for scoring. Table <ref type="table" target="#tab_1">1</ref> provides a description of the two datasets after preprocessing. To form our training, validation, and test sets, we kept one interaction per user for the test set, one interaction per user for the validation set, and all remaining interactions for the training set. This way of splitting ensures that the same set of users and items appears in the training, test, and validation sets.</p><p>For evaluation, we applied two measures: normalized discounted cumulative gain (NDCG) (Eq. 1) and recall (Eq. 2).</p><formula xml:id="formula_1">𝐷𝐶𝐺 𝑢 = |𝑅𝑢| ∑︁ 𝑖=1 2 𝑟𝑒𝑙 𝑖 − 1 𝑙𝑜𝑔 2 (𝑖 + 1) , 𝐼𝐷𝐶𝐺 𝑢 = |𝑅_𝑖𝑑𝑙𝑢| ∑︁ 𝑖=1 2 𝑟𝑒𝑙 𝑖 − 1 𝑙𝑜𝑔 2 (𝑖 + 1) , 𝑁 𝐷𝐶𝐺 = 1 |𝑈 | ∑︁ 𝑢∈𝑈 𝐷𝐶𝐺 𝑢 𝐼𝐷𝐶𝐺 𝑢 .</formula><p>(1)</p><formula xml:id="formula_2">𝑅𝑒𝑐𝑎𝑙𝑙 = 1 |𝑈 | ∑︁ 𝑢∈𝑈 |𝑇 𝑢 ∩ 𝑅 𝑢 | |𝑇 𝑢 | . (<label>2</label></formula><formula xml:id="formula_3">)</formula><p>where 𝑈 is the set of users, 𝑇 𝑢 is the test items for user 𝑢, 𝑅 𝑢 is the recommendation list for user 𝑢, 𝑟𝑒𝑙 𝑖 is the real rating value of the 𝑖 𝑡ℎ item in 𝑅 𝑢 , 𝐼𝐷𝐶𝐺 is the ideal DCG value, and 𝑅_𝑖𝑑𝑙 𝑢 is the ideal recommendation list for user 𝑢, that one can create based on the ground truth. Therefore, the 𝐷𝐶𝐺 value for each user is normalized with the ideal value (𝐼𝐷𝐶𝐺) to get the 𝑁 𝐷𝐶𝐺 value for that user.</p><p>As competing methods, we included five approaches:</p><p>• UKNN: user-based (UKNN) collaborative filtering (CF) <ref type="bibr" target="#b16">[17]</ref> is a memory-based CF methods that impute missing interactions between users and items based on the interactions of neighbor users. • IKNN: item-based (UKNN) collaborative filtering (CF) <ref type="bibr" target="#b17">[18]</ref> is a memory-based CF methods that impute missing interactions between users and items based on the interactions of neighbor items.</p><p>• WRMF: weighted regularized matrix factorization (WRMF) <ref type="bibr" target="#b18">[19]</ref> is a model-based CF method that utilizes the alternating-least-squares optimization algorithm to learn its parameters. • EASE: Embarrassingly Shallow Autoencoders (EASE) <ref type="bibr" target="#b19">[20]</ref> is a linear collaborative filtering model for implicit feedback datasets based on shallow autoencoders <ref type="bibr" target="#b20">[21]</ref>. • MVAE: multi-variational autoencoders (MVAE) <ref type="bibr" target="#b21">[22]</ref> is a CF-based recommender system for implicit feedback, based on variational autoencoders, with the main assumption that user interactions follow a multinomial distribution.</p><p>We selected these methods because they demonstrated high performance in collaborative filtering tasks, as reported in the award-winning paper by Dacrema et al. <ref type="bibr" target="#b22">[23]</ref>, which showed that memory-based approaches (UKNN and IKNN) and MVAE outperform recent complex deep neural network-based approaches .</p><p>As for the hyperparameters of our method, each forest was composed of 1000 trees. The forests employed bootstrapping to resample the set of input instances, and were grown until their maximum depth. The remaining hyperparameters were left to their default values in the scikit-learn package. For instance, the objective criterion was set to the mean squared error, and no feature sampling was performed.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.">Results and Discussion</head><p>The results of applying the proposed method and the competing approaches are reported in Tables <ref type="table" target="#tab_4">2 and 3</ref> for the learning material and trajectory recommendation tasks, respectively. The proposed PentaForest method is shown to clearly outperform the other competing approaches in both tasks. In terms of recall, PentaForest is especially superior when fewer recommendations are selected, which represents the most difficult tasks under study (see top 3 in Table <ref type="table" target="#tab_4">2</ref> and Table <ref type="table">3</ref>). Regarding NDCG, the superiority of our method is evident and consistent across all scenarios. Again, this suggests that the proposed technique is especially proficient for fewer recommendations, for the reason that NDCG prioritizes the recommendations predicted to be most likely. As such, NDCG is less sensitive to the number of recommendations we select for evaluation.</p><p>The promising results corroborate the hypothesized benefits of weak-label techniques in recommendation contexts. Note that using the label matrix as features for the Random Forests means that each tree will cluster instances with similar labels. When we use self-learned label probabilities to train a new forest, we are instructing this forest to cluster labels that are expected to be similar, even if not similar from the original label annotations. This is what mitigates the effect of uncertain annotations.</p><p>Furthermore, we argue that our model proficiency is in great part due to the diversity of the forests employed. This diversity might not be apparent at a first glance, since all the component models are based on the same Random Forest algorithm. Notwithstanding, notice that the secondary user recommender is trained on the outputs of the primary item recommender, and vice-versa. This compels the secondary estimators to learn a representation of the problem that is different from the primary models. That is, the predictions made based on the user-wise information now need to be explained from the item information alone, and vice-versa. Similarly, the representation learned at the final step by the global Random Forest is also different. This forest will induce a biclustering of the interaction matrix, grouping interactions that are similar both in terms of user and item profiles. This differs from the local forests used in the first two steps, that only consider either user-user or item-item label similarity. Finally, using a global forest that receives the original binary labels yields a crucial advantage for the deployment of these systems: it enables acknowledging new interactions, that were added to the dataset after training the models. This is due to the fact that the final estimator only needs the user and item interaction profiles to estimate new interaction probabilities. As such, updated interaction profiles could be provided to obtain updated recommendations without rebuilding the models. Even interactions between unseen items and unseen users can be inferred, based on their interactions with the users and items in the training set.</p><p>The global forest also allows estimating probabilities only for a subset of items and users. This can be used, for example, to provide recommendations within a user-specified category of activities, without generating probabilities for all activities in the training set.</p><p>An educational system could also benefit from the known transparency of forest estimators. It could, for instance, prioritize items with higher importance when making recommendations, instead of only relying on the predicted probabilities. This could enable the system to discover new interests of a user, making recommendations that are not necessarily similar to the user's previous interactions, but are useful for profiling the user's preference.</p><p>In summary, the trained final forest of a PentaForest model can be employed to create highly adaptive recommendation systems, for a dynamic and personalized learning experience.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="7.">Conclusion</head><p>In this paper we proposed PentaForest, a new collaborative filtering approach for recommending learning material and learning trajectories to students in a web-based educational platform. Our method combines local and global Random Forests in a way to acknowledge the weaklysupervised nature of our problem. Our results suggest state-of-the-art performance in our current learning settings, prompting future studies to further investigate decision forests for recommendation tasks in general. In subsequent work, we would like to explore semi-supervised split evaluation metrics, since they have been shown to be useful for other PU-learning scenarios <ref type="bibr" target="#b23">[24,</ref><ref type="bibr" target="#b15">16]</ref>. We would also like to evaluate the potential of deeper ensembles of decision forests as recommender systems, adapting deep forests presented by Wang et al. <ref type="bibr" target="#b6">[7]</ref>, Ilídio et al. <ref type="bibr" target="#b7">[8]</ref>. In contrast, shallower versions of our estimators might also be an interesting topic of study in a broader variety of problems, performing more detailed ablation analysis to elucidate the effectiveness of the reconstruction steps. Furthermore, exploring the explainability of decision trees could be a valuable method for selecting the most informative activities to recommend when profiling new users. Finally, the use of side-features is also a topic for further exploration, as a way of mitigating the cold start problem or possibly improving transductive predictions.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: The PentaForest procedure.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Algorithm 1 :</head><label>1</label><figDesc>PentaForest: Random Forests for Collaborative Filtering Input: User-item interaction matrix 𝑌 Output: Completed matrix 𝑌 Step 1: Train primary recommenders 𝑌 ˜items, 1 ← 𝐿𝑜𝑐𝑎𝑙𝑅𝑎𝑛𝑑𝑜𝑚𝐹 𝑜𝑟𝑒𝑠𝑡.𝑓 𝑖𝑡_𝑝𝑟𝑒𝑑𝑖𝑐𝑡_𝑝𝑟𝑜𝑏𝑎(𝑌, 𝑌 ) 𝑌 ˜𝑇 users, 1 ← 𝐿𝑜𝑐𝑎𝑙𝑅𝑎𝑛𝑑𝑜𝑚𝐹 𝑜𝑟𝑒𝑠𝑡.𝑓 𝑖𝑡_𝑝𝑟𝑒𝑑𝑖𝑐𝑡_𝑝𝑟𝑜𝑏𝑎(𝑌 𝑇 , 𝑌 𝑇 )</figDesc><table /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>Step 2: Train secondary recommenders</head><label></label><figDesc>𝑌 ˜𝑇 users, 2 ← 𝐿𝑜𝑐𝑎𝑙𝑅𝑎𝑛𝑑𝑜𝑚𝐹 𝑜𝑟𝑒𝑠𝑡.𝑓 𝑖𝑡_𝑝𝑟𝑒𝑑𝑖𝑐𝑡_𝑝𝑟𝑜𝑏𝑎(𝑌 𝑇 , 𝑌 𝑇 items, 1 ) 𝑌 ˜items, 2 ← 𝐿𝑜𝑐𝑎𝑙𝑅𝑎𝑛𝑑𝑜𝑚𝐹 𝑜𝑟𝑒𝑠𝑡.𝑓 𝑖𝑡_𝑝𝑟𝑒𝑑𝑖𝑐𝑡_𝑝𝑟𝑜𝑏𝑎(𝑌, 𝑌 users, 1 ) 𝑎𝑣𝑒𝑟𝑎𝑔𝑒(𝑌 ˜users, 2 , 𝑌 ˜items, 2 ) 𝑌 ˜final ← 𝐺𝑙𝑜𝑏𝑎𝑙𝑅𝑎𝑛𝑑𝑜𝑚𝐹 𝑜𝑟𝑒𝑠𝑡.𝑓 𝑖𝑡_𝑝𝑟𝑒𝑑𝑖𝑐𝑡_𝑝𝑟𝑜𝑏𝑎(𝑌, 𝑌 𝑇 , 𝑌 ˜avg )</figDesc><table><row><cell>Step 3: Final prediction</cell></row><row><cell>𝑌 ˜avg ←</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_3"><head>Table 1</head><label>1</label><figDesc>Datasets Description</figDesc><table><row><cell></cell><cell cols="4">Learning material dataset Learning trajectory dataset</cell></row><row><cell></cell><cell>Original</cell><cell>preprocessed</cell><cell>Original</cell><cell>preprocessed</cell></row><row><cell># users</cell><cell>8573</cell><cell>1171</cell><cell>9211</cell><cell>1183</cell></row><row><cell># Items</cell><cell>3306</cell><cell>1057</cell><cell>1234</cell><cell>577</cell></row><row><cell>Density</cell><cell>0.2%</cell><cell>1.8%</cell><cell>0.2%</cell><cell>1.5%</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_4"><head>Table 2</head><label>2</label><figDesc>Results for learning material recommendation</figDesc><table><row><cell></cell><cell cols="2">top 3</cell><cell cols="2">top 5</cell><cell cols="2">top 10</cell></row><row><cell></cell><cell cols="6">NDCG Recall NDCG Recall NDCG Recall</cell></row><row><cell>UKNN</cell><cell>0.709</cell><cell>0.746</cell><cell>0.751</cell><cell>0.837</cell><cell>0.782</cell><cell>0.924</cell></row><row><cell>IKNN</cell><cell>0.717</cell><cell>0.756</cell><cell>0.760</cell><cell>0.846</cell><cell>0.790</cell><cell>0.933</cell></row><row><cell>WRMF</cell><cell>0.668</cell><cell>0.726</cell><cell>0.729</cell><cell>0.811</cell><cell>0.750</cell><cell>0.873</cell></row><row><cell>EASE</cell><cell>0.710</cell><cell>0.746</cell><cell>0.752</cell><cell>0.836</cell><cell>0.783</cell><cell>0.923</cell></row><row><cell>MVAE</cell><cell>0.792</cell><cell>0.830</cell><cell>0.826</cell><cell>0.915</cell><cell>0.840</cell><cell>0.977</cell></row><row><cell>PentaForest</cell><cell>0.885</cell><cell>0.900</cell><cell>0.909</cell><cell>0.952</cell><cell>0.921</cell><cell>0.987</cell></row><row><cell>Table 3</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell cols="3">Results for learning trajectory recommendation</cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell></cell><cell cols="2">top 3</cell><cell cols="2">top 5</cell><cell cols="2">top 10</cell></row><row><cell></cell><cell cols="6">NDCG Recall NDCG Recall NDCG Recall</cell></row><row><cell>UKNN</cell><cell>0.814</cell><cell>0.854</cell><cell>0.847</cell><cell>0.922</cell><cell>0.862</cell><cell>0.963</cell></row><row><cell>IKNN</cell><cell>0.819</cell><cell>0.860</cell><cell>0.852</cell><cell>0.927</cell><cell>0.866</cell><cell>0.967</cell></row><row><cell>WRMF</cell><cell>0.582</cell><cell>0.609</cell><cell>0.616</cell><cell>0.680</cell><cell>0.634</cell><cell>0.731</cell></row><row><cell>EASE</cell><cell>0.812</cell><cell>0.850</cell><cell>0.847</cell><cell>0.923</cell><cell>0.861</cell><cell>0.963</cell></row><row><cell>MVAE</cell><cell>0.840</cell><cell>0.881</cell><cell>0.882</cell><cell>0.968</cell><cell>0.889</cell><cell>0.997</cell></row><row><cell>PentaForest</cell><cell>0.931</cell><cell>0.952</cell><cell>0.954</cell><cell>0.979</cell><cell>0.950</cell><cell>0.995</cell></row></table></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_0">The implementation of the PentaForest algorithm is available at https://itec.kuleuven-kulak.be/supporting-material/.</note>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgments</head><p>The work received funding from the Flemish Government (AI Research Program). The authors would like to also thank FWO, 1235924N.</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Adaptive lifelong learning (all)</title>
		<author>
			<persName><forename type="first">A</forename><surname>Gharahighehi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Van Schoors</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Topali</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Ooge</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">International Conference on Artificial Intelligence in Education</title>
				<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2024">2024</date>
			<biblScope unit="page" from="452" to="459" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">A review of artificial intelligence (ai) in education from 2010 to 2020</title>
		<author>
			<persName><forename type="first">X</forename><surname>Zhai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Chu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">S</forename><surname>Chai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">S Y</forename><surname>Jong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Istenic</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Spector</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J.-B</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Yuan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Li</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Complexity</title>
		<imprint>
			<biblScope unit="volume">2021</biblScope>
			<biblScope unit="page" from="1" to="18" />
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Ai technologies for education: Recent research &amp; future directions</title>
		<author>
			<persName><forename type="first">K</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">B</forename><surname>Aslan</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Computers and Education: Artificial Intelligence</title>
		<imprint>
			<biblScope unit="volume">2</biblScope>
			<biblScope unit="page">100025</biblScope>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<monogr>
		<author>
			<persName><forename type="first">L</forename><surname>Breiman</surname></persName>
		</author>
		<title level="m">Random forests, Machine Learning</title>
				<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2001">2001</date>
			<biblScope unit="volume">45</biblScope>
			<biblScope unit="page" from="5" to="32" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Recent advances in decision trees: an updated survey</title>
		<author>
			<persName><forename type="first">V</forename><forename type="middle">G</forename><surname>Costa</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">E</forename><surname>Pedreira</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Artificial Intelligence Review</title>
		<imprint>
			<biblScope unit="volume">56</biblScope>
			<biblScope unit="page" from="4765" to="4800" />
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Semi-supervised self-training for decision tree classifiers</title>
		<author>
			<persName><forename type="first">J</forename><surname>Tanha</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Van Someren</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Afsarmanesh</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">International Journal of Machine Learning and Cybernetics</title>
		<imprint>
			<biblScope unit="volume">8</biblScope>
			<biblScope unit="page" from="355" to="370" />
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Learning from weak-label data: a deep forest expedition</title>
		<author>
			<persName><forename type="first">Q.-W</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Yang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y.-F</forename><surname>Li</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 34th AAAI Conference on Artificial Intelligence</title>
				<meeting>the 34th AAAI Conference on Artificial Intelligence<address><addrLine>New York</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="6251" to="6258" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Deep forests with tree-embeddings and label imputation for weak-label learning</title>
		<author>
			<persName><forename type="first">P</forename><surname>Ilídio</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Vens</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Cerri</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><forename type="middle">K</forename><surname>Nakano</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2024 International joint conference on neural networks, IJCNN</title>
				<meeting>the 2024 International joint conference on neural networks, IJCNN</meeting>
		<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Drug-target interaction prediction with tree-ensemble learning and output space reconstruction</title>
		<author>
			<persName><forename type="first">K</forename><surname>Pliakos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Vens</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">BMC Bioinformatics</title>
		<imprint>
			<biblScope unit="volume">21</biblScope>
			<biblScope unit="page" from="1" to="11" />
			<date type="published" when="2020">2020</date>
			<publisher>Springer</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Addressing the cold-start problem in collaborative filtering through positive-unlabeled learning and multi-target prediction</title>
		<author>
			<persName><forename type="first">A</forename><surname>Gharahighehi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Pliakos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Vens</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Ieee Access</title>
		<imprint>
			<biblScope unit="volume">10</biblScope>
			<biblScope unit="page" from="117189" to="117198" />
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Slim: Sparse linear methods for top-n recommender systems</title>
		<author>
			<persName><forename type="first">X</forename><surname>Ning</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Karypis</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">2011 IEEE 11th international conference on data mining</title>
				<imprint>
			<publisher>IEEE</publisher>
			<date type="published" when="2011">2011</date>
			<biblScope unit="page" from="497" to="506" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Extending bayesian personalized ranking with survival analysis for mooc recommendation</title>
		<author>
			<persName><forename type="first">A</forename><surname>Gharahighehi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Venturini</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Ghinis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Cornillie</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Vens</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Adjunct Proceedings of the 31st ACM Conference on User Modeling, Adaptation and Personalization</title>
				<imprint>
			<date type="published" when="2023">2023</date>
			<biblScope unit="page" from="56" to="59" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<monogr>
		<author>
			<persName><forename type="first">S</forename><surname>Rendle</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Freudenthaler</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Gantner</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Schmidt-Thieme</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1205.2618</idno>
		<title level="m">Bpr: Bayesian personalized ranking from implicit feedback</title>
				<imprint>
			<date type="published" when="2012">2012</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">A Multi-Dimensional Context-Aware Recommendation Approach Based on Improved Random Forest Algorithm</title>
		<author>
			<persName><forename type="first">X</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Hu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><surname>Zhu</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Access</title>
		<imprint>
			<biblScope unit="volume">6</biblScope>
			<biblScope unit="page" from="45071" to="45085" />
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">A dual hybrid recommender system based on SCoR and the random forest</title>
		<author>
			<persName><forename type="first">C</forename><surname>Panagiotakis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Papadakis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Fragopoulou</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Computer Science and Information Systems</title>
		<imprint>
			<biblScope unit="volume">18</biblScope>
			<biblScope unit="page" from="115" to="128" />
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">Fast Bipartite Forests for Semi-supervised Interaction Prediction</title>
		<author>
			<persName><forename type="first">P</forename><surname>Ilídio</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Alves</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Cerri</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 39th ACM/SIGAPP Symposium on Applied Computing, SAC &apos;24</title>
				<meeting>the 39th ACM/SIGAPP Symposium on Applied Computing, SAC &apos;24<address><addrLine>New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>Association for Computing Machinery</publisher>
			<date type="published" when="2024">2024</date>
			<biblScope unit="page" from="979" to="986" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">Item-based collaborative filtering recommendation algorithms</title>
		<author>
			<persName><forename type="first">B</forename><surname>Sarwar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Karypis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Konstan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Riedl</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 10th international conference on World Wide Web</title>
				<meeting>the 10th international conference on World Wide Web</meeting>
		<imprint>
			<date type="published" when="2001">2001</date>
			<biblScope unit="page" from="285" to="295" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">Content-based recommender systems: State of the art and trends</title>
		<author>
			<persName><forename type="first">P</forename><surname>Lops</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>De Gemmis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Semeraro</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Recommender systems handbook</title>
				<imprint>
			<date type="published" when="2011">2011</date>
			<biblScope unit="page" from="73" to="105" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<analytic>
		<title level="a" type="main">One-class collaborative filtering</title>
		<author>
			<persName><forename type="first">R</forename><surname>Pan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Zhou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Cao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><forename type="middle">N</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Lukose</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Scholz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><surname>Yang</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">2008 Eighth IEEE International Conference on Data Mining</title>
				<imprint>
			<publisher>IEEE</publisher>
			<date type="published" when="2008">2008</date>
			<biblScope unit="page" from="502" to="511" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<analytic>
		<title level="a" type="main">Embarrassingly shallow autoencoders for sparse data</title>
		<author>
			<persName><forename type="first">H</forename><surname>Steck</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">The World Wide Web Conference</title>
				<imprint>
			<date type="published" when="2019">2019</date>
			<biblScope unit="page" from="3251" to="3257" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<analytic>
		<title level="a" type="main">Wide &amp; deep learning for recommender systems</title>
		<author>
			<persName><forename type="first">H.-T</forename><surname>Cheng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Koc</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Harmsen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Shaked</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Chandra</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Aradhye</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Anderson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Corrado</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Chai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Ispir</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 1st workshop on deep learning for recommender systems</title>
				<meeting>the 1st workshop on deep learning for recommender systems</meeting>
		<imprint>
			<date type="published" when="2016">2016</date>
			<biblScope unit="page" from="7" to="10" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b21">
	<analytic>
		<title level="a" type="main">Variational autoencoders for collaborative filtering</title>
		<author>
			<persName><forename type="first">D</forename><surname>Liang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">G</forename><surname>Krishnan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">D</forename><surname>Hoffman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Jebara</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2018 world wide web conference</title>
				<meeting>the 2018 world wide web conference</meeting>
		<imprint>
			<date type="published" when="2018">2018</date>
			<biblScope unit="page" from="689" to="698" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b22">
	<analytic>
		<title level="a" type="main">Are we really making much progress? a worrying analysis of recent neural recommendation approaches</title>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">F</forename><surname>Dacrema</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Cremonesi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Jannach</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 13th ACM conference on recommender systems</title>
				<meeting>the 13th ACM conference on recommender systems</meeting>
		<imprint>
			<date type="published" when="2019">2019</date>
			<biblScope unit="page" from="101" to="109" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b23">
	<analytic>
		<title level="a" type="main">Semi-supervised hybrid predictive bi-clustering trees for drug-target interaction prediction</title>
		<author>
			<persName><forename type="first">A</forename><surname>Alves</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Ilidio</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Cerri</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 38th ACM/SIGAPP Symposium on Applied Computing</title>
				<meeting>the 38th ACM/SIGAPP Symposium on Applied Computing<address><addrLine>Tallinn</addrLine></address></meeting>
		<imprint>
			<publisher>ACM/SIGAPP</publisher>
			<date type="published" when="2023">2023</date>
			<biblScope unit="page" from="1163" to="1170" />
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
